| |
| """Extract lexicon data for Avestan (ave), Old Persian (peo), and Phrygian (xpg). |
| |
| Data sourced from Wiktionary Swadesh lists and category lemma pages. |
| Romanized forms are taken directly from Wiktionary's parenthesized |
| transliterations; IPA is generated via transliteration_maps; SCA via |
| cognate_pipeline's ipa_to_sound_class. |
| |
| Usage: |
| python scripts/extract_ave_peo_xpg.py |
| """ |
|
|
| from __future__ import annotations |
|
|
| import json |
| import re |
| import subprocess |
| import sys |
| import time |
| import urllib.parse |
| import urllib.request |
| from pathlib import Path |
|
|
| |
| |
| |
| ROOT = Path(__file__).resolve().parent.parent |
| sys.path.insert(0, str(ROOT / "cognate_pipeline" / "src")) |
| sys.path.insert(0, str(ROOT / "scripts")) |
|
|
| from cognate_pipeline.normalise.sound_class import ipa_to_sound_class |
| from transliteration_maps import transliterate |
|
|
|
|
| |
| |
| |
| USER_AGENT = "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36" |
| MAX_RETRIES = 3 |
| RETRY_DELAY = 10 |
|
|
|
|
| def wiki_api(params: dict) -> dict: |
| """Call the Wiktionary API and return parsed JSON, with retry on 429.""" |
| params["format"] = "json" |
| qs = "&".join(f"{k}={urllib.parse.quote(str(v))}" for k, v in params.items()) |
| url = f"https://en.wiktionary.org/w/api.php?{qs}" |
| req = urllib.request.Request(url, headers={"User-Agent": USER_AGENT}) |
|
|
| for attempt in range(MAX_RETRIES): |
| try: |
| with urllib.request.urlopen(req, timeout=30) as resp: |
| return json.loads(resp.read().decode("utf-8")) |
| except urllib.error.HTTPError as e: |
| if e.code == 429 and attempt < MAX_RETRIES - 1: |
| wait = RETRY_DELAY * (attempt + 1) |
| print(f" Rate limited, waiting {wait}s...", file=sys.stderr) |
| time.sleep(wait) |
| else: |
| raise |
|
|
|
|
| def fetch_rendered_page(page_title: str) -> str: |
| """Fetch a Wiktionary page and return text-stripped HTML.""" |
| data = wiki_api({ |
| "action": "parse", |
| "page": page_title, |
| "prop": "text", |
| }) |
| if "parse" not in data: |
| return "" |
| html = data["parse"]["text"]["*"] |
| clean = re.sub(r"<[^>]+>", " ", html) |
| clean = re.sub(r"\s+", " ", clean) |
| return clean |
|
|
|
|
| def fetch_page_by_id(page_id: int) -> tuple[str, str]: |
| """Fetch a Wiktionary page by ID, return (title, cleaned_text).""" |
| data = wiki_api({ |
| "action": "parse", |
| "pageid": str(page_id), |
| "prop": "text", |
| }) |
| if "parse" not in data: |
| return ("", "") |
| html = data["parse"]["text"]["*"] |
| clean = re.sub(r"<[^>]+>", " ", html) |
| clean = re.sub(r"\s+", " ", clean) |
| return (data["parse"]["title"], clean) |
|
|
|
|
| def fetch_category_members(category: str) -> list[dict]: |
| """Fetch all members of a Wiktionary category.""" |
| all_members = [] |
| params = { |
| "action": "query", |
| "list": "categorymembers", |
| "cmtitle": category, |
| "cmlimit": "500", |
| } |
| while True: |
| data = wiki_api(params) |
| members = data.get("query", {}).get("categorymembers", []) |
| all_members.extend(members) |
| if "continue" not in data: |
| break |
| params["cmcontinue"] = data["continue"]["cmcontinue"] |
| time.sleep(1) |
| return all_members |
|
|
|
|
| |
| |
| |
|
|
| |
| |
| |
| |
| |
|
|
| AVESTAN_SWADESH = [ |
| |
| ("azə̄m", "I"), |
| ("tūm", "you"), |
| ("tū", "you"), |
| ("tuuə̄m", "you"), |
| ("vaēm", "we"), |
| ("vā", "we"), |
| ("ahma", "we"), |
| ("nā̊", "we"), |
| ("yūžəm", "you_pl"), |
| ("vā̊", "you_pl"), |
| ("ha", "this"), |
| ("aiiə̄m", "this"), |
| ("aēm", "this"), |
| ("ta", "that"), |
| ("auua", "that"), |
| ("ka", "who"), |
| ("vīspa", "all"), |
| ("hauruua", "whole"), |
| ("aniia", "other"), |
| ("aṇtara", "other"), |
| ("aēuua", "one"), |
| ("duua", "two"), |
| ("θri", "three"), |
| ("caθβar", "four"), |
| ("paṇca", "five"), |
| ("maza", "big"), |
| ("darəga", "long"), |
| ("pərəθu", "wide"), |
| ("gouruš", "heavy"), |
| ("strī", "woman"), |
| ("nāirī", "woman"), |
| ("nar", "man"), |
| ("vīra", "man"), |
| ("manuš", "human"), |
| ("maṣ̌iia", "human"), |
| ("paiti", "husband"), |
| ("barəθrī", "mother"), |
| ("mātar", "mother"), |
| ("pitar", "father"), |
| ("ząθar", "father"), |
| ("masiia", "fish"), |
| ("vaii", "bird"), |
| ("mərəγa", "bird"), |
| ("span", "dog"), |
| ("aži", "snake"), |
| ("varəša", "tree"), |
| ("gao", "meat"), |
| ("vohunī", "blood"), |
| ("ast", "bone"), |
| ("aēm", "egg"), |
| ("sāra", "head"), |
| ("gaoša", "ear"), |
| ("karəna", "ear"), |
| ("aši", "eye"), |
| ("cašman", "eye"), |
| ("daēman", "eye"), |
| ("hizuuā", "tongue"), |
| ("pad", "foot"), |
| ("pāδa", "foot"), |
| ("zānu", "knee"), |
| ("zasta", "hand"), |
| ("udara", "belly"), |
| ("maršū", "belly"), |
| ("mərəzāna", "belly"), |
| ("paršta", "back"), |
| ("varah", "breast"), |
| ("fštāna", "breast"), |
| ("zərəd", "heart"), |
| ("xᵛaraiti", "to_drink"), |
| ("vāiti", "to_blow"), |
| ("zānaiti", "to_know"), |
| ("vid", "to_know"), |
| ("juuaiti", "to_live"), |
| ("iriθiieiti", "to_die"), |
| ("miriieitē", "to_die"), |
| ("kərəntaiti", "to_cut"), |
| ("kan", "to_dig"), |
| ("gam", "to_come"), |
| ("dā", "to_give"), |
| ("dadāiti", "to_give"), |
| ("naēnižaiti", "to_wash"), |
| ("xᵛan", "sun"), |
| ("huuarə", "sun"), |
| ("māh", "moon"), |
| ("star", "star"), |
| ("ap", "water"), |
| ("āpō", "water"), |
| ("zraiiaŋh", "sea"), |
| ("pąsnu", "dust"), |
| ("zam", "earth"), |
| ("būmi", "earth"), |
| ("maēγa", "cloud"), |
| ("aβra", "cloud"), |
| ("snaoδa", "cloud"), |
| ("dunman", "cloud"), |
| ("dyaoš", "sky"), |
| ("vāta", "wind"), |
| ("snaēžaiti", "to_snow"), |
| ("ātar", "fire"), |
| ("ātriia", "ash"), |
| ("dažaiti", "to_burn"), |
| ("dažiiete", "to_burn"), |
| ("kaofa", "mountain"), |
| ("gairi", "mountain"), |
| ("pauruuatā", "mountain"), |
| ("spaēta", "white"), |
| ("sāma", "black"), |
| ("siāuua", "black"), |
| ("xṣ̌ap", "night"), |
| ("xṣ̌apan", "night"), |
| ("asarə", "day"), |
| ("xṣ̌apara", "day"), |
| ("yārə", "year"), |
| ("sarəta", "cold"), |
| ("aota", "cold"), |
| ("pərənāiiu", "old"), |
| ("hu", "good"), |
| ("friia", "good"), |
| ("aka", "bad"), |
| ("aγa", "bad"), |
| ("aŋra", "bad"), |
| ("āhīta", "dirty"), |
| ("huška", "dry"), |
| ("ərəš", "correct"), |
| ("haiθiia", "correct"), |
| ("dūirē", "far"), |
| ("dašina", "right"), |
| ("hauuiia", "left"), |
| ("aṇtara", "in"), |
| ("aṇtarə", "in"), |
| ("pairi", "with"), |
| ("ca", "and"), |
| ("cā", "and"), |
| ("utā", "and"), |
| ("zī", "because"), |
| ("nāman", "name"), |
| ] |
|
|
| OLD_PERSIAN_SWADESH = [ |
| |
| ("adam", "I"), |
| ("tuvam", "you"), |
| ("hauv", "he"), |
| ("iyam", "this"), |
| ("aita", "this"), |
| ("idā", "here"), |
| ("avadā", "there"), |
| ("naiy", "not"), |
| ("aniya", "other"), |
| ("aiva", "one"), |
| ("vazạrka", "big"), |
| ("darga", "long"), |
| ("dargam", "long"), |
| ("martiyaʰ", "man"), |
| ("zana", "man"), |
| ("mātā", "mother"), |
| ("pitā", "father"), |
| ("cašman", "eye"), |
| ("nāham", "nose"), |
| ("hazānam", "tongue"), |
| ("pādaʰ", "foot"), |
| ("dasta", "hand"), |
| ("tạrsatiy", "to_fear"), |
| ("jīvatiy", "to_live"), |
| ("θātiy", "to_say"), |
| ("dārayatiy", "to_hold"), |
| ("ʰuvar", "sun"), |
| ("māh", "moon"), |
| ("draya", "sea"), |
| ("aθaⁿgaʰ", "stone"), |
| ("asan", "stone"), |
| ("būmiš", "earth"), |
| ("asman", "sky"), |
| ("ātar", "fire"), |
| ("kaufaʰ", "mountain"), |
| ("xšapa", "night"), |
| ("raucah", "day"), |
| ("θarda", "year"), |
| ("hašiyam", "correct"), |
| ("hadā", "with"), |
| ("utā", "and"), |
| ("yaθā", "because"), |
| ("nāma", "name"), |
| ("nāmā", "name"), |
| ("apiyā", "water"), |
| ("paruv", "many"), |
| ] |
|
|
| |
| |
| OLD_PERSIAN_RECONSTRUCTION = [ |
| ("aivacadaθa", "province"), |
| ("amah", "we"), |
| ("apamah", "farthest"), |
| ("arbah", "small"), |
| ("arušah", "white"), |
| ("arvāh", "value"), |
| ("badrah", "treasure"), |
| ("baivaram", "ten_thousand"), |
| ("baivarapatiš", "commander_of_ten_thousand"), |
| ("bādukah", "wind"), |
| ("bāgah", "god"), |
| ("bātah", "tribute"), |
| ("ciθrah", "lineage"), |
| ("darikah", "gold_coin"), |
| ("daθa", "law"), |
| ("daθapatiš", "judge"), |
| ("farnah", "glory"), |
| ("fšupā", "shepherd"), |
| ("gaiθā", "possession"), |
| ("gandabarah", "treasurer"), |
| ("ganzabarah", "treasurer"), |
| ("gaupā", "cowherd"), |
| ("gāuš", "cow"), |
| ("hadāram", "together"), |
| ("hammārakarah", "accountant"), |
| ("haxā", "companion"), |
| ("hazārapatiš", "chiliarch"), |
| ("hvatah", "self"), |
| ("jamānā", "time"), |
| ("kāratākah", "worker"), |
| ("kāravā", "people"), |
| ("madu", "wine"), |
| ("maniš", "thought"), |
| ("margā", "death"), |
| ("marzapā", "satrap"), |
| ("maθištah", "greatest"), |
| ("mānā", "house"), |
| ("naftah", "naphtha"), |
| ("nauciš", "nine"), |
| ("navacadaθa", "nine_provinces"), |
| ("patiš", "lord"), |
| ("raivah", "wealth"), |
| ("rauxšnah", "light"), |
| ("rauxšnā", "light"), |
| ("rāivāh", "rich"), |
| ("spantah", "holy"), |
| ("sparabarah", "shield_bearer"), |
| ("spiθrah", "white"), |
| ("upaganzabarah", "sub_treasurer"), |
| ("vahištah", "best"), |
| ("vahuš", "good"), |
| ("vahār", "spring"), |
| ("varāzah", "boar"), |
| ("vazdāh", "prosperity"), |
| ("vātah", "wind"), |
| ("xšayah", "king"), |
| ("xšaθram", "kingdom"), |
| ("xšaθrapā", "satrap"), |
| ("šarguš", "mule"), |
| ("θatam", "hundred"), |
| ("θatapatiš", "centurion"), |
| ("θigrah", "sharp"), |
| ("θuxrah", "firm"), |
| ("θāyakā", "decree"), |
| ("çahma", "which"), |
| ] |
|
|
| |
| |
| PHRYGIAN_ADDITIONAL = [ |
| ("ad", "to"), |
| ("aini", "if"), |
| ("knais", "woman"), |
| ("abberet", "carries"), |
| ("akke", "make"), |
| ("as", "from"), |
| ("autos", "self"), |
| ("aōrō", "plowed"), |
| ("bekos", "bread"), |
| ("bratere", "brother"), |
| ("breit", "cut"), |
| ("daket", "makes"), |
| ("deos", "god"), |
| ("edaes", "placed"), |
| ("eitou", "let"), |
| ("estaes", "stood"), |
| ("ke", "and"), |
| ("kin", "who"), |
| ("lawagtaei", "leader"), |
| ("matar", "mother"), |
| ("onoman", "name"), |
| ("paterēs", "father"), |
| ("pinke", "five"), |
| ] |
|
|
| PHRYGIAN_SWADESH = [ |
| |
| ("semoun", "this"), |
| ("semou", "this"), |
| ("kin", "who"), |
| ("thri", "three"), |
| ("pinke", "five"), |
| ("zemelōs", "man"), |
| ("knaikan", "wife"), |
| ("anar", "husband"), |
| ("matar", "mother"), |
| ("estaes", "to_stand"), |
| ("bédu", "water"), |
| ("akala", "water"), |
| ("onoman", "name"), |
| ("ke", "and"), |
| ("eti", "and"), |
| ("ab", "near"), |
| ("nōrikon", "skin"), |
| ("ew", "good"), |
| ("bago", "good"), |
| ("waso", "good"), |
| ] |
|
|
|
|
| def get_avestan_swadesh_entries() -> list[dict]: |
| """Return hardcoded Avestan Swadesh entries.""" |
| return [{"word": w, "gloss": g} for w, g in AVESTAN_SWADESH] |
|
|
|
|
| def get_old_persian_swadesh_entries() -> list[dict]: |
| """Return hardcoded Old Persian Swadesh + Reconstruction entries.""" |
| entries = [{"word": w, "gloss": g} for w, g in OLD_PERSIAN_SWADESH] |
| entries += [{"word": w, "gloss": g} for w, g in OLD_PERSIAN_RECONSTRUCTION] |
| return entries |
|
|
|
|
| def get_phrygian_swadesh_entries() -> list[dict]: |
| """Return hardcoded Phrygian Swadesh + additional lemma entries.""" |
| entries = [{"word": w, "gloss": g} for w, g in PHRYGIAN_SWADESH] |
| entries += [{"word": w, "gloss": g} for w, g in PHRYGIAN_ADDITIONAL] |
| return entries |
|
|
|
|
| |
| |
| |
|
|
| def fetch_avestan_word_romanizations(page_ids: list[int]) -> list[dict]: |
| """Fetch individual Avestan word pages and extract romanizations. |
| |
| Each page has pattern: Avestan_script • ( romanization ) with gloss below. |
| No artificial limit — fetches ALL pages in the category. |
| """ |
| entries = [] |
| fetched = 0 |
|
|
| for pid in page_ids: |
| try: |
| title, text = fetch_page_by_id(pid) |
| if not text: |
| continue |
|
|
| |
| |
| |
| romans = re.findall(r'•\s*\(\s*([a-zA-Zāēīōūəąęðθšžŋɣβγñδ\u0300-\u036f\u0323\u0331\u0325ᵛ\s\-]+?)\s*\)', text) |
|
|
| if not romans: |
| |
| romans = re.findall( |
| r'(?:Avestan|Etymology|Noun|Verb|Adjective|Numeral|Pronoun|Adverb|Conjunction|Preposition)' |
| r'.*?\(\s*([a-zA-Zāēīōūəąęðθšžŋɣβγñδ\u0300-\u036f\u0323\u0331\u0325ᵛ\s\-]+?)\s*\)', |
| text[:1500] |
| ) |
|
|
| |
| def_match = re.search(r'(?:Noun|Verb|Adjective|Numeral|Pronoun|Adverb)\s.*?(\d+\.\s*)?([a-zA-Z][a-zA-Z\s,;]+?)(?:\.|$)', text[:2000]) |
| gloss = "" |
| if def_match: |
| gloss = def_match.group(2).strip()[:50] if def_match.group(2) else "" |
|
|
| for roman in romans: |
| roman_clean = re.sub(r'\s+', '', roman.strip()) |
| if roman_clean and len(roman_clean) >= 1 and len(roman_clean) <= 25: |
| |
| if any(skip in roman_clean.lower() for skip in ['old', 'young', 'avestan', 'edit']): |
| continue |
| entries.append({ |
| "word": roman_clean, |
| "gloss": gloss, |
| }) |
|
|
| fetched += 1 |
| if fetched % 20 == 0: |
| print(f" Fetched {fetched} Avestan pages...", file=sys.stderr) |
| time.sleep(3) |
| else: |
| time.sleep(1.5) |
|
|
| except Exception as e: |
| print(f" Warning: failed to fetch page {pid}: {e}", file=sys.stderr) |
| continue |
|
|
| return entries |
|
|
|
|
| |
| |
| |
|
|
| def extract_old_persian_reconstruction_words(members: list[dict]) -> list[dict]: |
| """Extract romanized words from Reconstruction:Old Persian/ page titles.""" |
| entries = [] |
| for m in members: |
| if m.get("ns") != 118: |
| continue |
| title = m.get("title", "") |
| |
| parts = title.split("/") |
| if len(parts) < 2: |
| continue |
| word = parts[-1].strip() |
| if not word: |
| continue |
| |
| if " " in word: |
| continue |
| |
| is_proper = word[0].isupper() |
| entries.append({ |
| "word": word.lower() if is_proper else word, |
| "gloss": "proper_noun" if is_proper else "-", |
| "is_proper": is_proper, |
| }) |
| return entries |
|
|
|
|
| |
| |
| |
|
|
| def fetch_phrygian_lemma_romanizations(members: list[dict]) -> list[dict]: |
| """Fetch individual Phrygian lemma pages to extract romanized forms. |
| |
| No artificial limit — fetches ALL main-namespace pages in the category. |
| """ |
| entries = [] |
| fetched = 0 |
|
|
| for m in members: |
| if m.get("ns") != 0: |
| continue |
|
|
| title = m["title"] |
|
|
| try: |
| text = fetch_rendered_page(title) |
| if not text: |
| continue |
|
|
| |
| if "Phrygian" not in text: |
| continue |
|
|
| |
| |
| romans = re.findall(r'•\s*\(\s*([a-zA-Zāēīōūə\-]+?)\s*\)', text[:1500]) |
|
|
| if not romans: |
| |
| |
| if all(c.isascii() or c in 'āēīōū' for c in title): |
| romans = [title] |
|
|
| |
| def_match = re.search(r'(?:Noun|Verb|Adjective|Numeral|Pronoun|Adverb|Conjunction|Preposition|Determiner)\s.*?(\d+\.\s*)?([a-zA-Z][a-zA-Z\s,;]+?)(?:\.|$)', text[:2000]) |
| gloss = "" |
| if def_match: |
| gloss = def_match.group(2).strip()[:50] if def_match.group(2) else "" |
|
|
| for roman in romans: |
| roman_clean = re.sub(r'\s+', '', roman.strip()) |
| if roman_clean and len(roman_clean) >= 1: |
| entries.append({ |
| "word": roman_clean, |
| "gloss": gloss, |
| }) |
|
|
| fetched += 1 |
| time.sleep(1.5) |
|
|
| except Exception as e: |
| print(f" Warning: failed to fetch Phrygian page '{title}': {e}", file=sys.stderr) |
| time.sleep(2) |
| continue |
|
|
| return entries |
|
|
|
|
| |
| |
| |
|
|
| |
| _JUNK_WORDS = { |
| "cosmicandmoral", "deverbaladjectivesuffix", "transliterationneeded", |
| "zoroastrianism", "swiftly", "-chested", "key", "antara", |
| "adultmale", "asaknife", "asinabed", "output", |
| } |
|
|
| |
| _MIN_WORD_LEN = {"ave": 2, "peo": 2, "xpg": 2} |
|
|
|
|
| def _is_valid_word(word: str, iso: str) -> bool: |
| """Check if a word looks like a valid lexical entry (not junk).""" |
| |
| min_len = _MIN_WORD_LEN.get(iso, 2) |
| if len(word) < min_len: |
| return False |
| |
| if word.lower() in _JUNK_WORDS: |
| return False |
| |
| if word.isascii() and word[0].isupper() and len(word) > 3: |
| return False |
| |
| if " " in word: |
| return False |
| |
| if len(word) > 25: |
| return False |
| |
| if word.startswith("-") and len(word) > 5: |
| return False |
| return True |
|
|
|
|
| def write_tsv(entries: list[dict], iso: str, output_dir: Path) -> Path: |
| """Write entries to TSV, applying transliteration->IPA->SCA pipeline. |
| |
| Each entry dict has: word (romanized), gloss (optional). |
| """ |
| output_dir.mkdir(parents=True, exist_ok=True) |
| output_path = output_dir / f"{iso}.tsv" |
|
|
| processed = [] |
| seen = set() |
| errors = [] |
|
|
| for entry in entries: |
| word = entry["word"].strip() |
| if not word: |
| continue |
|
|
| |
| if word in seen: |
| continue |
| seen.add(word) |
|
|
| |
| if not _is_valid_word(word, iso): |
| continue |
|
|
| |
| try: |
| ipa = transliterate(word, iso) |
| except Exception as e: |
| errors.append(f"Transliteration error for '{word}': {e}") |
| ipa = word |
|
|
| if not ipa: |
| ipa = word |
|
|
| |
| try: |
| sca = ipa_to_sound_class(ipa) |
| except Exception as e: |
| errors.append(f"SCA error for '{word}' (IPA: {ipa}): {e}") |
| sca = "" |
|
|
| gloss = entry.get("gloss", "").strip() |
| |
| if gloss.lower() in ("output", "edit", "-", ""): |
| gloss = "" |
| elif any(junk in gloss.lower() for junk in [ |
| "encyclopedia", "iranica", "online", "university", "leiden", |
| "edit this", "height haraiti", "reckoning", "following the other", |
| ]): |
| gloss = "" |
| elif len(gloss) > 40: |
| |
| gloss = "" |
| concept_id = gloss.lower().replace(" ", "_")[:50] if gloss else "-" |
|
|
| processed.append((word, ipa, sca, "wiktionary", concept_id, "-")) |
|
|
| |
| processed.sort(key=lambda x: x[0].lower()) |
|
|
| with open(output_path, "w", encoding="utf-8", newline="") as f: |
| f.write("Word\tIPA\tSCA\tSource\tConcept_ID\tCognate_Set_ID\n") |
| for word, ipa, sca, source, concept_id, cognate_set_id in processed: |
| f.write(f"{word}\t{ipa}\t{sca}\t{source}\t{concept_id}\t{cognate_set_id}\n") |
|
|
| return output_path, len(processed), errors |
|
|
|
|
| |
| |
| |
|
|
| def main(): |
| output_dir = ROOT / "data" / "training" / "lexicons" |
|
|
| print("=" * 60) |
| print("Extracting Avestan (ave), Old Persian (peo), Phrygian (xpg)") |
| print("Source: Wiktionary Swadesh lists + category lemmas") |
| print("=" * 60) |
|
|
| |
| |
| |
| print("\n--- AVESTAN (ave) ---") |
|
|
| |
| ave_entries = get_avestan_swadesh_entries() |
| print(f" Swadesh list entries: {len(ave_entries)}") |
|
|
| |
| print(" Fetching Avestan category members...") |
| try: |
| ave_members = fetch_category_members("Category:Avestan_lemmas") |
| ave_main_ids = [m["pageid"] for m in ave_members if m.get("ns") == 0] |
| print(f" Category members: {len(ave_main_ids)} main-namespace pages") |
|
|
| |
| print(" Fetching individual Avestan word pages for romanization...") |
| ave_page_entries = fetch_avestan_word_romanizations(ave_main_ids) |
| print(f" Additional entries from pages: {len(ave_page_entries)}") |
| except Exception as e: |
| print(f" Warning: could not fetch category members: {e}", file=sys.stderr) |
| ave_page_entries = [] |
|
|
| all_ave = ave_entries + ave_page_entries |
|
|
| ave_path, ave_count, ave_errors = write_tsv(all_ave, "ave", output_dir) |
| print(f" Written: {ave_count} entries to {ave_path}") |
| if ave_errors: |
| print(f" Errors: {len(ave_errors)}") |
| for e in ave_errors[:5]: |
| print(f" {e}") |
|
|
| |
| |
| |
| print("\n--- OLD PERSIAN (peo) ---") |
|
|
| |
| peo_swadesh_entries = get_old_persian_swadesh_entries() |
| print(f" Swadesh list entries: {len(peo_swadesh_entries)}") |
|
|
| |
| print(" Fetching Old Persian category members...") |
| try: |
| peo_members = fetch_category_members("Category:Old_Persian_lemmas") |
| peo_recon_entries = extract_old_persian_reconstruction_words(peo_members) |
| print(f" Reconstruction entries: {len(peo_recon_entries)}") |
| except Exception as e: |
| print(f" Warning: could not fetch category members: {e}", file=sys.stderr) |
| peo_recon_entries = [] |
|
|
| all_peo = peo_swadesh_entries + peo_recon_entries |
|
|
| peo_path, peo_count, peo_errors = write_tsv(all_peo, "peo", output_dir) |
| print(f" Written: {peo_count} entries to {peo_path}") |
| if peo_errors: |
| print(f" Errors: {len(peo_errors)}") |
| for e in peo_errors[:5]: |
| print(f" {e}") |
|
|
| |
| |
| |
| print("\n--- PHRYGIAN (xpg) ---") |
|
|
| |
| xpg_swadesh_entries = get_phrygian_swadesh_entries() |
| print(f" Swadesh list entries: {len(xpg_swadesh_entries)}") |
|
|
| |
| print(" Fetching Phrygian category members...") |
| try: |
| xpg_members = fetch_category_members("Category:Phrygian_lemmas") |
| xpg_main = [m for m in xpg_members if m.get("ns") == 0] |
| print(f" Category members: {len(xpg_main)} main-namespace pages") |
|
|
| |
| print(" Fetching individual Phrygian word pages...") |
| xpg_page_entries = fetch_phrygian_lemma_romanizations(xpg_members) |
| print(f" Additional entries from pages: {len(xpg_page_entries)}") |
| except Exception as e: |
| print(f" Warning: could not fetch category members: {e}", file=sys.stderr) |
| xpg_page_entries = [] |
|
|
| all_xpg = xpg_swadesh_entries + xpg_page_entries |
|
|
| xpg_path, xpg_count, xpg_errors = write_tsv(all_xpg, "xpg", output_dir) |
| print(f" Written: {xpg_count} entries to {xpg_path}") |
| if xpg_errors: |
| print(f" Errors: {len(xpg_errors)}") |
| for e in xpg_errors[:5]: |
| print(f" {e}") |
|
|
| |
| |
| |
| print("\n" + "=" * 60) |
| print("SUMMARY") |
| print("=" * 60) |
| print(f" Avestan (ave): {ave_count} entries -> {ave_path}") |
| print(f" Old Persian (peo): {peo_count} entries -> {peo_path}") |
| print(f" Phrygian (xpg): {xpg_count} entries -> {xpg_path}") |
|
|
| total_errors = len(ave_errors) + len(peo_errors) + len(xpg_errors) |
| if total_errors: |
| print(f" Total errors: {total_errors}") |
| else: |
| print(" No errors.") |
| print("=" * 60) |
|
|
|
|
| if __name__ == "__main__": |
| main() |
|
|