| | # Ancient Scripts Datasets — Master Database Reference
|
| |
|
| | > **Last updated:** 2026-03-14 | **Total entries:** 3,466,000+ across 1,178 languages
|
| |
|
| | This document is the single source of truth for understanding, modifying, and extending this database. It is designed for both human researchers and AI agents.
|
| |
|
| | ---
|
| |
|
| | ## Table of Contents
|
| |
|
| | 1. [Database Overview](#1-database-overview)
|
| | 2. [TSV Schema & Format](#2-tsv-schema--format)
|
| | 3. [Ancient Languages — Complete Registry](#3-ancient-languages--complete-registry)
|
| | 4. [Non-Ancient Languages — Summary](#4-non-ancient-languages--summary)
|
| | 5. [Source Registry](#5-source-registry)
|
| | 6. [IPA & Phonetic Processing Pipeline](#6-ipa--phonetic-processing-pipeline)
|
| | 7. [Transliteration Maps System](#7-transliteration-maps-system)
|
| | 8. [Sound Class (SCA) System](#8-sound-class-sca-system)
|
| | 9. [Scripts & Data Flow](#9-scripts--data-flow)
|
| | 10. [PRD: Adding New Data](#10-prd-adding-new-data)
|
| | 11. [PRD: Adding New Languages](#11-prd-adding-new-languages)
|
| | 12. [Data Acquisition Rules (Iron Law)](#12-data-acquisition-rules-iron-law)
|
| | 13. [Adversarial Review Protocol](#13-adversarial-review-protocol)
|
| | 14. [Re-processing & Cleaning Runbook](#14-re-processing--cleaning-runbook)
|
| | 15. [Known Limitations & Future Work](#15-known-limitations--future-work)
|
| |
|
| | ---
|
| |
|
| | ## 1. Database Overview
|
| |
|
| | ### Locations
|
| |
|
| | | Location | Path / URL | What |
|
| | |----------|-----------|------|
|
| | | **HuggingFace dataset** | `https://huggingface.co/datasets/PhaistosLabs/ancient-scripts-datasets` | **PRIMARY cloud copy.** All lexicons, cognate pairs, metadata, sources, scripts, docs. Push here after any data change. |
|
| | | **HuggingFace local clone** | `C:\Users\alvin\hf-ancient-scripts\` | Local clone of HuggingFace repo. Use `huggingface_hub` API or `git push` to sync. |
|
| | | **GitHub repo** | `https://github.com/Nacryos/ancient-scripts-datasets.git` | Scripts, docs, pipeline code. Lexicon data is gitignored but committed via force-add for some ancient langs. |
|
| | | **Local working copy** | `C:\Users\alvin\ancient-scripts-datasets\` | Full repo + generated data + CLDF sources |
|
| | | **CLDF sources** | `sources/` (593 MB) | **Gitignored.** Cloned separately: `northeuralex`, `ids`, `abvd`, `wold`, `sinotibetan`, `wikipron` |
|
| | | **Total local footprint** | 2.2 GB | Includes all generated data + CLDF source repos |
|
| |
|
| | ### What IS Tracked in Git (GitHub)
|
| |
|
| | - `scripts/` — All extraction and processing scripts
|
| | - `cognate_pipeline/` — Python package for phonetic processing
|
| | - `docs/` — PRDs, audit reports, this reference doc
|
| | - `data/training/metadata/` — `languages.tsv`, `source_stats.tsv`, `phylo_pairs.tsv` (small summary/lookup files)
|
| | - `data/training/validation/` — Validation sets (via Git LFS)
|
| | - `data/training/lexicons/*.tsv` — Ancient language TSVs (force-added despite gitignore)
|
| |
|
| | ### What is NOT Tracked in Git (gitignored)
|
| |
|
| | - `data/training/lexicons/` — Modern language TSVs (1,113 files, regenerated from scripts)
|
| | - `data/training/cognate_pairs/` — Cognate pair datasets (regenerated)
|
| | - `sources/` — CLDF source repositories (cloned separately, ~593 MB)
|
| |
|
| | ### What IS on HuggingFace (everything)
|
| |
|
| | **HuggingFace is the single source of truth for ALL data files.** It contains:
|
| | - All 1,136 lexicon TSVs (ancient + modern)
|
| | - All cognate pair datasets
|
| | - All metadata files
|
| | - All scripts, docs, and pipeline code
|
| | - All CLDF source repos (2,928 files in `sources/`)
|
| | - Raw audit trails and intermediate extraction files
|
| |
|
| | ### HuggingFace Push Rules
|
| |
|
| | 1. **After any data change** (new entries, IPA reprocessing, map fixes): push updated TSVs to HF
|
| | 2. **After any script change** that affects output: push scripts to HF
|
| | 3. **Use `huggingface_hub` API** for individual file uploads:
|
| | ```python
|
| | from huggingface_hub import HfApi
|
| | api = HfApi()
|
| | api.upload_file(
|
| | path_or_fileobj="data/training/lexicons/ave.tsv",
|
| | path_in_repo="data/training/lexicons/ave.tsv",
|
| | repo_id="PhaistosLabs/ancient-scripts-datasets",
|
| | repo_type="dataset",
|
| | commit_message="fix: reprocess Avestan IPA with expanded transliteration map"
|
| | )
|
| | ```
|
| | 4. **For bulk uploads** (many files): use `upload_large_folder()` from the HF local clone at `C:\Users\alvin\hf-ancient-scripts\`
|
| | 5. **Always push to BOTH** GitHub (scripts/docs) and HuggingFace (data + scripts/docs)
|
| | 6. **Never let HF fall behind** — if data exists locally but not on HF, it's not deployed
|
| |
|
| | **To reconstruct all data from scratch:**
|
| | ```bash
|
| | # 1. Clone CLDF sources
|
| | git clone https://github.com/lexibank/northeuralex sources/northeuralex
|
| | git clone https://github.com/lexibank/ids sources/ids
|
| | git clone https://github.com/lexibank/abvd sources/abvd
|
| | git clone https://github.com/lexibank/wold sources/wold
|
| | git clone https://github.com/lexibank/sinotibetan sources/sinotibetan
|
| | # WikiPron: download from https://github.com/CUNY-CL/wikipron
|
| |
|
| | # 2. Run extraction pipeline
|
| | python scripts/expand_cldf_full.py # Modern languages from CLDF
|
| | python scripts/ingest_wikipron.py # WikiPron IPA data
|
| | python scripts/run_lexicon_expansion.py # Ancient language extraction (requires internet)
|
| | python scripts/reprocess_ipa.py # Apply transliteration maps
|
| | python scripts/assemble_lexicons.py # Generate metadata
|
| | ```
|
| |
|
| | ### Directory Structure
|
| |
|
| | ```
|
| | ancient-scripts-datasets/
|
| | data/training/
|
| | lexicons/ # 1,136 TSV files (one per language) [GITIGNORED]
|
| | metadata/ # languages.tsv, source_stats.tsv, phylo_pairs.tsv [TRACKED]
|
| | cognate_pairs/ # inherited, similarity, borrowing pairs [GITIGNORED]
|
| | validation/ # stratified ML training/test sets [GIT LFS]
|
| | language_profiles/ # per-language markdown profiles
|
| | raw/ # raw JSON audit trails
|
| | audit_trails/ # JSONL provenance logs
|
| | scripts/ # 23 extraction scripts + 7 parsers [TRACKED]
|
| | cognate_pipeline/ # Python package for phonetic processing [TRACKED]
|
| | docs/ # PRDs, audit reports, this file [TRACKED]
|
| | sources/ # CLDF repos [GITIGNORED, clone separately]
|
| | ```
|
| |
|
| | **Scale:**
|
| | - 1,178 languages (68 ancient/reconstructed + 1,113 modern — 3 overlap)
|
| | - 3,466,000+ total lexical entries
|
| | - 170,756 ancient language entries (68 languages)
|
| | - 3,296,156 modern language entries (1,113 languages)
|
| | - 23,375,899 cognate/borrowing/similarity pairs
|
| |
|
| | ### Cognate Pairs (v2)
|
| |
|
| | Three TSV files in `data/training/cognate_pairs/`, 14-column schema:
|
| |
|
| | ```
|
| | Lang_A Word_A IPA_A Lang_B Word_B IPA_B Concept_ID Relationship Score Source Relation_Detail Donor_Language Confidence Source_Record_ID
|
| | ```
|
| |
|
| | | File | Rows | Description |
|
| | |------|------|-------------|
|
| | | `cognate_pairs_inherited.tsv` | 22,893,519 | Expert-classified cognates ONLY (ABVD + ACD + IE-CoR + Sino-Tibetan) |
|
| | | `cognate_pairs_borrowing.tsv` | 17,147 | Verified donor→recipient borrowings from WOLD BorrowingTable |
|
| | | `cognate_pairs_similarity.tsv` | 465,233 | Algorithmic phonetic similarity: concept_aligned (219,519, score ≥ 0.5) + similarity_only (245,714, 0.3 ≤ score < 0.5) |
|
| |
|
| | **Sources & provenance:**
|
| | - **ABVD** CognateTable (20,697,739 pairs, 1,682 Austronesian languages) — Expert-classified by Greenhill, Blust & Gray (2008). Sister-sister relationships within cognate sets. `loan_flagged` for 160,768 pairs where ABVD Loan column non-empty.
|
| | - **ACD** Cognacy groups (1,825,701 pairs, 727 Austronesian languages) — Expert-classified by Blust, Trussel & Smith (2023), DOI: 10.5281/zenodo.7737547. Cross-language pairs within cognacy groups. Score = -1 for all (no reliable IPA; pseudo-IPA from orthography). `loan_flagged` for 573 pairs where ACD Loan column is True.
|
| | - **IE-CoR** CognateTable (365,913 pairs, 159 Indo-European languages) — Expert-classified by Heggarty et al. (2023, Science). Mixed PIE-level (32%) and branch-level (55%) cognacy. `loan_involved` for 3,598 pairs where cognateset appears in loans.csv.
|
| | - **Sino-Tibetan** CognateTable (4,166 pairs) — Expert-classified from STEDT-derived data (Jacques & List). Borrowings (146 entries) pre-filtered at extraction.
|
| | - **WOLD** BorrowingTable (17,147 pairs) — Expert borrowing judgments from Haspelmath & Tadmor (2009). `borrowed_immediate` (13,967) vs `borrowed_earlier` (3,180). Score = -1 sentinel for pseudo-IPA entries. Entries with "no evidence for borrowing" (802) filtered out.
|
| | - **Concept-aligned** (219,519 pairs) + **Similarity-only** (245,714 pairs) — Algorithmically generated via SCA phonetic comparison within same language family. NOT expert cognates. Isolate and constructed languages excluded.
|
| |
|
| | **Deduplication:** Priority ordering expert_cognate > borrowing > concept_aligned > similarity_only. Pass 1.5 pre-populates expert language-concept keys across ALL files before writing, ensuring no concept_aligned/similarity pair duplicates an expert pair. See `docs/prd/PRD_COGNATE_PAIRS_V2.md` for full specification.
|
| |
|
| | **Adversarial audit status (2026-03-14):** All 3 output files PASS final audit. Zero cross-file contamination, zero self-pairs, zero isolate/constructed language leakage, all Source_Record_IDs traceable to source databases.
|
| |
|
| | ### Phylogenetic Relationship Metadata
|
| |
|
| | **File:** `data/training/metadata/phylo_pairs.tsv` (386,101 unique language pairs)
|
| |
|
| | A lookup table mapping every unique `(Lang_A, Lang_B)` pair in the cognate dataset to its phylogenetic relationship, based on Glottolog CLDF v5.x (Hammarström et al.). Not stored inline in the 23M-row cognate files to avoid redundancy.
|
| |
|
| | **Schema (9 columns, tab-separated):**
|
| |
|
| | | Column | Type | Description |
|
| | |--------|------|-------------|
|
| | | `Lang_A` | str | ISO 639-3 code (alphabetically first) |
|
| | | `Lang_B` | str | ISO 639-3 code (alphabetically second) |
|
| | | `Phylo_Relation` | enum | `near_ancestral`, `close_sister`, `distant_sister`, `cross_family`, `unclassified` |
|
| | | `Tree_Distance` | int | Edge count through MRCA (99 = unclassified/cross-family) |
|
| | | `MRCA_Clade` | str | Glottocode of MRCA node |
|
| | | `MRCA_Depth` | int | Depth of MRCA in tree (0 = root) |
|
| | | `Ancestor_Lang` | str | For `near_ancestral`: ISO of the ancestor; `-` otherwise |
|
| | | `Family_A` | str | Top-level Glottolog family of Lang_A |
|
| | | `Family_B` | str | Top-level Glottolog family of Lang_B |
|
| |
|
| | **Distribution:**
|
| |
|
| | | Relation | Count | Percentage |
|
| | |----------|-------|------------|
|
| | | `distant_sister` | 249,392 | 64.6% |
|
| | | `close_sister` | 87,078 | 22.6% |
|
| | | `cross_family` | 45,267 | 11.7% |
|
| | | `unclassified` | 4,302 | 1.1% |
|
| | | `near_ancestral` | 62 | 0.0% |
|
| |
|
| | **Usage:** Join at query time using `pair_key = (min(a,b), max(a,b))`. The classification is orthogonal to the cognate data and can be updated independently when Glottolog releases new versions.
|
| |
|
| | **Scripts:** `scripts/ingest_glottolog.py` (download), `scripts/build_glottolog_tree.py` (parse), `scripts/build_phylo_pairs.py` (classify), `scripts/validate_phylo_pairs.py` (validate). See `docs/prd/PRD_PHYLO_ENRICHMENT.md` for full specification.
|
| |
|
| | **Validation (2026-03-14):** 45,363/45,363 tests passed. 13 known-answer checks, 62 near-ancestral integrity, 45,267 cross-family integrity, 99.4% ISO coverage, 20/20 random audit.
|
| |
|
| | ---
|
| |
|
| | ## 2. TSV Schema & Format
|
| |
|
| | Every lexicon file follows this 6-column tab-separated schema:
|
| |
|
| | ```
|
| | Word IPA SCA Source Concept_ID Cognate_Set_ID
|
| | ```
|
| |
|
| | | Column | Description | Example |
|
| | |--------|-------------|---------|
|
| | | **Word** | Orthographic/transliterated form | `pahhur`, `*wódr̥`, `𐬀𐬵𐬎𐬭𐬀` |
|
| | | **IPA** | Broad phonemic IPA transcription | `paxːur`, `wodr̩`, `ahura` |
|
| | | **SCA** | Sound Class Alphabet encoding (18C + 5V) | `PAKUR`, `WOTR`, `AHURA` |
|
| | | **Source** | Data provenance identifier | `wiktionary`, `ediana`, `wikipron` |
|
| | | **Concept_ID** | Semantic concept (first gloss word, snake_case) | `fire`, `water`, `-` |
|
| | | **Cognate_Set_ID** | Cognate grouping identifier | `PIE_fire_001`, `-` |
|
| |
|
| | **Rules:**
|
| | - Header row MUST be present as line 1
|
| | - UTF-8 encoding, Unix line endings preferred
|
| | - No empty IPA fields — use Word as fallback if no conversion possible
|
| | - Source field must accurately reflect actual data origin
|
| | - `-` for unknown/unavailable fields
|
| |
|
| | ---
|
| |
|
| | ## 3. Ancient Languages — Complete Registry
|
| |
|
| | ### Entry Counts & IPA Quality (as of 2026-03-12)
|
| |
|
| | | # | Language | ISO | Family | Entries | Identity% | Top Sources | IPA Type |
|
| | |---|----------|-----|--------|---------|-----------|-------------|----------|
|
| | | 1 | Avestan | ave | Indo-Iranian | 3,455 | 14.4% | avesta_org (2,716), wiktionary_cat (384), wiktionary (355) | Broad phonemic (Hoffmann & Forssman) |
|
| | | 2 | Tocharian B | txb | Indo-European | 2,386 | 25.2% | wiktionary_cat (2,386) | Broad phonemic (Tocharian map) |
|
| | | 3 | Luwian | xlw | Anatolian | 2,230 | 26.2% | ediana (1,985), palaeolexicon (225) | Broad phonemic (Luwian map) |
|
| | | 4 | Proto-Indo-European | ine-pro | Indo-European | 1,704 | 0.2% | wiktionary_cat (863), wiktionary (841) | Broad phonemic (reconstructed) |
|
| | | 5 | Lycian | xlc | Anatolian | 1,098 | 36.7% | ediana (517), palaeolexicon (482) | Broad phonemic (Melchert 2004) |
|
| | | 6 | Etruscan | ett | Tyrsenian | 753 | 25.5% | palaeolexicon (503), wikipron (207) | Broad phonemic (Bonfante) |
|
| | | 7 | Urartian | xur | Hurro-Urartian | 748 | 54.4% | oracc_ecut (704), wiktionary (44) | Partial (cuneiform sign names) |
|
| | | 8 | Lydian | xld | Anatolian | 693 | 53.0% | ediana (447), palaeolexicon (187) | Broad phonemic (Gusmani 1964) |
|
| | | 9 | Carian | xcr | Anatolian | 532 | 39.7% | palaeolexicon (304), ediana (174) | Broad phonemic (Adiego 2007) |
|
| | | 10 | Proto-Kartvelian | ccs-pro | Kartvelian | 504 | 22.2% | wiktionary (254), wiktionary_cat (250) | Broad phonemic (Klimov 1998) |
|
| | | 11 | Old Persian | peo | Indo-Iranian | 486 | 10.5% | wiktionary (244), wiktionary_cat (242) | Broad phonemic (Kent 1953) |
|
| | | 12 | Tocharian A | xto | Indo-European | 467 | 23.1% | wiktionary_cat (467) | Broad phonemic (Tocharian map) |
|
| | | 13 | Proto-Dravidian | dra-pro | Dravidian | 406 | 7.1% | wiktionary_cat (235), wiktionary (171) | Broad phonemic (Krishnamurti) |
|
| | | 14 | Proto-Semitic | sem-pro | Afroasiatic | 386 | 26.9% | wiktionary_cat (247), wiktionary (139) | Broad phonemic (Huehnergard) |
|
| | | 15 | Ugaritic | uga | Afroasiatic | 371 | 15.6% | wiktionary (344), wiktionary_cat (27) | Broad phonemic (Tropper 2000) |
|
| | | 16 | Hittite | hit | Anatolian | 266 | 20.3% | wiktionary (266) | Broad phonemic (Hoffner & Melchert) |
|
| | | 17 | Hurrian | xhu | Hurro-Urartian | 260 | 50.4% | palaeolexicon (259) | Broad phonemic (Wegner 2007) |
|
| | | 18 | Elamite | elx | Isolate | 301 | 71.1% | wiktionary (301) | Minimal (transparent orthography) |
|
| | | 19 | Rhaetic | xrr | Tyrsenian | 187 | 55.1% | tir_raetica (142), wiktionary (45) | Partial (North Italic alphabet) |
|
| | | 20 | Phoenician | phn | Afroasiatic | 180 | 18.3% | wiktionary (180) | Broad phonemic (abjad reconstruction) |
|
| | | 21 | Phrygian | xpg | Indo-European | 79 | 36.7% | wiktionary (79) | Partial (small corpus, Greek-script support) |
|
| | | 22 | Messapic | cms | Indo-European | 45 | 88.9% | wiktionary (45) | Minimal (Greek-alphabet, mostly identity) |
|
| | | 23 | Lemnian | xle | Tyrsenian | 30 | 53.3% | wiktionary (30) | Minimal (very small corpus) |
|
| | | | | | | | | | |
|
| | | **--- Tier 2 (Phase 6) ---** | | | | | | | |
|
| | | 24 | Old English | ang | Germanic | 31,319 | 10.5% | wiktionary_cat (31,319) | Broad phonemic (Hogg 1992) |
|
| | | 25 | Biblical Hebrew | hbo | Semitic | 12,182 | 0.1% | wiktionary_cat (12,182) | Broad phonemic (Blau 2010) |
|
| | | 26 | Coptic | cop | Egyptian | 11,180 | 0.1% | wiktionary_cat (7,987), kellia (3,193) | Broad phonemic (Layton 2000) |
|
| | | 27 | Old Armenian | xcl | Indo-European | 6,277 | 0.0% | wiktionary_cat (6,277) | Broad phonemic (Meillet 1913) |
|
| | | 28 | Pali | pli | Indo-Aryan | 2,792 | 19.1% | wiktionary_cat (2,792) | Broad phonemic (Geiger 1943) |
|
| | | 29 | Ge'ez | gez | Semitic | 496 | 0.0% | wiktionary_cat (496) | Broad phonemic (Dillmann 1857) |
|
| | | 30 | Hattic | xht | Isolate | 269 | 37.9% | wiktionary_cat (269) | Partial (cuneiformist conventions) |
|
| | | | | | | | | | |
|
| | | **--- Tier 3 (Phase 7) ---** | | | | | | | |
|
| | | 31 | Old Irish | sga | Celtic | 41,300 | 39.4% | edil (40,309), wiktionary_cat (991) | Broad phonemic (Thurneysen) |
|
| | | 32 | Old Japanese | ojp | Japonic | 5,393 | 59.7% | oncoj (4,974), wiktionary_cat (419) | Broad phonemic (Frellesvig 2010) |
|
| | | 33 | Classical Nahuatl | nci | Uto-Aztecan | 3,873 | 5.7% | wiktionary_cat (3,873) | Broad phonemic |
|
| | | 34 | Oscan | osc | Italic | 2,122 | 15.1% | ceipom (2,122) | Broad phonemic (CEIPoM Standard_aligned) |
|
| | | 35 | Umbrian | xum | Italic | 1,631 | 3.7% | ceipom (1,631) | Broad phonemic (CEIPoM Standard_aligned) |
|
| | | 36 | Venetic | xve | Italic | 721 | 86.5% | ceipom (721) | Minimal (Latin transliteration) |
|
| | | 37 | Gaulish | xtg | Celtic | 271 | 92.3% | diacl (183), wiktionary_cat (88) | Minimal (Latin transliteration) |
|
| | | 38 | Middle Persian | pal | Indo-Iranian | 242 | 62.8% | wiktionary_cat (242) | Broad phonemic (MacKenzie 1971) |
|
| | | 39 | Sogdian | sog | Indo-Iranian | 194 | 37.1% | iecor (161), wiktionary_cat (33) | Broad phonemic (Gharib 1995) |
|
| | | | | | | | | | |
|
| | | **--- Proto-Languages (Phase 7) ---** | | | | | | | |
|
| | | 40 | Proto-Austronesian | map | Austronesian | 11,624 | 41.1% | acd (11,624) | Broad phonemic (Blust notation) |
|
| | | 41 | Proto-Germanic | gem-pro | Germanic | 5,399 | 32.9% | wiktionary_cat (5,399) | Broad phonemic (reconstructed) |
|
| | | 42 | Proto-Celtic | cel-pro | Celtic | 1,584 | 68.3% | wiktionary_cat (1,584) | Partial (mixed Latin/IPA) |
|
| | | 43 | Proto-Uralic | urj-pro | Uralic | 585 | 50.3% | wiktionary_cat (585) | Broad phonemic (Sammallahti 1988) |
|
| | | 44 | Proto-Bantu | bnt-pro | Niger-Congo | 467 | 54.0% | wiktionary_cat (467) | Broad phonemic (BLR notation) |
|
| | | 45 | Proto-Sino-Tibetan | sit-pro | Sino-Tibetan | 358 | 100.0% | wiktionary_cat (358) | Already IPA (Wiktionary provides IPA) |
|
| | | | | | | | | | |
|
| | | **--- Phase 8 Batch 1 (Proto-Languages + Italic/Celtic) ---** | | | | | | | |
|
| | | 46 | Proto-Slavic | sla-pro | Balto-Slavic | 5,068 | 18.4% | wiktionary_cat (5,068) | Broad phonemic (reconstructed) |
|
| | | 47 | Proto-Turkic | trk-pro | Turkic | 1,027 | 27.8% | wiktionary_cat (1,027) | Broad phonemic (reconstructed) |
|
| | | 48 | Proto-Italic | itc-pro | Italic | 739 | 46.7% | wiktionary_cat (739) | Broad phonemic (reconstructed) |
|
| | | 49 | Faliscan | xfa | Italic | 566 | 67.1% | ceipom (566) | Partial (CEIPoM Standard_aligned) |
|
| | | 50 | Proto-Japonic | jpx-pro | Japonic | 426 | 70.2% | wiktionary_cat (426) | Partial (mixed notation) |
|
| | | 51 | Lepontic | xlp | Celtic | 421 | 27.6% | lexlep (421) | Broad phonemic (Lexicon Leponticum) |
|
| | | 52 | Proto-Iranian | ira-pro | Indo-Iranian | 366 | 4.6% | wiktionary_cat (366) | Broad phonemic (reconstructed) |
|
| | | 53 | Ancient South Arabian | xsa | Semitic | 127 | 25.2% | wiktionary (127) | Broad phonemic (Musnad abjad) |
|
| | | 54 | Celtiberian | xce | Celtic | 11 | 100.0% | wiktionary_cat (11) | Minimal (very small corpus) |
|
| | | | | | | | | | |
|
| | | **--- Phase 8 Batch 2 (Proto-Languages + Ancient) ---** | | | | | | | |
|
| | | 55 | Meroitic | xmr | Nilo-Saharan | 1,978 | 39.8% | meroitic-corpus (1,978) | Broad phonemic (Rilly 2007) |
|
| | | 56 | Proto-Algonquian | alg-pro | Algic | 258 | 28.7% | wiktionary_cat (258) | Broad phonemic (reconstructed) |
|
| | | 57 | Proto-Albanian | sqj-pro | Albanian | 210 | 43.8% | wiktionary_cat (210) | Broad phonemic (reconstructed) |
|
| | | 58 | Proto-Austroasiatic | aav-pro | Austroasiatic | 180 | 100.0% | wiktionary_cat (180) | Already IPA (Wiktionary provides IPA) |
|
| | | 59 | Proto-Polynesian | poz-pol-pro | Austronesian | 157 | 100.0% | wiktionary_cat (157) | Already IPA (Wiktionary provides IPA) |
|
| | | 60 | Proto-Tai | tai-pro | Kra-Dai | 148 | 0.7% | wiktionary_cat (148) | Broad phonemic (Li 1977) |
|
| | | 61 | Proto-Tocharian | xto-pro | Tocharian | 138 | 22.5% | wiktionary_cat (138) | Broad phonemic (reconstructed) |
|
| | | 62 | Proto-Mongolic | xgn-pro | Mongolic | 126 | 41.3% | wiktionary_cat (126) | Broad phonemic (reconstructed) |
|
| | | 63 | Proto-Oceanic | poz-oce-pro | Austronesian | 114 | 92.1% | wiktionary_cat (114) | Minimal (transparent orthography) |
|
| | | 64 | Moabite | obm | Semitic | 31 | 0.0% | wiktionary_cat (31) | Broad phonemic (Canaanite abjad) |
|
| | | | | | | | | | |
|
| | | **--- Phase 8 Batch 3 (Proto-Languages + Iberian) ---** | | | | | | | |
|
| | | 65 | Proto-Mayan | myn-pro | Mayan | 65 | 20.0% | wiktionary_cat (65) | Broad phonemic (Kaufman 2003) |
|
| | | 66 | Proto-Afroasiatic | afa-pro | Afroasiatic | 48 | 54.2% | wiktionary_cat (48) | Broad phonemic (Ehret 1995) |
|
| | | 67 | Iberian | xib | Isolate | 39 | 74.4% | wiktionary_cat (39) | Partial (undeciphered script) |
|
| | | | | | | | | | |
|
| | | **--- Phase 8 Eblaite ---** | | | | | | | |
|
| | | 68 | Eblaite | xeb | Semitic | 667 | 0.3% | dcclt-ebla (667) | Broad phonemic (Krebernik 1982) |
|
| |
|
| | **Total ancient + classical: 170,756 entries across 68 languages | Overall identity rate: ~30%**
|
| |
|
| | ### Understanding Identity Rate
|
| |
|
| | **Identity rate = % of entries where Word == IPA** (no phonetic conversion applied).
|
| |
|
| | | Rate | Meaning | Example Languages |
|
| | |------|---------|-------------------|
|
| | | <10% | Excellent IPA conversion | ine-pro (0.2%), dra-pro (7.1%) |
|
| | | 10-30% | Good conversion | peo (10.5%), ave (14.4%), hit (20.3%), ccs-pro (22.2%), txb (25.2%) |
|
| | | 30-50% | Moderate — some chars unmapped | xlc (36.7%), xcr (39.7%), xhu (50.4%) |
|
| | | 50-70% | Partial — significant gaps | xld (53.0%), xur (54.4%), elx (71.1%) |
|
| | | >70% | Minimal — mostly passthrough | cms (88.9%) |
|
| |
|
| | **Causes of high identity:**
|
| | - **Cuneiform sign notation** (xur): Uppercase Sumerograms like `LUGAL`, `URU` aren't phonemic — 156 entries in xur
|
| | - **Already-IPA characters** (cms): Some scripts use characters that ARE IPA (θ, ə, ŋ)
|
| | - **Transparent orthography** (elx): Latin letters already map 1:1 to IPA
|
| | - **eDiAna pre-transliterated forms** (xlc, xld): Source provides Latin transliterations that are already close to IPA
|
| | - **Plain ASCII stems** (txb, xto): Short roots like `ak`, `aik` are valid in both orthography and IPA
|
| |
|
| | ### IPA Quality Categories
|
| |
|
| | | Category | Definition | Ancient Languages |
|
| | |----------|-----------|-------------------|
|
| | | **FULL** | >80% WikiPron-sourced IPA | (none — ancient langs don't have WikiPron) |
|
| | | **BROAD PHONEMIC** | Scholarly transliteration → IPA via cited map | hit, uga, phn, ave, peo, ine-pro, sem-pro, ccs-pro, dra-pro, xlw, xhu, ett, txb, xto, xld, xcr, xpg |
|
| | | **PARTIAL** | Some chars converted, gaps remain | xlc, xrr |
|
| | | **MINIMAL** | Mostly identity / transparent orthography | elx, xle, cms |
|
| | | **CUNEIFORM MIXED** | Mix of converted transliterations + unconverted sign names | xur |
|
| |
|
| | **Important:** For dead languages, **broad phonemic is the ceiling**. Narrow allophonic IPA is not possible because allophonic variation is unrecoverable from written records. The IPA column represents the best scholarly reconstruction of phonemic values, not actual pronunciation.
|
| |
|
| | ---
|
| |
|
| | ## 4. Non-Ancient Languages — Summary
|
| |
|
| | - **1,113 languages** with 3,296,156 entries
|
| | - **Dominant source:** WikiPron (85.3% of entries = 2,822,808)
|
| | - **Other sources:** ABVD (6.7%), NorthEuraLex (5.7%), WOLD (1.8%), sinotibetan (0.1%)
|
| |
|
| | **WikiPron entries** have true broad phonemic IPA (scraped from Wiktionary pronunciation sections by trained linguists). These are the gold standard.
|
| |
|
| | **ABVD entries** are often orthographic (Word == IPA). The `fix_abvd_ipa.py` script applies rule-based G2P conversion for Austronesian languages.
|
| |
|
| | ---
|
| |
|
| | ## 5. Source Registry
|
| |
|
| | | Source ID | Full Name | Type | URL | Languages Covered |
|
| | |-----------|-----------|------|-----|-------------------|
|
| | | `wikipron` | WikiPron Pronunciation Dictionary | Scraped IPA | `sources/wikipron/` (local) | 800+ modern languages |
|
| | | `abvd` | Austronesian Basic Vocabulary Database | CLDF | `sources/abvd/` (local) | 500+ Austronesian |
|
| | | `northeuralex` | NorthEuraLex | CLDF | `sources/northeuralex/` (local) | 100+ Eurasian |
|
| | | `wold` | World Loanword Database | CLDF | `sources/wold/` (local) | 40+ worldwide |
|
| | | `sinotibetan` | Sino-Tibetan Etymological Database | CLDF | `sources/sinotibetan/` (local) | 50+ Sino-Tibetan |
|
| | | `wiktionary` | Wiktionary (appendix/lemma pages) | Web scrape | `en.wiktionary.org` | All ancient langs |
|
| | | `wiktionary_cat` | Wiktionary (category pagination) | MediaWiki API | `en.wiktionary.org/w/api.php` | ine-pro, uga, peo, ave, dra-pro, sem-pro, ccs-pro, txb, xto |
|
| | | `ediana` | eDiAna (LMU Munich) | POST API | `ediana.gwi.uni-muenchen.de` | xlc, xld, xcr, xlw |
|
| | | `palaeolexicon` | Palaeolexicon | REST API | `palaeolexicon.com/api/Search/` | xlc, xld, xcr, xlw, xhu, ett |
|
| | | `oracc_ecut` | Oracc eCUT (Urartian texts) | JSON API | `oracc.museum.upenn.edu/ecut/` | xur |
|
| | | `tir_raetica` | TIR (Thesaurus Inscriptionum Raeticarum) | Web scrape | `tir.univie.ac.at` | xrr |
|
| | | `wikipedia` | Wikipedia vocabulary tables | Web scrape | `en.wikipedia.org` | xur (supplement) |
|
| | | `avesta_org` | Avesta.org Avestan Dictionary | Web scrape | `avesta.org/avdict/avdict.htm` | ave |
|
| | | `kaikki` | Kaikki Wiktionary Dump | JSON dump | `kaikki.org` | Various |
|
| | | `kellia` | Kellia Coptic Lexicon | XML | `data.copticscriptorium.org` | cop |
|
| | | `ceipom` | CEIPoM (Italian Epigraphy) | CSV | `zenodo.org` (CC BY-SA 4.0) | osc, xum, xve |
|
| | | `edil` | eDIL (Electronic Dict of Irish Lang) | XML | `github.com/e-dil/dil` | sga |
|
| | | `acd` | ACD (Austronesian Comparative Dict) | CLDF | `github.com/lexibank/acd` (CC BY 4.0) | map |
|
| | | `oncoj` | ONCOJ (Oxford-NINJAL OJ Corpus) | XML | `github.com/ONCOJ/data` (CC BY 4.0) | ojp |
|
| | | `diacl` | DiACL (Diachronic Atlas of Comp Ling) | CLDF | `github.com/lexibank/diacl` (CC BY 4.0) | xtg |
|
| | | `iecor` | IE-CoR (IE Cognate Relationships) | CLDF | `github.com/lexibank/iecor` (CC BY 4.0) | sog |
|
| | | `lexlep` | Lexicon Leponticum (Zurich) | Web/CSV | `lexlep.univie.ac.at` | xlp |
|
| | | `meroitic-corpus` | Meroitic Language Corpus (GitHub) | JSON/CSV | `github.com/MeroiticLanguage/Meroitic-Corpus` | xmr |
|
| | | `dcclt-ebla` | DCCLT/Ebla (ORACC) | JSON ZIP | `oracc.museum.upenn.edu/dcclt-ebla/` (CC0) | xeb |
|
| |
|
| | ---
|
| |
|
| | ## 6. IPA & Phonetic Processing Pipeline
|
| |
|
| | ### Pipeline Architecture
|
| |
|
| | ```
|
| | Source Data (Word column)
|
| | ↓
|
| | transliterate(word, iso) ← scripts/transliteration_maps.py
|
| | ↓ (greedy longest-match, NFC-normalized)
|
| | IPA string (broad phonemic)
|
| | ↓
|
| | ipa_to_sound_class(ipa) ← cognate_pipeline/.../sound_class.py
|
| | ↓ (tokenize → segment_to_class → join)
|
| | SCA string (e.g., "PATA")
|
| | ```
|
| |
|
| | ### IPA Generation Methods (by source type)
|
| |
|
| | | Source | IPA Method | Quality |
|
| | |--------|-----------|---------|
|
| | | WikiPron | Pre-extracted from Wiktionary pronunciation | True broad IPA |
|
| | | Wiktionary (ancient) | `transliterate(word, iso)` via language-specific map | Broad phonemic |
|
| | | ABVD | Orthographic passthrough → `fix_abvd_ipa.py` G2P | Variable |
|
| | | eDiAna | `transliterate(word, iso)` | Broad phonemic |
|
| | | Palaeolexicon | Source IPA if available, else `transliterate()` | Broad phonemic |
|
| | | Oracc | `transliterate(word, iso)` | Partial (cuneiform) |
|
| | | NorthEuraLex/WOLD | CLDF Segments column → joined IPA | Good |
|
| |
|
| | ### Never-Regress Re-processing Rule
|
| |
|
| | When re-applying transliteration maps to existing data (`scripts/reprocess_ipa.py`):
|
| |
|
| | ```python
|
| | candidate_ipa = transliterate(word, iso)
|
| |
|
| | if candidate_ipa != word:
|
| | final_ipa = candidate_ipa # New map converts — use it
|
| | elif old_ipa != word:
|
| | final_ipa = old_ipa # New map can't, but old was good — keep
|
| | else:
|
| | final_ipa = word # Both identity — nothing to do
|
| | ```
|
| |
|
| | **This ensures:** IPA quality can only improve or stay the same. It never regresses.
|
| |
|
| | ---
|
| |
|
| | ## 7. Transliteration Maps System
|
| |
|
| | **File:** `scripts/transliteration_maps.py` (~800 lines)
|
| |
|
| | ### How It Works
|
| |
|
| | Each ancient language has a `Dict[str, str]` mapping scholarly transliteration conventions to broad IPA. The `transliterate()` function applies these via **greedy longest-match**: keys sorted by descending length, first match consumed at each position.
|
| |
|
| | ### Map Registry (updated 2026-03-13 — 180+ new rules across 13 original maps + 15 new maps in Phases 6-7 + 24 new maps in Phase 8)
|
| |
|
| | | ISO | Language | Keys | Academic Reference |
|
| | |-----|----------|------|--------------------|
|
| | | `hit` | Hittite | 49 | Hoffner & Melchert (2008) — added š, ḫ, macron vowels |
|
| | | `uga` | Ugaritic | 68 | Tropper (2000) — added ʾ, macron/circumflex vowels, ḫ, ṣ, Ugaritic script (U+10380-1039F) |
|
| | | `phn` | Phoenician | 23 | Standard 22-letter abjad |
|
| | | `xur` | Urartian | 27 | Wegner (2007) — added ṣ, ṭ, y, w, ə, ʾ |
|
| | | `elx` | Elamite | 19 | Grillot-Susini (1987), Stolper (2004) |
|
| | | `xlc` | Lycian | 33 | Melchert (2004) — added x, j, o, long vowels |
|
| | | `xld` | Lydian | 38 | Gusmani (1964), Melchert — added ã, ẽ, ũ (nasalized vowels), c, h, z, x |
|
| | | `xcr` | Carian | 35 | Adiego (2007) — added β, z, v, j, f, ŋ, ĺ, ỳ, ý |
|
| | | `ave` | Avestan | 97 | Hoffmann & Forssman (1996) + Unicode 5.2 (U+10B00-10B3F) |
|
| | | `peo` | Old Persian | 68 | Kent (1953) — added z, č, Old Persian cuneiform syllabary (U+103A0-103C3, 31 signs) |
|
| | | `ine` | Proto-Indo-European | 61 | Fortson (2010), Beekes (2011) — added ḗ, ṓ, morpheme boundaries, accented syllabic sonorants |
|
| | | `sem` | Proto-Semitic | 44 | Huehnergard (2019) |
|
| | | `ccs` | Proto-Kartvelian | 66 | Klimov (1998) — added s₁/z₁/c₁/ʒ₁ subscript series, morpheme boundaries |
|
| | | `dra` | Proto-Dravidian | 49 | Krishnamurti (2003) |
|
| | | `xpg` | Phrygian | 55 | Brixhe & Lejeune (1984), Obrador-Cursach (2020) — added Greek alphabet support (22 letters) |
|
| | | `xle` | Lemnian | 24 | Greek-alphabet reconstruction |
|
| | | `xrr` | Rhaetic | 26 | North Italic alphabet reconstruction |
|
| | | `cms` | Messapic | 25 | Greek-alphabet reconstruction |
|
| | | `xlw` | Luwian | 39 | Melchert (2003), Yakubovich (2010) |
|
| | | `xhu` | Hurrian | 31 | Wegner (2007), Wilhelm (2008) |
|
| | | `ett` | Etruscan | 61 | Bonfante & Bonfante (2002), Rix (1963) + Old Italic Unicode — added z, o, d, g, b, q, σ→s |
|
| | | `txb`/`xto` | Tocharian A/B | 35 | Krause & Thomas (1960), Adams (2013), Peyrot (2008) — added retroflex series (ṭ, ḍ, ṇ, ḷ) |
|
| | | | | | |
|
| | | **--- Phase 6: Tier 2 Maps ---** | | | |
|
| | | `cop` | Coptic | 40+ | Layton (2000), Loprieno (1995) — Sahidic dialect |
|
| | | `pli` | Pali (IAST) | 30+ | Geiger (1943), Oberlies (2001) |
|
| | | `xcl` | Old Armenian | 40+ | Meillet (1913), Schmitt (1981) |
|
| | | `ang` | Old English | 30+ | Hogg (1992), Campbell (1959) |
|
| | | `gez` | Ge'ez (Ethiopic) | 50+ | Dillmann (1857), Tropper (2002) |
|
| | | `hbo` | Biblical Hebrew | 40+ | Blau (2010), Khan (2020) |
|
| | | | | | |
|
| | | **--- Phase 7: Tier 3 + Proto Maps ---** | | | |
|
| | | `osc` | Oscan | 12 | CEIPoM Standard_aligned conventions |
|
| | | `xum` | Umbrian | 12 | CEIPoM Standard_aligned conventions |
|
| | | `xve` | Venetic | 6 | CEIPoM Token_clean conventions |
|
| | | `sga` | Old Irish | 25 | Thurneysen (1946), Stifter (2006) — lenition + macron vowels |
|
| | | `xeb` | Eblaite | 20 | Standard Semitist notation |
|
| | | `nci` | Classical Nahuatl | 15 | Andrews (2003), Launey (2011) |
|
| | | `ojp` | Old Japanese | 20 | Frellesvig (2010), ONCOJ conventions |
|
| | | `pal` | Middle Persian | 25 | MacKenzie (1971), Skjærvø (2009) |
|
| | | `sog` | Sogdian | 25 | Gharib (1995), Sims-Williams (2000) |
|
| | | `xtg` | Gaulish | 15 | Delamarre (2003) |
|
| | | `gem-pro` | Proto-Germanic | 20 | Ringe (2006), Kroonen (2013) |
|
| | | `cel-pro` | Proto-Celtic | 15 | Matasović (2009) |
|
| | | `urj-pro` | Proto-Uralic | 12 | Sammallahti (1988), Janhunen (1981) |
|
| | | `bnt-pro` | Proto-Bantu | 20 | Bastin et al. (2002), Meeussen (1967) |
|
| | | `sit-pro` | Proto-Sino-Tibetan | 18 | Matisoff (2003), Sagart (2004) |
|
| | | | | | |
|
| | | **--- Phase 8 Maps ---** | | | |
|
| | | `sla-pro` | Proto-Slavic | 25+ | Shevelov (1964), Holzer (2007) |
|
| | | `trk-pro` | Proto-Turkic | 20+ | Clauson (1972), Róna-Tas (1991) |
|
| | | `itc-pro` | Proto-Italic | 15+ | Meiser (1998), Bakkum (2009) |
|
| | | `jpx-pro` | Proto-Japonic | 15+ | Vovin (2005), Frellesvig (2010) |
|
| | | `ira-pro` | Proto-Iranian | 20+ | Cheung (2007), Lubotsky (2001) |
|
| | | `xfa` | Faliscan | 12 | CEIPoM Standard_aligned conventions |
|
| | | `xlp` | Lepontic | 25 | Lexicon Leponticum (Stifter et al.) |
|
| | | `xce` | Celtiberian | 15+ | De Bernardo Stempel (1999) |
|
| | | `xsa` | Ancient South Arabian | 30+ | Stein (2003), Beeston (1984) |
|
| | | `alg-pro` | Proto-Algonquian | 15+ | Bloomfield (1946), Goddard (1994) |
|
| | | `sqj-pro` | Proto-Albanian | 15+ | Orel (1998), Demiraj (1997) |
|
| | | `aav-pro` | Proto-Austroasiatic | 10+ | Shorto (2006), Sidwell (2015) |
|
| | | `poz-pol-pro` | Proto-Polynesian | 10+ | Biggs (1978), Pawley (1966) |
|
| | | `tai-pro` | Proto-Tai | 20+ | Li (1977), Pittayaporn (2009) |
|
| | | `xto-pro` | Proto-Tocharian | 15+ | Adams (2013), Peyrot (2008) |
|
| | | `poz-oce-pro` | Proto-Oceanic | 10+ | Ross et al. (1998, 2003, 2008) |
|
| | | `xgn-pro` | Proto-Mongolic | 15+ | Poppe (1955), Nugteren (2011) |
|
| | | `xmr` | Meroitic | 30+ | Rilly (2007), Griffith (1911) |
|
| | | `obm` | Moabite | 22 | Canaanite abjad (shares Phoenician map base) |
|
| | | `myn-pro` | Proto-Mayan | 20+ | Kaufman (2003), Campbell & Kaufman (1985) |
|
| | | `afa-pro` | Proto-Afroasiatic | 15+ | Ehret (1995), Orel & Stolbova (1995) |
|
| | | `xib` | Iberian | 25+ | De Hoz (2010), Untermann (1990) |
|
| | | `xeb` | Eblaite | 20+ | Krebernik (1982), Fronzaroli (2003) |
|
| |
|
| | ### NFC Normalization
|
| |
|
| | All map keys and input text are NFC-normalized before comparison. This ensures `š` (U+0161, composed) matches `s` + combining caron (U+0073 + U+030C, decomposed). Cache is per-ISO to prevent cross-language leakage.
|
| |
|
| | ### ISO Code Mapping for Proto-Languages
|
| |
|
| | TSV filenames use hyphenated codes but `ALL_MAPS` uses short codes:
|
| |
|
| | | TSV filename ISO | Map ISO |
|
| | |-----------------|---------|
|
| | | `ine-pro` | `ine` |
|
| | | `sem-pro` | `sem` |
|
| | | `ccs-pro` | `ccs` |
|
| | | `dra-pro` | `dra` |
|
| | | `gem-pro` | `gem-pro` |
|
| | | `cel-pro` | `cel-pro` |
|
| | | `urj-pro` | `urj-pro` |
|
| | | `bnt-pro` | `bnt-pro` |
|
| | | `sit-pro` | `sit-pro` |
|
| |
|
| | ### Adding a New Map
|
| |
|
| | 1. Add the `Dict[str, str]` constant (e.g., `NEW_LANG_MAP`) with cited reference
|
| | 2. Register in `ALL_MAPS`: `"iso_code": NEW_LANG_MAP`
|
| | 3. Clear `_nfc_cache` implicitly (happens on next call with new ISO)
|
| | 4. Run `reprocess_ipa.py --language iso_code` to apply
|
| | 5. Deploy adversarial auditor to verify
|
| |
|
| | ---
|
| |
|
| | ## 8. Sound Class (SCA) System
|
| |
|
| | **File:** `cognate_pipeline/src/cognate_pipeline/normalise/sound_class.py`
|
| |
|
| | ### Class Inventory
|
| |
|
| | | Class | IPA Segments | Description |
|
| | |-------|-------------|-------------|
|
| | | A | a, ɑ, æ, ɐ | Open vowels |
|
| | | E | e, ɛ, ə, ɘ, ø, œ | Mid vowels |
|
| | | I | i, ɪ, ɨ | Close front vowels |
|
| | | O | o, ɔ, ɵ | Mid back vowels |
|
| | | U | u, ʊ, ʉ, ɯ, y | Close back vowels |
|
| | | P/B | p, b, ɸ, β | Labial stops |
|
| | | T/D | t, d, ʈ, ɖ | Coronal stops |
|
| | | K/G | k, g, ɡ, q, ɢ | Dorsal stops |
|
| | | S | s, z, ʃ, ʒ, ɕ, ʑ, f, v, θ, ð, x, ɣ, χ, ts, dz, tʃ, dʒ | Fricatives + affricates |
|
| | | M/N | m, n, ɲ, ŋ, ɳ, ɴ | Nasals |
|
| | | L/R | l, ɫ, ɭ, ɬ, r, ɾ, ɽ, ʀ, ɹ, ʁ | Liquids |
|
| | | W/Y | w, ʋ, ɰ, j | Glides |
|
| | | H | ʔ, h, ɦ, ʕ, ħ | Glottals/pharyngeals |
|
| | | 0 | (anything unmapped) | Unknown |
|
| |
|
| | ### Processing Chain
|
| |
|
| | ```python
|
| | ipa_to_sound_class("paxːur")
|
| | → tokenize_ipa("paxːur") → ["p", "a", "xː", "u", "r"]
|
| | → [segment_to_class(s) for s in segments] → ["P", "A", "K", "U", "R"]
|
| | → "PAKUR"
|
| | ```
|
| |
|
| | ---
|
| |
|
| | ## 9. Scripts & Data Flow
|
| |
|
| | ### Data Flow Diagram
|
| |
|
| | ```
|
| | EXTERNAL SOURCES
|
| | ├── Wiktionary API ──────────→ extract_ave_peo_xpg.py
|
| | │ extract_phn_elx.py
|
| | │ extract_pie_urartian.py
|
| | │ extract_wiktionary_lexicons.py
|
| | │ expand_wiktionary_categories.py
|
| | │ expand_xpg.py
|
| | ├── eDiAna API ──────────────→ scrape_ediana.py
|
| | ├── Palaeolexicon API ───────→ scrape_palaeolexicon.py
|
| | ├── Oracc JSON API ──────────→ scrape_oracc_urartian.py
|
| | ├── avesta.org ──────────────→ scrape_avesta_org.py
|
| | ├── TIR (Vienna) ────────────→ scrape_tir_rhaetic.py
|
| | ├── WikiPron TSVs ───────────→ ingest_wikipron.py
|
| | └── CLDF Sources ────────────→ expand_cldf_full.py
|
| | convert_cldf_to_tsv.py
|
| | ↓
|
| | data/training/lexicons/{iso}.tsv
|
| | ↓
|
| | normalize_lexicons.py (NFC, dedup, strip stress)
|
| | reprocess_ipa.py (re-apply updated transliteration maps)
|
| | fix_abvd_ipa.py (Austronesian G2P fix)
|
| | ↓
|
| | assemble_lexicons.py → metadata/languages.tsv
|
| | assign_cognate_links.py → cognate_pairs/*.tsv
|
| | build_validation_sets.py → validation/*.tsv
|
| | ```
|
| |
|
| | ### Script Quick Reference
|
| |
|
| | | Script | Purpose | Languages |
|
| | |--------|---------|-----------|
|
| | | `extract_ave_peo_xpg.py` | Wiktionary Swadesh + category | ave, peo, xpg |
|
| | | `extract_phn_elx.py` | Wiktionary + appendix | phn, elx |
|
| | | `extract_pie_urartian.py` | Wiktionary + Wikipedia | ine-pro, xur |
|
| | | `extract_wiktionary_lexicons.py` | Wiktionary appendix | sem-pro, ccs-pro, dra-pro, xle |
|
| | | `extract_anatolian_lexicons.py` | Multi-source | xlc, xld, xcr |
|
| | | `expand_wiktionary_categories.py` | Wiktionary category pagination | ine-pro, uga, peo, ave, dra-pro, sem-pro, ccs-pro |
|
| | | `expand_xpg.py` | Wiktionary category + appendix | xpg |
|
| | | `scrape_ediana.py` | eDiAna POST API | xlc, xld, xcr, xlw |
|
| | | `scrape_palaeolexicon.py` | Palaeolexicon REST API | xlc, xld, xcr, xlw, xhu, ett |
|
| | | `scrape_avesta.py` | avesta.org (old, superseded) | ave |
|
| | | `scrape_avesta_org.py` | avesta.org dictionary (current, adversarial-audited) | ave |
|
| | | `scrape_oracc_urartian.py` | Oracc eCUT JSON API | xur |
|
| | | `scrape_tir_rhaetic.py` | TIR web scrape | xrr |
|
| | | `ingest_wikipron.py` | WikiPron TSV ingestion | 800+ modern |
|
| | | `expand_cldf_full.py` | CLDF full extraction | All CLDF languages |
|
| | | `reprocess_ipa.py` | Re-apply transliteration maps | 23 ancient |
|
| | | `fix_abvd_ipa.py` | G2P for Austronesian | ABVD languages |
|
| | | `normalize_lexicons.py` | NFC + dedup + SCA recompute | All |
|
| | | `assemble_lexicons.py` | Generate metadata | All |
|
| | | `ingest_wiktionary_tier2.py` | Wiktionary category ingestion (Tier 2+) | Phase 6-8 Wiktionary languages |
|
| | | `fetch_wiktionary_raw.py` | Fetch raw Wiktionary category JSON | Phase 6-8 Wiktionary languages |
|
| | | `ingest_dcclt_ebla.py` | ORACC DCCLT/Ebla extraction | xeb |
|
| | | `ingest_meroitic.py` | Meroitic Language Corpus | xmr |
|
| | | `ingest_lexlep.py` | Lexicon Leponticum extraction | xlp |
|
| | | `ingest_ceipom_italic.py` | CEIPoM italic epigraphy | osc, xum, xve, xfa |
|
| | | `update_metadata.py` | Update languages.tsv from disk | All |
|
| | | `validate_all.py` | Comprehensive TSV validation | All |
|
| | | `push_to_hf.py` | Push files to HuggingFace | All Phase 6-8 |
|
| |
|
| | ---
|
| |
|
| | ## 10. PRD: Adding New Data to Existing Languages
|
| |
|
| | ### Prerequisites
|
| |
|
| | - The language already has a TSV file in `data/training/lexicons/`
|
| | - You have identified a new external source with verifiable data
|
| | - A transliteration map exists in `transliteration_maps.py` (if ancient)
|
| |
|
| | ### Step-by-Step
|
| |
|
| | #### Step 1: Identify Source
|
| | - Find a publicly accessible online source (API, web page, database)
|
| | - Verify it returns real lexical data (not AI-generated)
|
| | - Document the URL, API format, and expected entry count
|
| |
|
| | #### Step 2: Write Extraction Script
|
| | ```python
|
| | # Template: scripts/scrape_{source}_{iso}.py
|
| | #!/usr/bin/env python3
|
| | """Scrape {Source Name} for {Language} word lists.
|
| | Source: {URL}
|
| | """
|
| | import urllib.request # MANDATORY — proves data comes from HTTP
|
| | ...
|
| |
|
| | def fetch_data(url):
|
| | """Fetch from external source."""
|
| | req = urllib.request.Request(url, headers={"User-Agent": "..."})
|
| | with urllib.request.urlopen(req) as resp:
|
| | return json.loads(resp.read())
|
| |
|
| | def process_language(iso, config, dry_run=False):
|
| | """Process and deduplicate."""
|
| | existing = load_existing_words(tsv_path) # MUST deduplicate
|
| | entries = fetch_data(url)
|
| | new_entries = [e for e in entries if e["word"] not in existing]
|
| | ...
|
| | # Apply transliteration
|
| | ipa = transliterate(word, iso)
|
| | sca = ipa_to_sound_class(ipa)
|
| | f.write(f"{word}\t{ipa}\t{sca}\t{source_id}\t{concept_id}\t-\n")
|
| | ```
|
| |
|
| | **Critical:** Script MUST contain `urllib.request.urlopen()`, `requests.get()`, or equivalent HTTP fetch. No hardcoded word lists.
|
| |
|
| | #### Step 3: Run with --dry-run
|
| | ```bash
|
| | python scripts/scrape_new_source.py --dry-run --language {iso}
|
| | ```
|
| |
|
| | #### Step 4: Run Live
|
| | ```bash
|
| | python scripts/scrape_new_source.py --language {iso}
|
| | ```
|
| |
|
| | #### Step 5: Re-process IPA (if map was updated)
|
| | ```bash
|
| | python scripts/reprocess_ipa.py --language {iso}
|
| | ```
|
| |
|
| | #### Step 6: Deploy Adversarial Auditor
|
| | See [Section 13](#13-adversarial-review-protocol).
|
| |
|
| | #### Step 7: Commit & Push to Both Repos
|
| | ```bash
|
| | # GitHub
|
| | git add scripts/scrape_new_source.py data/training/lexicons/{iso}.tsv
|
| | git commit -m "Add {N} entries to {Language} from {Source}"
|
| | git push
|
| |
|
| | # HuggingFace (MANDATORY — HF is the primary data host)
|
| | python -c "
|
| | from huggingface_hub import HfApi
|
| | api = HfApi()
|
| | for f in ['data/training/lexicons/{iso}.tsv', 'scripts/scrape_new_source.py']:
|
| | api.upload_file(path_or_fileobj=f, path_in_repo=f,
|
| | repo_id='PhaistosLabs/ancient-scripts-datasets', repo_type='dataset',
|
| | commit_message='Add {N} entries to {Language} from {Source}')
|
| | "
|
| | ```
|
| |
|
| | ---
|
| |
|
| | ## 11. PRD: Adding New Languages
|
| |
|
| | ### Prerequisites
|
| |
|
| | - ISO 639-3 code identified
|
| | - At least one external source with verifiable word lists
|
| | - Script conventions for the relevant writing system understood
|
| |
|
| | ### Step-by-Step
|
| |
|
| | #### Step 1: Create Transliteration Map (if needed)
|
| |
|
| | Add to `scripts/transliteration_maps.py`:
|
| |
|
| | ```python
|
| | # ---------------------------------------------------------------------------
|
| | # N. NEW_LANGUAGE (Author Year, "Title")
|
| | # ---------------------------------------------------------------------------
|
| | NEW_LANGUAGE_MAP: Dict[str, str] = {
|
| | "a": "a", "b": "b", ...
|
| | # Every key MUST have a cited academic reference
|
| | }
|
| | ```
|
| |
|
| | Register in `ALL_MAPS`:
|
| | ```python
|
| | ALL_MAPS = {
|
| | ...
|
| | "new_iso": NEW_LANGUAGE_MAP,
|
| | }
|
| | ```
|
| |
|
| | #### Step 2: Write Extraction Script
|
| |
|
| | Follow the template in [Section 10](#10-prd-adding-new-data). The script must:
|
| | - Fetch from an external source via HTTP
|
| | - Parse the response (HTML, JSON, XML)
|
| | - Apply `transliterate()` and `ipa_to_sound_class()`
|
| | - Write to `data/training/lexicons/{iso}.tsv`
|
| | - Save raw JSON to `data/training/raw/` for audit trail
|
| | - Deduplicate by Word column
|
| |
|
| | #### Step 3: Add to Language Config (optional)
|
| |
|
| | If the language will be part of the ancient languages pipeline, add to `scripts/language_configs.py`.
|
| |
|
| | #### Step 4: Add to Re-processing List
|
| |
|
| | Add the ISO code to `ANCIENT_LANGUAGES` in `scripts/reprocess_ipa.py` and to `ISO_TO_MAP_ISO` if the TSV filename differs from the map ISO.
|
| |
|
| | #### Step 5: Run Extraction
|
| | ```bash
|
| | python scripts/scrape_{source}.py --language {iso} --dry-run
|
| | python scripts/scrape_{source}.py --language {iso}
|
| | ```
|
| |
|
| | #### Step 6: Verify
|
| |
|
| | ```bash
|
| | # Check entry count and IPA quality
|
| | python scripts/reprocess_ipa.py --dry-run --language {iso}
|
| | ```
|
| |
|
| | #### Step 7: Deploy Adversarial Auditor
|
| |
|
| | See [Section 13](#13-adversarial-review-protocol).
|
| |
|
| | #### Step 8: Commit and Push
|
| |
|
| | ---
|
| |
|
| | ## 12. Data Acquisition Rules (Iron Law)
|
| |
|
| | ```
|
| | ┌─────────────────────────────────────────────────────────────────────┐
|
| | │ DATA MAY ONLY ENTER THE DATASET THROUGH CODE THAT DOWNLOADS IT │
|
| | │ FROM AN EXTERNAL SOURCE. │
|
| | │ │
|
| | │ NO EXCEPTIONS. NO "JUST THIS ONCE." NO "IT'S FASTER." │
|
| | └─────────────────────────────────────────────────────────────────────┘
|
| | ```
|
| |
|
| | ### What IS Allowed
|
| |
|
| | | Action | Example | Why OK |
|
| | |--------|---------|--------|
|
| | | Write a script with `urllib.request.urlopen()` | `scrape_palaeolexicon.py` | Data comes from HTTP |
|
| | | Parse HTML/JSON from downloaded content | `BeautifulSoup(html)` | Deterministic extraction |
|
| | | Apply transliteration map (CODE, not DATA) | `transliterate(word, "hit")` | Transformation rules are code |
|
| | | Re-compute SCA from IPA | `ipa_to_sound_class(ipa)` | Deterministic function |
|
| |
|
| | ### What is FORBIDDEN
|
| |
|
| | | Action | Example | Why Forbidden |
|
| | |--------|---------|---------------|
|
| | | Write data rows directly | `f.write("water\twɔːtər\t...")` | Data authoring |
|
| | | Hardcode word lists from memory | `WORDS = [("fire", "paxːur")]` | LLM knowledge ≠ source |
|
| | | Fill in missing fields with guesses | `ipa = "probably θ"` | Hallucination risk |
|
| | | Generate translations/transcriptions | `ipa = "wɔːtər" # I know how water sounds` | Not from a source |
|
| | | Pad entries to reach a target count | Adding 13 entries to make it 200 | Fabrication |
|
| |
|
| | ### The Cached-Fetch Pattern (Acceptable Gray Area)
|
| |
|
| | If a source requires JavaScript rendering or CAPTCHAs:
|
| | 1. Use WebFetch/browser to access the source
|
| | 2. Save raw content to `data/training/raw/{source}_{iso}_{date}.html`
|
| | 3. Write a parsing script that reads from the saved file
|
| | 4. The auditor spot-checks 5 entries against the live source
|
| |
|
| | ### Transliteration Maps Are CODE, Not DATA
|
| |
|
| | Transliteration maps (e.g., `"š": "ʃ"`) are **transformation rules** derived from published grammars, not lexical content. Adding or modifying map entries is a code change, not data authoring. However, every map entry MUST cite an academic reference.
|
| |
|
| | ---
|
| |
|
| | ## 13. Adversarial Review Protocol
|
| |
|
| | ### Architecture: Dual-Agent System
|
| |
|
| | ```
|
| | Team A (Extraction Agent) Team B (Adversarial Auditor)
|
| | ├── Writes code ├── Reviews code
|
| | ├── Runs scripts ├── Spot-checks output
|
| | ├── Produces TSV data ├── Verifies provenance
|
| | └── NEVER writes data └── Has VETO POWER
|
| | directly
|
| | ```
|
| |
|
| | ### When to Deploy
|
| |
|
| | - After ANY new data is added to the database
|
| | - After ANY transliteration map change
|
| | - After ANY re-processing run
|
| | - After ANY script modification that affects output
|
| |
|
| | ### Audit Checklist (per modular step)
|
| |
|
| | #### Code Review
|
| | - [ ] Script contains `urllib`/`requests`/`curl` (not hardcoded data)
|
| | - [ ] No literal IPA data in `f.write()` calls
|
| | - [ ] Source attribution matches actual source
|
| | - [ ] Deduplication against existing entries
|
| |
|
| | #### Data Quality
|
| | - [ ] Entry count is non-round and plausible
|
| | - [ ] No duplicate Word values
|
| | - [ ] No empty IPA fields
|
| | - [ ] Identity rate is explainable (not suspiciously low or high)
|
| | - [ ] SCA matches `ipa_to_sound_class(IPA)` for 20 random samples
|
| |
|
| | #### Never-Regress Verification
|
| | - [ ] No entry went from non-identity IPA to identity (regression)
|
| | - [ ] Entry counts did not decrease
|
| | - [ ] Existing Word/Source/Concept_ID/Cognate_Set_ID unchanged
|
| |
|
| | #### Provenance
|
| | - [ ] 20 random entries traced back to source URL
|
| | - [ ] Raw JSON/HTML audit trail saved in `data/training/raw/`
|
| |
|
| | ### Red Flags (STOP immediately)
|
| |
|
| | | Red Flag | What It Means |
|
| | |----------|---------------|
|
| | | No `urllib`/`requests` in extraction code | Agent is authoring data |
|
| | | Entry count is exactly round (100, 200, 500) | Likely padded |
|
| | | >90% of entries have empty required fields | Extraction didn't work |
|
| | | Script contains `f.write("word\tipa\t...")` with literal data | Direct data authoring |
|
| | | Transformation output == input for >80% without cited justification | Map not actually applied |
|
| |
|
| | ### Report Format
|
| |
|
| | ```markdown
|
| | # Adversarial Audit: {Step} — {Language} ({iso})
|
| | ## Checks:
|
| | - [ ] No data authoring: PASS/FAIL
|
| | - [ ] Entry count: PASS/FAIL (expected X, got Y)
|
| | - [ ] IPA quality: PASS/FAIL (identity rate: Z%)
|
| | - [ ] SCA consistency: PASS/FAIL (N/N verified)
|
| | - [ ] Provenance: PASS/FAIL (N/20 traced to source)
|
| | ## Verdict: PASS / WARN / FAIL
|
| | ## Blocking: YES (if FAIL)
|
| | ```
|
| |
|
| | ---
|
| |
|
| | ## 14. Re-processing & Cleaning Runbook
|
| |
|
| | ### When to Re-process
|
| |
|
| | - After modifying any transliteration map in `transliteration_maps.py`
|
| | - After fixing a bug in `transliterate()` or `ipa_to_sound_class()`
|
| | - After adding a new language to `ALL_MAPS`
|
| |
|
| | ### How to Re-process
|
| |
|
| | ```bash
|
| | # Dry run first (ALWAYS)
|
| | python scripts/reprocess_ipa.py --dry-run
|
| |
|
| | # Check: identity rates should decrease or stay the same, NEVER increase
|
| | # Check: "Changed" column shows expected number of modifications
|
| | # Check: "Errors" column is 0
|
| |
|
| | # Run live
|
| | python scripts/reprocess_ipa.py
|
| |
|
| | # Or for a single language
|
| | python scripts/reprocess_ipa.py --language xlw
|
| | ```
|
| |
|
| | ### Common Cleaning Operations
|
| |
|
| | #### Remove entries with HTML artifacts
|
| | ```python
|
| | # Check for HTML entities
|
| | grep -P '&\w+;' data/training/lexicons/{iso}.tsv
|
| | # Remove affected lines via Python script (not manual edit)
|
| | ```
|
| |
|
| | #### Remove entries from wrong source (contamination)
|
| | ```python
|
| | # Example: Hurrian TSV had Hittite entries from wrong Palaeolexicon ID
|
| | # Write a Python script that identifies and removes contaminated entries
|
| | # Save removed entries to audit trail
|
| | ```
|
| |
|
| | #### Deduplicate
|
| | ```python
|
| | # reprocess_ipa.py handles dedup by Word column
|
| | # For more complex dedup, use normalize_lexicons.py
|
| | ```
|
| |
|
| | #### Fix ABVD fake-IPA
|
| | ```bash
|
| | python scripts/fix_abvd_ipa.py
|
| | ```
|
| |
|
| | ### Post-Cleaning Verification
|
| |
|
| | ```bash
|
| | # Verify entry counts
|
| | python -c "
|
| | for iso in ['hit','uga',...]:
|
| | with open(f'data/training/lexicons/{iso}.tsv') as f:
|
| | print(f'{iso}: {sum(1 for _ in f) - 1} entries')
|
| | "
|
| |
|
| | # Verify no empty IPA
|
| | python -c "
|
| | for iso in [...]:
|
| | with open(f'data/training/lexicons/{iso}.tsv') as f:
|
| | for line in f:
|
| | parts = line.strip().split('\t')
|
| | if len(parts) >= 2 and not parts[1]:
|
| | print(f'EMPTY IPA: {iso} {parts[0]}')
|
| | "
|
| | ```
|
| |
|
| | ---
|
| |
|
| | ## 15. Known Limitations & Future Work
|
| |
|
| | ### Linguistic Limitations
|
| |
|
| | | Issue | Languages Affected | Root Cause |
|
| | |-------|-------------------|------------|
|
| | | Broad phonemic only (no allophonic) | All ancient | Dead languages — allophonic variation unrecoverable |
|
| | | Cuneiform sign names as entries | xur, xhu | Source provides sign-level notation, not phonemic. ~156 Sumerograms in xur. |
|
| | | High identity for transparent orthographies | elx, cms, xle | Writing system maps 1:1 to IPA |
|
| | | Old Persian ç → θ debatable | peo | Kent (1953) says /θ/, Kloekhorst (2008) says /ts/ |
|
| | | Old Persian cuneiform inherent vowels | peo | Syllabary signs (𐎣=ka, 𐎫=ta) include inherent vowels that may be redundant in context |
|
| | | eDiAna entries drive high identity | xlc, xld | eDiAna provides already-transliterated forms; identity is expected, not a map gap |
|
| |
|
| | ### Technical Debt
|
| |
|
| | | Issue | Priority | Fix |
|
| | |-------|----------|-----|
|
| | | `use_word_for_ipa` dead config in expand_wiktionary_categories.py | Low | Remove the config key |
|
| | | Some extraction scripts have hardcoded word lists from pre-Iron-Law era | Medium | Rewrite with HTTP fetch |
|
| | | ABVD entries still ~50% fake-IPA after G2P fix | Medium | Better G2P or manual review |
|
| | | NorthEuraLex/WOLD join segments with spaces | Low | Handled by normalize_lexicons.py |
|
| | | Combining diacritics in Lycian/Carian (U+0303, U+0302) | Low | Normalize in preprocessing before transliteration |
|
| | | Greek letter leaks in Carian source data | Low | Data cleaning script to normalize σ→s, α→a, etc. |
|
| | | HTML entities in 4 PIE IPA entries | Low | Decode with `html.unescape()` in reprocess_ipa.py |
|
| | | 15 Old Persian proper nouns have wrong-language IPA | Low | Filter or manually correct Akkadian/Greek transcriptions |
|
| |
|
| | ### Expansion Opportunities
|
| |
|
| | | Language | Current | Available | Source |
|
| | |----------|---------|-----------|--------|
|
| | | Sumerian | 0 | 5,000+ | EPSD2 (ePSD), Oracc |
|
| | | Akkadian | 0 | 10,000+ | CAD, CDA, ePSD2 |
|
| | | Egyptian | 0 | 3,000+ | TLA (Thesaurus Linguae Aegyptiae) |
|
| | | Sanskrit | (modern only) | 50,000+ | Monier-Williams, DCS |
|
| | | Linear B | 0 | 500+ | DAMOS, Wingspread |
|
| | | Luvian Hieroglyphic | (mixed with xlw) | 500+ | Hawkins (2000) |
|
| |
|
| | ---
|
| |
|
| | ## Appendix A: Quick Commands
|
| |
|
| | ```bash
|
| | # Count entries for a language
|
| | wc -l data/training/lexicons/{iso}.tsv
|
| |
|
| | # Check identity rate
|
| | python -c "
|
| | with open('data/training/lexicons/{iso}.tsv') as f:
|
| | lines = f.readlines()[1:]
|
| | total = len(lines)
|
| | identity = sum(1 for l in lines if l.split('\t')[0] == l.split('\t')[1])
|
| | print(f'{identity}/{total} = {identity/total*100:.1f}%')
|
| | "
|
| |
|
| | # Test a transliteration map
|
| | python -c "
|
| | import sys; sys.path.insert(0, 'scripts')
|
| | from transliteration_maps import transliterate
|
| | print(transliterate('test_word', 'iso_code'))
|
| | "
|
| |
|
| | # Re-process single language (dry run)
|
| | python scripts/reprocess_ipa.py --dry-run --language {iso}
|
| |
|
| | # Run adversarial audit (deploy via AI agent)
|
| | # See Section 13 for protocol
|
| | ```
|
| |
|
| | ## Appendix B: File Checksums Reference
|
| |
|
| | Run after any batch operation to create a baseline:
|
| | ```bash
|
| | find data/training/lexicons -name "*.tsv" -exec wc -l {} \; | sort -k2 > /tmp/lexicon_counts.txt
|
| | ```
|
| |
|