File size: 8,927 Bytes
26786e3
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
#!/usr/bin/env python3
"""Parser for DSAL DEDR (Dravidian Etymological Dictionary Revised).

Extracts Proto-Dravidian reconstructions with glosses from the
University of Chicago DSAL digital edition:
    https://dsal.uchicago.edu/dictionaries/burrow/

Entries are organized by entry number and typically contain:
- Proto-Dravidian reconstruction (starred form)
- Cognates across Dravidian languages (Tamil, Telugu, Kannada, etc.)
- English gloss

Uses only stdlib (urllib, html.parser, re).

Reference: Burrow & Emeneau, "A Dravidian Etymological Dictionary" (1984, revised)
"""

from __future__ import annotations

import logging
import re
import urllib.request
import urllib.error
from html.parser import HTMLParser
from typing import Any

logger = logging.getLogger(__name__)


class DEDRPageParser(HTMLParser):
    """Parse DEDR entry pages to extract reconstructions and glosses."""

    def __init__(self) -> None:
        super().__init__()
        self.entries: list[dict] = []
        self.in_entry_div = False
        self.in_bold = False
        self.entry_text = ""
        self.bold_text = ""
        self.current_entry_text = ""
        self._tag_stack: list[str] = []

    def handle_starttag(self, tag: str, attrs: list[tuple[str, str | None]]) -> None:
        attr_dict = dict(attrs)
        cls = attr_dict.get("class", "") or ""

        self._tag_stack.append(tag)

        if any(kw in cls.lower() for kw in ("entry", "hw", "result", "body")):
            self.in_entry_div = True
            self.current_entry_text = ""

        if tag in ("b", "strong"):
            self.in_bold = True
            self.bold_text = ""

    def handle_endtag(self, tag: str) -> None:
        if self._tag_stack and self._tag_stack[-1] == tag:
            self._tag_stack.pop()

        if tag in ("b", "strong") and self.in_bold:
            self.in_bold = False

        if tag in ("div", "p", "br") and self.in_entry_div:
            if self.current_entry_text.strip():
                self._process_entry_text(self.current_entry_text)
            self.current_entry_text = ""

    def handle_data(self, data: str) -> None:
        if self.in_bold:
            self.bold_text += data
        if self.in_entry_div:
            self.current_entry_text += data

    def _process_entry_text(self, text: str) -> None:
        """Process accumulated entry text to extract reconstruction + gloss."""
        text = text.strip()
        if not text or len(text) < 5:
            return

        # Look for reconstruction pattern: *form or starred form
        m = re.search(r"\*([A-Za-zÀ-žḀ-ỿ\-āēīōūṭḍṇṛḷṃṁ]+)", text)
        if m:
            form = m.group(1).strip()
            # Extract gloss: text after the form, typically after comma or dash
            rest = text[m.end():]
            gloss = ""
            # Try various delimiters
            gm = (
                re.match(r"\s*[-–—,]\s+(.+?)(?:\.|;|$)", rest)
                or re.match(r"\s+'(.+?)'", rest)
                or re.match(r'\s+"(.+?)"', rest)
            )
            if gm:
                gloss = gm.group(1).strip()
            if form and gloss and len(form) < 40:
                self.entries.append({
                    "word": form,
                    "transliteration": form,
                    "gloss": gloss,
                })


def _regex_extract(html: str) -> list[dict]:
    """Extract DEDR entries using regex patterns on raw HTML.

    DEDR entries typically look like:
        <b>1234</b> Ta. <i>word</i> meaning. ...
    or:
        1234 *proto-form - meaning
    """
    entries: list[dict] = []

    # Strip HTML tags but keep structure hints
    text = html
    # Replace <br> and <p> with newlines
    text = re.sub(r"<br\s*/?>", "\n", text, flags=re.IGNORECASE)
    text = re.sub(r"</?p[^>]*>", "\n", text, flags=re.IGNORECASE)
    # Strip remaining tags
    text = re.sub(r"<[^>]+>", "", text)
    # Decode entities
    text = text.replace("&amp;", "&")
    text = text.replace("&lt;", "<")
    text = text.replace("&gt;", ">")
    text = text.replace("&nbsp;", " ")

    # Pattern 1: Entry number + reconstruction
    # e.g. "1234 *kāl- leg, foot"
    pattern1 = re.compile(
        r"(\d{1,5})\s+\*([A-Za-zÀ-žḀ-ỿāēīōūṭḍṇṛḷṃṁ\-]+)"
        r"\s*[-–—,]?\s+"
        r"([A-Za-z][A-Za-z\s,'-]{2,80}?)(?=[.;]|\n|$)"
    )
    for m in pattern1.finditer(text):
        entry_num = m.group(1)
        form = m.group(2).strip()
        gloss = m.group(3).strip()
        gloss = re.sub(r"[,;:\s]+$", "", gloss)
        if form and gloss:
            entries.append({
                "word": form,
                "transliteration": form,
                "gloss": gloss,
            })

    # Pattern 2: Tamil/Dravidian language abbreviation + word + meaning
    # e.g. "Ta. kāl foot, leg"
    if not entries:
        pattern2 = re.compile(
            r"(?:Ta|Te|Ka|Ma|Tu|Ko|Go|Ku|Mal|Kur|Br)\.\s+"
            r"([A-Za-zÀ-žḀ-ỿāēīōūṭḍṇṛḷṃṁ\-]+)"
            r"\s+"
            r"([A-Za-z][A-Za-z\s,'-]{2,60}?)(?=[.;]|\n|$)"
        )
        for m in pattern2.finditer(text):
            form = m.group(1).strip()
            gloss = m.group(2).strip()
            gloss = re.sub(r"[,;:\s]+$", "", gloss)
            if form and gloss and len(form) < 30:
                entries.append({
                    "word": form,
                    "transliteration": form,
                    "gloss": gloss,
                })

    return entries


def parse(url: str, **kwargs: Any) -> list[dict]:
    """Download and parse DEDR dictionary entries.

    Args:
        url: URL to DEDR, e.g.
             https://dsal.uchicago.edu/dictionaries/burrow/
        **kwargs:
            entry_range: Tuple (start, end) of entry numbers to fetch.
                         Default fetches the main page only.
            max_pages: Maximum number of pages to fetch (default 10).

    Returns:
        List of dicts with keys: word, transliteration, gloss.
        Returns empty list if URL is unreachable.
    """
    # No artificial page limit — follow ALL pagination links
    max_pages = kwargs.get("max_pages", 9999)
    logger.info("DEDR: downloading %s", url)

    try:
        req = urllib.request.Request(url, headers={"User-Agent": "PhaiPhon/1.0"})
        with urllib.request.urlopen(req, timeout=30) as resp:
            html = resp.read().decode("utf-8", errors="replace")
    except (urllib.error.URLError, urllib.error.HTTPError, OSError) as exc:
        logger.warning("DEDR: failed to download %s: %s", url, exc)
        return []

    # Try structured parsing
    parser = DEDRPageParser()
    parser.feed(html)
    entries = parser.entries

    # Fallback to regex
    if not entries:
        logger.info("DEDR: structured parsing found 0 entries, trying regex")
        entries = _regex_extract(html)

    # Follow pagination links if present
    page_links: set[str] = set()
    for m in re.finditer(r'href="([^"]*(?:page=\d+|/\d+/?)[^"]*)"', html, re.IGNORECASE):
        href = m.group(1)
        if href.startswith("/"):
            domain_match = re.match(r"(https?://[^/]+)", url)
            if domain_match:
                href = domain_match.group(1) + href
        elif not href.startswith("http"):
            href = url.rstrip("/") + "/" + href
        page_links.add(href)

    pages_fetched = 0
    for page_url in sorted(page_links):
        if pages_fetched >= max_pages:
            break
        logger.info("DEDR: downloading page %s", page_url)
        try:
            req = urllib.request.Request(page_url, headers={"User-Agent": "PhaiPhon/1.0"})
            with urllib.request.urlopen(req, timeout=30) as resp:
                page_html = resp.read().decode("utf-8", errors="replace")
        except (urllib.error.URLError, urllib.error.HTTPError, OSError) as exc:
            logger.warning("DEDR: failed to download page %s: %s", page_url, exc)
            continue

        page_entries = _regex_extract(page_html)
        entries.extend(page_entries)
        pages_fetched += 1

    # Deduplicate
    seen: set[tuple[str, str]] = set()
    unique: list[dict] = []
    for e in entries:
        key = (e["word"], e["gloss"])
        if key not in seen:
            seen.add(key)
            unique.append(e)

    logger.info("DEDR: extracted %d unique entries total", len(unique))
    return unique


if __name__ == "__main__":
    import sys
    logging.basicConfig(level=logging.INFO, format="%(levelname)s: %(message)s")

    test_url = (
        sys.argv[1] if len(sys.argv) > 1
        else "https://dsal.uchicago.edu/dictionaries/burrow/"
    )
    results = parse(test_url)
    print(f"\nExtracted {len(results)} entries:")
    for entry in results[:15]:
        print(f"  *{entry['word']:25s}  {entry['gloss']}")
    if len(results) > 15:
        print(f"  ... and {len(results) - 15} more")