| | |
| | """ |
| | Add Wikidata ID and main category to UVW 2026 dataset. |
| | |
| | Uses concurrent requests to Wikidata API for faster processing. |
| | |
| | Fields added: |
| | - wikidata_id: The Q-ID from Wikidata (e.g., Q1930) |
| | - main_category: The P31 (instance of) property label in Vietnamese |
| | |
| | API Reference: |
| | - https://www.wikidata.org/w/api.php?action=help&modules=wbgetentities |
| | """ |
| |
|
| | import json |
| | import time |
| | from concurrent.futures import ThreadPoolExecutor, as_completed |
| | from pathlib import Path |
| | from typing import Optional |
| | import urllib.request |
| | import urllib.parse |
| | import urllib.error |
| | import threading |
| |
|
| | from tqdm import tqdm |
| |
|
| | INPUT_PATH = Path(__file__).parent.parent / "data" / "processed" / "uvw_2026_quality.jsonl" |
| | OUTPUT_PATH = Path(__file__).parent.parent / "data" / "processed" / "uvw_2026_wikidata.jsonl" |
| | CACHE_PATH = Path(__file__).parent.parent / "data" / "processed" / "wikidata_cache.json" |
| |
|
| | |
| | WIKIDATA_API = "https://www.wikidata.org/w/api.php" |
| |
|
| | |
| | BATCH_SIZE = 50 |
| |
|
| | |
| | MAX_WORKERS = 8 |
| |
|
| | |
| | TIMEOUT = 60 |
| |
|
| | |
| | _label_cache: dict[str, Optional[str]] = {} |
| | _label_cache_lock = threading.Lock() |
| |
|
| |
|
| | def load_cache() -> dict: |
| | """Load cached Wikidata mappings.""" |
| | if CACHE_PATH.exists(): |
| | with open(CACHE_PATH, "r", encoding="utf-8") as f: |
| | return json.load(f) |
| | return {} |
| |
|
| |
|
| | def save_cache(cache: dict): |
| | """Save Wikidata mappings to cache.""" |
| | with open(CACHE_PATH, "w", encoding="utf-8") as f: |
| | json.dump(cache, f, ensure_ascii=False) |
| |
|
| |
|
| | def fetch_entity_labels_batch(entity_ids: list[str]) -> dict[str, Optional[str]]: |
| | """Fetch Vietnamese/English labels for a batch of Wikidata entity IDs.""" |
| | global _label_cache |
| |
|
| | with _label_cache_lock: |
| | to_fetch = [eid for eid in entity_ids if eid not in _label_cache] |
| |
|
| | if not to_fetch: |
| | with _label_cache_lock: |
| | return {eid: _label_cache.get(eid) for eid in entity_ids} |
| |
|
| | ids_param = "|".join(to_fetch[:50]) |
| |
|
| | params = { |
| | "action": "wbgetentities", |
| | "ids": ids_param, |
| | "props": "labels", |
| | "languages": "vi|en", |
| | "format": "json", |
| | } |
| |
|
| | url = f"{WIKIDATA_API}?{urllib.parse.urlencode(params)}" |
| |
|
| | try: |
| | req = urllib.request.Request( |
| | url, |
| | headers={"User-Agent": "UVW2026-Dataset/1.0 (https://github.com/undertheseanlp)"} |
| | ) |
| | with urllib.request.urlopen(req, timeout=TIMEOUT) as response: |
| | data = json.loads(response.read().decode("utf-8")) |
| |
|
| | with _label_cache_lock: |
| | if "entities" in data: |
| | for entity_id, entity in data["entities"].items(): |
| | if "labels" in entity: |
| | labels = entity["labels"] |
| | if "vi" in labels: |
| | _label_cache[entity_id] = labels["vi"]["value"] |
| | elif "en" in labels: |
| | _label_cache[entity_id] = labels["en"]["value"] |
| | else: |
| | _label_cache[entity_id] = None |
| | else: |
| | _label_cache[entity_id] = None |
| |
|
| | except Exception: |
| | with _label_cache_lock: |
| | for eid in to_fetch[:50]: |
| | if eid not in _label_cache: |
| | _label_cache[eid] = None |
| |
|
| | with _label_cache_lock: |
| | return {eid: _label_cache.get(eid) for eid in entity_ids} |
| |
|
| |
|
| | def fetch_wikidata_batch(titles: list[str], batch_id: int = 0) -> dict[str, dict]: |
| | """ |
| | Fetch Wikidata entities for a batch of Vietnamese Wikipedia titles. |
| | |
| | Args: |
| | titles: List of Wikipedia article titles (max 50) |
| | batch_id: Batch identifier for error reporting |
| | |
| | Returns: |
| | Dict mapping title to {wikidata_id, main_category} |
| | """ |
| | results = {} |
| | category_ids_to_fetch = [] |
| | title_to_category_id = {} |
| |
|
| | titles_param = "|".join(titles) |
| |
|
| | params = { |
| | "action": "wbgetentities", |
| | "sites": "viwiki", |
| | "titles": titles_param, |
| | "props": "claims|sitelinks", |
| | "format": "json", |
| | } |
| |
|
| | url = f"{WIKIDATA_API}?{urllib.parse.urlencode(params)}" |
| |
|
| | try: |
| | req = urllib.request.Request( |
| | url, |
| | headers={"User-Agent": "UVW2026-Dataset/1.0 (https://github.com/undertheseanlp)"} |
| | ) |
| | with urllib.request.urlopen(req, timeout=TIMEOUT) as response: |
| | data = json.loads(response.read().decode("utf-8")) |
| |
|
| | if "entities" not in data: |
| | |
| | for title in titles: |
| | results[title] = {"wikidata_id": None, "main_category": None} |
| | return results |
| |
|
| | |
| | found_titles = set() |
| |
|
| | for entity_id, entity in data["entities"].items(): |
| | if entity_id.startswith("-"): |
| | continue |
| |
|
| | |
| | if "sitelinks" in entity and "viwiki" in entity["sitelinks"]: |
| | title = entity["sitelinks"]["viwiki"]["title"] |
| | else: |
| | continue |
| |
|
| | found_titles.add(title) |
| |
|
| | |
| | wikidata_id = entity.get("id") |
| |
|
| | |
| | category_id = None |
| | if "claims" in entity and "P31" in entity["claims"]: |
| | p31_claims = entity["claims"]["P31"] |
| | if p31_claims: |
| | first_claim = p31_claims[0] |
| | if "mainsnak" in first_claim: |
| | mainsnak = first_claim["mainsnak"] |
| | if mainsnak.get("datatype") == "wikibase-item": |
| | datavalue = mainsnak.get("datavalue", {}) |
| | if datavalue.get("type") == "wikibase-entityid": |
| | category_id = datavalue["value"].get("id") |
| |
|
| | results[title] = { |
| | "wikidata_id": wikidata_id, |
| | "main_category": None, |
| | } |
| |
|
| | if category_id: |
| | category_ids_to_fetch.append(category_id) |
| | title_to_category_id[title] = category_id |
| |
|
| | |
| | for title in titles: |
| | if title not in found_titles: |
| | results[title] = {"wikidata_id": None, "main_category": None} |
| |
|
| | except Exception as e: |
| | |
| | for title in titles: |
| | results[title] = {"wikidata_id": None, "main_category": None} |
| | return results |
| |
|
| | |
| | if category_ids_to_fetch: |
| | unique_ids = list(set(category_ids_to_fetch)) |
| | labels = fetch_entity_labels_batch(unique_ids) |
| |
|
| | for title, category_id in title_to_category_id.items(): |
| | if title in results: |
| | results[title]["main_category"] = labels.get(category_id) |
| |
|
| | return results |
| |
|
| |
|
| | def process_batch(args: tuple) -> tuple[int, dict[str, dict]]: |
| | """Process a single batch, returns (batch_id, results).""" |
| | batch_id, titles = args |
| | results = fetch_wikidata_batch(titles, batch_id) |
| | return batch_id, results |
| |
|
| |
|
| | def count_lines(path: Path) -> int: |
| | """Count lines in a file without loading it all into memory.""" |
| | count = 0 |
| | with open(path, "r", encoding="utf-8") as f: |
| | for _ in f: |
| | count += 1 |
| | return count |
| |
|
| |
|
| | def iter_titles(path: Path): |
| | """Iterate over titles from JSONL file without loading full articles.""" |
| | with open(path, "r", encoding="utf-8") as f: |
| | for line in f: |
| | |
| | article = json.loads(line) |
| | yield article["title"] |
| |
|
| |
|
| | def iter_articles(path: Path): |
| | """Iterate over articles from JSONL file one at a time.""" |
| | with open(path, "r", encoding="utf-8") as f: |
| | for line in f: |
| | yield json.loads(line) |
| |
|
| |
|
| | def batch_iterator(iterable, batch_size: int): |
| | """Yield batches from an iterable without loading all into memory.""" |
| | batch = [] |
| | for item in iterable: |
| | batch.append(item) |
| | if len(batch) >= batch_size: |
| | yield batch |
| | batch = [] |
| | if batch: |
| | yield batch |
| |
|
| |
|
| | def main(): |
| | """Add Wikidata information to dataset.""" |
| | print("Adding Wikidata IDs and categories to UVW 2026 dataset...") |
| | print(f"Input: {INPUT_PATH}") |
| | print(f"Output: {OUTPUT_PATH}") |
| | print(f"Using {MAX_WORKERS} concurrent workers") |
| |
|
| | |
| | cache = load_cache() |
| | print(f"Loaded {len(cache):,} cached Wikidata mappings") |
| |
|
| | |
| | print("Counting articles...") |
| | total = count_lines(INPUT_PATH) |
| | print(f"Found {total:,} articles") |
| |
|
| | |
| | print("Finding uncached titles...") |
| | uncached_titles = [] |
| | for title in tqdm(iter_titles(INPUT_PATH), total=total, desc="Scanning titles"): |
| | if title not in cache: |
| | uncached_titles.append(title) |
| | print(f"Need to fetch {len(uncached_titles):,} titles from Wikidata API") |
| |
|
| | |
| | if uncached_titles: |
| | num_batches = (len(uncached_titles) + BATCH_SIZE - 1) // BATCH_SIZE |
| | print(f"Processing {num_batches:,} batches...") |
| |
|
| | completed = 0 |
| | last_save = 0 |
| |
|
| | |
| | SUBMIT_CHUNK_SIZE = MAX_WORKERS * 10 |
| |
|
| | with tqdm(total=num_batches, desc="Fetching Wikidata", smoothing=0.1) as pbar: |
| | with ThreadPoolExecutor(max_workers=MAX_WORKERS) as executor: |
| | batch_id = 0 |
| | for chunk_start in range(0, len(uncached_titles), BATCH_SIZE * SUBMIT_CHUNK_SIZE): |
| | chunk_end = min(chunk_start + BATCH_SIZE * SUBMIT_CHUNK_SIZE, len(uncached_titles)) |
| | chunk_titles = uncached_titles[chunk_start:chunk_end] |
| |
|
| | |
| | title_batches = [chunk_titles[i:i + BATCH_SIZE] |
| | for i in range(0, len(chunk_titles), BATCH_SIZE)] |
| |
|
| | |
| | future_to_batch = { |
| | executor.submit(process_batch, (batch_id + idx, titles_list)): batch_id + idx |
| | for idx, titles_list in enumerate(title_batches) |
| | } |
| | batch_id += len(title_batches) |
| |
|
| | for future in as_completed(future_to_batch): |
| | try: |
| | _, results = future.result() |
| | cache.update(results) |
| | completed += 1 |
| | pbar.update(1) |
| |
|
| | |
| | if completed - last_save >= 1000: |
| | save_cache(cache) |
| | last_save = completed |
| |
|
| | except Exception as e: |
| | print(f"Batch error: {e}") |
| |
|
| | |
| | del uncached_titles |
| |
|
| | |
| | save_cache(cache) |
| | print(f"Saved {len(cache):,} entries to cache") |
| |
|
| | |
| | stats = { |
| | "with_wikidata_id": 0, |
| | "with_category": 0, |
| | "missing": 0, |
| | } |
| | category_counts: dict[str, int] = {} |
| |
|
| | print("Writing output (streaming)...") |
| | with open(OUTPUT_PATH, "w", encoding="utf-8") as fout: |
| | for article in tqdm(iter_articles(INPUT_PATH), total=total, desc="Writing output"): |
| | title = article["title"] |
| | wikidata_info = cache.get(title, {}) |
| |
|
| | article["wikidata_id"] = wikidata_info.get("wikidata_id") |
| | article["main_category"] = wikidata_info.get("main_category") |
| |
|
| | if article["wikidata_id"]: |
| | stats["with_wikidata_id"] += 1 |
| | else: |
| | stats["missing"] += 1 |
| |
|
| | if article["main_category"]: |
| | stats["with_category"] += 1 |
| | category = article["main_category"] |
| | category_counts[category] = category_counts.get(category, 0) + 1 |
| |
|
| | fout.write(json.dumps(article, ensure_ascii=False) + "\n") |
| |
|
| | |
| | print("\nWikidata coverage:") |
| | print("-" * 50) |
| | print(f" With Wikidata ID: {stats['with_wikidata_id']:>10,} ({stats['with_wikidata_id']/total*100:5.1f}%)") |
| | print(f" With category: {stats['with_category']:>10,} ({stats['with_category']/total*100:5.1f}%)") |
| | print(f" Missing: {stats['missing']:>10,} ({stats['missing']/total*100:5.1f}%)") |
| |
|
| | print(f"\nTop 30 categories:") |
| | print("-" * 50) |
| | top_categories = sorted(category_counts.items(), key=lambda x: -x[1])[:30] |
| | for category, count in top_categories: |
| | pct = count / total * 100 |
| | print(f" {category:40s} {count:>8,} ({pct:5.2f}%)") |
| |
|
| | print(f"\nTotal unique categories: {len(category_counts):,}") |
| | print(f"Total articles: {total:,}") |
| | print(f"Output saved to: {OUTPUT_PATH}") |
| | print(f"Cache saved to: {CACHE_PATH}") |
| |
|
| |
|
| | if __name__ == "__main__": |
| | main() |
| |
|