|
|
|
|
|
""" |
|
|
Analyze UVW 2026 dataset statistics. |
|
|
|
|
|
UVW 2026: Underthesea Vietnamese Wikipedia Dataset |
|
|
https://github.com/undertheseanlp/underthesea/issues/896 |
|
|
""" |
|
|
|
|
|
import json |
|
|
from collections import Counter |
|
|
from pathlib import Path |
|
|
|
|
|
from tqdm import tqdm |
|
|
|
|
|
|
|
|
DATA_DIR = Path(__file__).parent.parent / "data" / "processed" |
|
|
SPLITS_DIR = Path(__file__).parent.parent / "data" / "splits" |
|
|
|
|
|
|
|
|
def load_jsonl(path: Path) -> list: |
|
|
"""Load data from JSONL file.""" |
|
|
data = [] |
|
|
with open(path, "r", encoding="utf-8") as f: |
|
|
for line in f: |
|
|
data.append(json.loads(line)) |
|
|
return data |
|
|
|
|
|
|
|
|
def analyze_content_length(articles: list) -> dict: |
|
|
"""Analyze content length distribution.""" |
|
|
lengths = [a["num_chars"] for a in articles] |
|
|
lengths.sort() |
|
|
|
|
|
n = len(lengths) |
|
|
return { |
|
|
"min": lengths[0], |
|
|
"max": lengths[-1], |
|
|
"mean": sum(lengths) // n, |
|
|
"median": lengths[n // 2], |
|
|
"p10": lengths[int(n * 0.1)], |
|
|
"p25": lengths[int(n * 0.25)], |
|
|
"p75": lengths[int(n * 0.75)], |
|
|
"p90": lengths[int(n * 0.9)], |
|
|
} |
|
|
|
|
|
|
|
|
def analyze_titles(articles: list) -> dict: |
|
|
"""Analyze article titles.""" |
|
|
titles = [a["title"] for a in articles] |
|
|
|
|
|
|
|
|
first_chars = Counter(t[0].upper() for t in titles if t) |
|
|
|
|
|
|
|
|
title_lengths = [len(t) for t in titles] |
|
|
|
|
|
return { |
|
|
"total": len(titles), |
|
|
"avg_title_length": sum(title_lengths) // len(title_lengths), |
|
|
"first_char_distribution": dict(first_chars.most_common(26)), |
|
|
} |
|
|
|
|
|
|
|
|
def count_vietnamese_chars(text: str) -> int: |
|
|
"""Count Vietnamese-specific characters.""" |
|
|
vietnamese_chars = set("àáảãạăằắẳẵặâầấẩẫậèéẻẽẹêềếểễệìíỉĩịòóỏõọôồốổỗộơờớởỡợùúủũụưừứửữựỳýỷỹỵđ") |
|
|
vietnamese_chars.update(c.upper() for c in vietnamese_chars) |
|
|
return sum(1 for c in text if c in vietnamese_chars) |
|
|
|
|
|
|
|
|
def analyze_language(articles: list) -> dict: |
|
|
"""Analyze Vietnamese language characteristics.""" |
|
|
total_chars = 0 |
|
|
vietnamese_chars = 0 |
|
|
|
|
|
for article in tqdm(articles[:1000], desc="Analyzing language"): |
|
|
content = article["content"] |
|
|
total_chars += len(content) |
|
|
vietnamese_chars += count_vietnamese_chars(content) |
|
|
|
|
|
return { |
|
|
"sample_size": min(1000, len(articles)), |
|
|
"vietnamese_char_ratio": vietnamese_chars / total_chars if total_chars else 0, |
|
|
} |
|
|
|
|
|
|
|
|
def main(): |
|
|
"""Analyze dataset statistics.""" |
|
|
jsonl_path = DATA_DIR / "uvw_2026.jsonl" |
|
|
|
|
|
if not jsonl_path.exists(): |
|
|
print(f"Dataset not found: {jsonl_path}") |
|
|
print("Please run extract_articles.py first.") |
|
|
return |
|
|
|
|
|
print("Loading dataset...") |
|
|
articles = load_jsonl(jsonl_path) |
|
|
print(f"Total articles: {len(articles)}") |
|
|
|
|
|
print("\n" + "=" * 50) |
|
|
print("CONTENT LENGTH ANALYSIS") |
|
|
print("=" * 50) |
|
|
length_stats = analyze_content_length(articles) |
|
|
for key, value in length_stats.items(): |
|
|
print(f" {key}: {value:,} chars") |
|
|
|
|
|
print("\n" + "=" * 50) |
|
|
print("TITLE ANALYSIS") |
|
|
print("=" * 50) |
|
|
title_stats = analyze_titles(articles) |
|
|
print(f" Total titles: {title_stats['total']:,}") |
|
|
print(f" Avg title length: {title_stats['avg_title_length']} chars") |
|
|
print(f" First character distribution (top 10):") |
|
|
for char, count in list(title_stats["first_char_distribution"].items())[:10]: |
|
|
print(f" {char}: {count:,}") |
|
|
|
|
|
print("\n" + "=" * 50) |
|
|
print("LANGUAGE ANALYSIS") |
|
|
print("=" * 50) |
|
|
lang_stats = analyze_language(articles) |
|
|
print(f" Sample size: {lang_stats['sample_size']}") |
|
|
print(f" Vietnamese char ratio: {lang_stats['vietnamese_char_ratio']:.2%}") |
|
|
|
|
|
print("\n" + "=" * 50) |
|
|
print("OVERALL STATISTICS") |
|
|
print("=" * 50) |
|
|
total_chars = sum(a["num_chars"] for a in articles) |
|
|
total_sentences = sum(a["num_sentences"] for a in articles) |
|
|
print(f" Total articles: {len(articles):,}") |
|
|
print(f" Total characters: {total_chars:,}") |
|
|
print(f" Total sentences: {total_sentences:,}") |
|
|
print(f" Avg chars/article: {total_chars // len(articles):,}") |
|
|
print(f" Avg sentences/article: {total_sentences // len(articles)}") |
|
|
|
|
|
|
|
|
size_categories = { |
|
|
"small (<1K chars)": 0, |
|
|
"medium (1K-10K chars)": 0, |
|
|
"large (10K-100K chars)": 0, |
|
|
"very large (>100K chars)": 0, |
|
|
} |
|
|
for a in articles: |
|
|
chars = a["num_chars"] |
|
|
if chars < 1000: |
|
|
size_categories["small (<1K chars)"] += 1 |
|
|
elif chars < 10000: |
|
|
size_categories["medium (1K-10K chars)"] += 1 |
|
|
elif chars < 100000: |
|
|
size_categories["large (10K-100K chars)"] += 1 |
|
|
else: |
|
|
size_categories["very large (>100K chars)"] += 1 |
|
|
|
|
|
print("\n Article size distribution:") |
|
|
for category, count in size_categories.items(): |
|
|
pct = count / len(articles) * 100 |
|
|
print(f" {category}: {count:,} ({pct:.1f}%)") |
|
|
|
|
|
|
|
|
analysis = { |
|
|
"total_articles": len(articles), |
|
|
"total_characters": total_chars, |
|
|
"total_sentences": total_sentences, |
|
|
"content_length": length_stats, |
|
|
"title_stats": title_stats, |
|
|
"language_stats": lang_stats, |
|
|
"size_distribution": size_categories, |
|
|
} |
|
|
|
|
|
analysis_path = DATA_DIR / "analysis.json" |
|
|
with open(analysis_path, "w", encoding="utf-8") as f: |
|
|
json.dump(analysis, f, ensure_ascii=False, indent=2) |
|
|
print(f"\nAnalysis saved to: {analysis_path}") |
|
|
|
|
|
|
|
|
if __name__ == "__main__": |
|
|
main() |
|
|
|