Data Processing Pipeline
This document describes the technical workflow for constructing and maintaining the Nepali Text Corpus.
Overview
The corpus is built using a DuckDB-based pipeline (scripts/merge.py) that ingests raw CSVs, applies filtering and normalization, and produces stratified parquet outputs optimized for different research domains.
Raw Data (CSV) → Validation & Normalization → Domain Stratification → Parquet Output
Input Sources
| Source | File | Format | Records | Notes |
|---|---|---|---|---|
| IRIISNEPAL | iriisnepal_raw.csv |
CSV | ~6.1M | Manually curated formal Nepali |
| YouTube | youtube_comments_clean.csv |
CSV | ~431K | Pre-cleaned by clean.py |
| Wikipedia | wikipedia_nepali.csv |
CSV | ~291K | Extracted from wiki dump |
| News | nepali_news.csv |
CSV | ~87K | Scraped from live news feeds |
Preprocessing Steps
1. Text Validation
All records are filtered on:
- Non-null check:
text IS NOT NULL - Non-empty check:
trim(text) <> '' - Minimum length (IRIIS only):
length(split(trim(text), ' ')) >= 5(5+ words) - Script validation (IRIIS only): Must contain at least one Devanagari character
2. Script Detection
Automatic classification for news and YouTube sources:
IF text contains [ऀ-ॿ] THEN 'devanagari'
ELSE IF text contains [A-Za-z] THEN 'latin'
ELSE 'other'
3. Metadata Assignment
Each row is enriched with:
- source: Origin identifier (e.g.,
iriisnepal,youtube_comments,wikipedia_nepali) - domain: Content category (formal, colloquial, encyclopedia, news)
- script: Writing system detected
- lang: ISO 639-1 code (
nefor Nepali) - date_collected: Processing date
- license: Source-specific license
4. Deduplication
- No exact-duplicate removal (preserves all unique utterances)
- Partial duplicates retained (colloquial speech naturally repeats common phrases)
Output Datasets
nepali_corpus_full.parquet
Purpose: Complete merged corpus for general research
Rows: 7,167,456
Ordering:
- Formal domain (IRIISNEPAL)
- Encyclopedia domain (Wikipedia)
- News domain
- Colloquial domain (YouTube)
Within each domain, ordered by source, then length DESC for visibility.
nepali_corpus_formal.parquet
Purpose: Formal writing for LM pretraining
Rows: 6,378,206 (formal + encyclopedia + news)
Domains included: formal, encyclopedia, news
Ordering: Domain priority → source → length DESC
Professional dataset preview: Leading rows are formal Wikipedia and news articles (clean, representative examples).
nepali_corpus_colloquial.parquet
Purpose: Conversational Nepali for sociolinguistic analysis
Rows: 431,648 (YouTube comments only)
Script distribution:
- Devanagari: 123,804 comments
- Latin (Roman): 307,999 comments
- Mixed: 19,845 comments
Ordering: Devanagari (longest first) → Latin → Mixed (ensures Hugging Face viewer surfaces Devanagari examples first)
nepali_corpus_roman.parquet
Purpose: Roman-script Nepali subset
Rows: 307,999 (latin script from YouTube)
Derived from: Colloquial corpus with script = 'latin'
nepali_corpus_wikipedia.parquet
Purpose: Encyclopedia-style Nepali (NOT published separately; kept for analysis)
Rows: 291,767
Note: Wikipedia data is merged into nepali_corpus_formal.parquet for public release.
Performance Optimizations
DuckDB Configuration
PRAGMA threads = 2 -- CPU parallelism
PRAGMA preserve_insertion_order = false
PRAGMA memory_limit = '8GB' -- RAM cap
PRAGMA temp_directory = '...' -- Disk spillover
Parquet Compression
- Codec: Zstandard (ZSTD)
- Compression Level: Default (best balance)
- Result: 5.95 GB file for 7.1M rows (≈850 bytes/row average)
Processing Speed
- Typical merge run: ~2-5 minutes on 8GB RAM
- DuckDB streaming keeps memory footprint constant
Quality Metrics
Content Representation
- Formal (88%): Academic, journalistic, encyclopedic content
- Colloquial (6%): Conversational, social media discourse
- Roman script (4%): Transliterated Nepali
- Mixed script (<1%): Code-switching examples
Text Statistics
- Average length: ~120 UTF-8 characters
- Median length: ~85 characters
- Max length: ~50,000 characters (rare outliers)
- Devanagari percentage: ~87% of rows
Coverage
- Unique sources: 5 primary (IRIIS, Wikipedia, YouTube, Kantipur, Setopati, +3 others)
- Time span: Circa 2016–2026 (mixed historical and current)
- Geographic scope: Nepal-centric; diaspora content included in YouTube
Maintenance & Updates
Incremental Updates
To add new data:
- Append new rows to source CSV (e.g.,
nepali_news.csv) - Re-run
scripts/merge.py - Output parquets are regenerated from scratch
- Run
scripts/publish.pyto sync to Hugging Face
Re-running the Pipeline
cd /Users/ad/research/nepali-text
venv/bin/python scripts/merge.py
Schema Reference
All parquet files use this schema:
| Column | Type | Description |
|---|---|---|
| text | string | UTF-8 Nepali text |
| source | string | Data source identifier |
| domain | string | Content type (formal, colloquial, encyclopedia, news) |
| script | string | Writing system (devanagari, latin, mixed) |
| lang | string | ISO 639-1 language code (always 'ne') |
| date_collected | string | ISO 8601 processing date |
| license | string | Source license (MIT, CC BY 4.0, CC BY-SA 4.0, source-dependent) |
Last Updated: April 2, 2026
Pipeline Version: 1.0
Maintainer: Boredoom17