--- license: apache-2.0 datasets: - Remeinium/WWHO_30m language: - si - hi - en pipeline_tag: feature-extraction library_name: transformers tags: - tokenizer - WWHO - SGPE - linguis_trie - token - tokenization - Syllable - remeinium - transformer - linguistics - NLP - sinhala - hindi - english - BPE - GPE model-index: - name: WWHO results: - task: type: feature-extraction dataset: name: WWHO_30m type: Remeinium/WWHO_30m metrics: - name: Token-to-Word Ratio (TWR) - Sinhala type: twr value: 1.274 verified: false - name: Token-to-Word Ratio (TWR) - Hindi type: twr value: 1.181 verified: false - name: Token-to-Word Ratio (TWR) - Overall type: twr value: 1.240 verified: false --- # Separate before you Compress ## The Next Architectural Primitive in Tokenization Large language models remain linguistically blind to Abugida scripts. Byte-Pair Encoding and its descendants routinely shatter complex conjuncts — atomic multi-codepoint grapheme clusters that constitute the fundamental phonetic units of Indic and Southeast Asian writing systems — into meaningless sub-character fragments. The result is degraded reasoning, inflated inference costs, and a systemic “Token Tax” that disproportionately burdens more than one billion speakers. **WWHO (Where-What-How Often) introduces the clean separation of concerns the field has been missing.** By decoupling linguistic structural constraints from statistical compression, WWHO builds a unified meta-vocabulary space: 1. **Layer 1 (Where): Code-Switching Router** A linear $O(N)$ block scanner that evaluates characters in $O(1)$ time to inherently identify script boundaries, routing Latin text to proven frontier tokenizers (like `o200k_base`) while sending Abugida text for specialized processing. 2. **Layer 2 (What): LinguisTrie** Enforces linguistic integrity by construction: a DFA based syllabifier segments raw Unicode into well-formed syllables with a formal zero-breakage guarantee. 3. **Layer 3 (How Often): SGPE & Meta-Vocabulary** Performs statistical pair merging exclusively over this linguistically sound stream, safely projecting the resulting tokens into a unified, mathematically offset ID space. Sinhala and Devanagari serve as the high-complexity proofs-of-concept. The same architecture generalizes directly to Tamil, Khmer, Myanmar, and the broader Abugida family. --- ## Multi-Script Stratified Benchmarks (122.2M Characters) We evaluated WWHO against frontier models across a 1.5 million sentence code-switched corpus containing Sinhala, Hindi (Devanagari), and English. ### 1. Sinhala Efficiency | Tokenizer | Tokens | TWR | Chr/Tok | % Reduction | |---|---|---|---|---| | **SGPE(WWHO)** | **6,654,288** | **1.274** | **4.83** | **-** | | OpenAI (o200k_base) | 17,360,196 | 3.324 | 1.85 | 61.7% | | Llama 4 Scout | 18,157,707 | 3.476 | 1.77 | 63.4% | | DeepSeek V3 | 29,152,698 | 5.581 | 1.10 | 77.2% | ### 2. Hindi (Devanagari) Efficiency | Tokenizer | Tokens | TWR | Chr/Tok | % Reduction | |---|---|---|---|---| | **SGPE(WWHO)** | **13,433,554** | **1.181** | **4.29** | **-** | | OpenAI (o200k_base) | 18,394,075 | 1.617 | 3.13 | 27.0% | | Llama 4 Scout | 19,566,121 | 1.720 | 2.94 | 31.3% | | DeepSeek V3 | 31,682,218 | 2.786 | 1.82 | 57.6% | ### 3. English | Tokenizer | Tokens | TWR | Chr/Tok | % Reduction | |---|---|---|---|---| | **SGPE(WWHO)** | **7,240,147** | **1.330** | **4.46** | **-** | | OpenAI (o200k_base) | 7,420,527 | 1.364 | 4.35 | 2.4% | | Llama 4 Scout | 7,512,843 | 1.381 | 4.30 | 3.6% | | DeepSeek V3 | 7,904,670 | 1.453 | 4.09 | 8.4% | *(Note: Because WWHO routes Latin text directly to the native Tiktoken sequence, English performance is mathematically identical. The minor delta in total tokens emerges solely from boundary crossing mechanics.)* ### 4. Overall (Mixed-Script) | Tokenizer | Tokens | TWR | Chr/Tok | % Reduction | |---|---|---|---|---| | **SGPE(WWHO)** | **27,327,989** | **1.240** | **4.47** | **-** | | OpenAI (o200k_base) | 43,174,798 | 1.959 | 2.83 | 36.7% | | Llama 4 Scout | 45,236,671 | 2.053 | 2.70 | 39.6% | | DeepSeek V3 | 68,739,586 | 3.119 | 1.78 | 60.2% | - **Zero-Breakage Guarantee**: Validated through exhaustive testing permutations across all supported Abugida scripts (0 violations). - **Full-corpus reconstruction**: 1.5M code-switched sentences encoded and decoded with 0 non-UNK mismatches. - **UNK rate**: 0.08 % (restricted strictly to rare compounds without violating structural boundaries). WWHO radically compresses the context window for Abugida text, effectively ending the Token Tax without penalizing existing state-of-the-art programming and reasoning capabilities. --- ## Quick Start with Hugging Face ```python from transformers import AutoTokenizer tokenizer = AutoTokenizer.from_pretrained("remeinium/SGPE") text = "ආයුබෝවන් ශ්‍රී ලංකා" tokens = tokenizer.tokenize(text) # ['ආයුබෝවන්', ' ශ්‍රී', ' ලංකා'] print(tokenizer.encode(text)) ``` --- ## Resources - **Pre-trained Tokenizer**: [Hugging Face](https://huggingface.co/remeinium/WWHO) - **Cleaned Training Corpus**: [Hugging Face](https://huggingface.co/datasets/remeinium/WWHO_30m) - **Full Code & Evaluation Harness**: [GitHub](https://github.com/remeinium/WWHO) --- ## License Apache License 2.0 — see [LICENSE](LICENSE). **Remeinium Research | Remeinium AI | Intelligence for a Greater Tomorrow** ---