omarkamali commited on
Commit
b8ce0e4
·
verified ·
1 Parent(s): ff80080

Upload all models and assets for ba (latest)

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +1 -0
  2. README.md +204 -164
  3. models/embeddings/aligned/ba_128d.bin +3 -0
  4. models/embeddings/aligned/ba_128d.meta.json +1 -0
  5. models/embeddings/aligned/ba_128d.projection.npy +3 -0
  6. models/embeddings/aligned/ba_128d_metadata.json +8 -0
  7. models/embeddings/aligned/ba_32d.bin +3 -0
  8. models/embeddings/aligned/ba_32d.meta.json +1 -0
  9. models/embeddings/aligned/ba_32d.projection.npy +3 -0
  10. models/embeddings/aligned/ba_32d_metadata.json +8 -0
  11. models/embeddings/aligned/ba_64d.bin +3 -0
  12. models/embeddings/aligned/ba_64d.meta.json +1 -0
  13. models/embeddings/aligned/ba_64d.projection.npy +3 -0
  14. models/embeddings/aligned/ba_64d_metadata.json +8 -0
  15. models/embeddings/monolingual/ba_128d.bin +2 -2
  16. models/embeddings/monolingual/ba_128d_metadata.json +1 -1
  17. models/embeddings/monolingual/ba_32d.bin +2 -2
  18. models/embeddings/monolingual/ba_32d_metadata.json +1 -1
  19. models/embeddings/monolingual/ba_64d.bin +2 -2
  20. models/embeddings/monolingual/ba_64d_metadata.json +1 -1
  21. models/subword_markov/ba_markov_ctx1_subword.parquet +2 -2
  22. models/subword_markov/ba_markov_ctx1_subword_metadata.json +2 -2
  23. models/subword_markov/ba_markov_ctx2_subword.parquet +2 -2
  24. models/subword_markov/ba_markov_ctx2_subword_metadata.json +2 -2
  25. models/subword_markov/ba_markov_ctx3_subword.parquet +2 -2
  26. models/subword_markov/ba_markov_ctx3_subword_metadata.json +2 -2
  27. models/subword_markov/ba_markov_ctx4_subword.parquet +2 -2
  28. models/subword_markov/ba_markov_ctx4_subword_metadata.json +2 -2
  29. models/subword_ngram/ba_2gram_subword.parquet +2 -2
  30. models/subword_ngram/ba_2gram_subword_metadata.json +2 -2
  31. models/subword_ngram/ba_3gram_subword.parquet +2 -2
  32. models/subword_ngram/ba_3gram_subword_metadata.json +2 -2
  33. models/subword_ngram/ba_4gram_subword.parquet +2 -2
  34. models/subword_ngram/ba_4gram_subword_metadata.json +2 -2
  35. models/subword_ngram/ba_5gram_subword.parquet +3 -0
  36. models/subword_ngram/ba_5gram_subword_metadata.json +7 -0
  37. models/tokenizer/ba_tokenizer_16k.model +2 -2
  38. models/tokenizer/ba_tokenizer_16k.vocab +0 -0
  39. models/tokenizer/ba_tokenizer_32k.model +2 -2
  40. models/tokenizer/ba_tokenizer_32k.vocab +0 -0
  41. models/tokenizer/ba_tokenizer_64k.model +2 -2
  42. models/tokenizer/ba_tokenizer_64k.vocab +0 -0
  43. models/tokenizer/ba_tokenizer_8k.model +2 -2
  44. models/tokenizer/ba_tokenizer_8k.vocab +0 -0
  45. models/vocabulary/ba_vocabulary.parquet +2 -2
  46. models/vocabulary/ba_vocabulary_metadata.json +9 -9
  47. models/word_markov/ba_markov_ctx1_word.parquet +2 -2
  48. models/word_markov/ba_markov_ctx1_word_metadata.json +2 -2
  49. models/word_markov/ba_markov_ctx2_word.parquet +2 -2
  50. models/word_markov/ba_markov_ctx2_word_metadata.json +2 -2
.gitattributes CHANGED
@@ -39,3 +39,4 @@ visualizations/position_encoding_comparison.png filter=lfs diff=lfs merge=lfs -t
39
  visualizations/tsne_sentences.png filter=lfs diff=lfs merge=lfs -text
40
  visualizations/tsne_words.png filter=lfs diff=lfs merge=lfs -text
41
  visualizations/zipf_law.png filter=lfs diff=lfs merge=lfs -text
 
 
39
  visualizations/tsne_sentences.png filter=lfs diff=lfs merge=lfs -text
40
  visualizations/tsne_words.png filter=lfs diff=lfs merge=lfs -text
41
  visualizations/zipf_law.png filter=lfs diff=lfs merge=lfs -text
42
+ visualizations/embedding_tsne_multilingual.png filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  language: ba
3
- language_name: BA
4
  language_family: turkic_kipchak
5
  tags:
6
  - wikilangs
@@ -10,11 +10,21 @@ tags:
10
  - n-gram
11
  - markov
12
  - wikipedia
 
 
 
 
 
 
 
 
 
 
13
  - monolingual
14
  - family-turkic_kipchak
15
  license: mit
16
  library_name: wikilangs
17
- pipeline_tag: feature-extraction
18
  datasets:
19
  - omarkamali/wikipedia-monthly
20
  dataset_info:
@@ -23,20 +33,20 @@ dataset_info:
23
  metrics:
24
  - name: best_compression_ratio
25
  type: compression
26
- value: 4.673
27
  - name: best_isotropy
28
  type: isotropy
29
- value: 0.7751
30
  - name: vocabulary_size
31
  type: vocab
32
  value: 0
33
  generated: 2026-01-03
34
  ---
35
 
36
- # BA - Wikilangs Models
37
  ## Comprehensive Research Report & Full Ablation Study
38
 
39
- This repository contains NLP models trained and evaluated by Wikilangs, specifically on **BA** Wikipedia data.
40
  We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and word embeddings.
41
 
42
  ## 📋 Repository Contents
@@ -60,7 +70,7 @@ We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and
60
  - [3. Markov Chain Evaluation](#3-markov-chain-evaluation)
61
  - [4. Vocabulary Analysis](#4-vocabulary-analysis)
62
  - [5. Word Embeddings Evaluation](#5-word-embeddings-evaluation)
63
- - [6. Morphological Analysis (Experimental)](#6-morphological-analysis)
64
  - [7. Summary & Recommendations](#7-summary--recommendations)
65
  - [Metrics Glossary](#appendix-metrics-glossary--interpretation-guide)
66
  - [Visualizations Index](#visualizations-index)
@@ -80,47 +90,47 @@ We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and
80
 
81
  | Vocab Size | Compression | Avg Token Len | UNK Rate | Total Tokens |
82
  |------------|-------------|---------------|----------|--------------|
83
- | **8k** | 3.556x | 3.56 | 0.3956% | 1,547,491 |
84
- | **16k** | 3.995x | 4.00 | 0.4444% | 1,377,561 |
85
- | **32k** | 4.373x | 4.37 | 0.4864% | 1,258,583 |
86
- | **64k** | 4.673x 🏆 | 4.68 | 0.5198% | 1,177,657 |
87
 
88
  ### Tokenization Examples
89
 
90
  Below are sample sentences tokenized with each vocabulary size:
91
 
92
- **Sample 1:** `йыл шишәмбе көнөнән башланған йыл, кәбисә түгел. Ваҡиғалар Тыуғандар Вафат бул...`
93
 
94
  | Vocab | Tokens | Count |
95
  |-------|--------|-------|
96
- | 8k | `▁йыл ▁— ▁шиш әм бе ▁көнөнән ▁башланған ▁йыл , ▁кәбисә ... (+10 more)` | 20 |
97
- | 16k | `▁йыл ▁— ▁шиш әм бе ▁көнөнән ▁башланған ▁йыл , ▁кәбисә ... (+10 more)` | 20 |
98
- | 32k | `▁йыл ▁— ▁шишәмбе ▁көнөнән ▁башланған ▁йыл , ▁кәбисә ▁түгел . ... (+8 more)` | 18 |
99
- | 64k | `▁йыл ▁— ▁шишәмбе ▁көнөнән ▁башланған ▁йыл , ▁кәбисә ▁түгел . ... (+8 more)` | 18 |
100
 
101
- **Sample 2:** `Азимут: Азимут геодезияла бирелгән йүнәлеш менән төньяҡҡа табан булған йүнәлеш...`
102
 
103
  | Vocab | Tokens | Count |
104
  |-------|--------|-------|
105
- | 8k | `▁аз им ут : ▁аз им ут ▁— ▁ге од ... (+29 more)` | 39 |
106
- | 16k | `▁аз им ут : ▁аз им ут ▁— ▁геод ез ... (+27 more)` | 37 |
107
- | 32k | `▁аз им ут : ▁аз им ут ▁— ▁геодез ияла ... (+23 more)` | 33 |
108
- | 64k | `▁азим ут : ▁азим ут ▁— ▁геодез ияла ▁бирелгән ▁йүнәлеш ... (+19 more)` | 29 |
109
 
110
- **Sample 3:** `Апанай мәсете ( ) Ҡазан мәсете , татар архитектура культы ҡомартҡыһы. Ҡаҙанда ...`
111
 
112
  | Vocab | Tokens | Count |
113
  |-------|--------|-------|
114
- | 8k | `▁ап ан ай ▁мәсете ▁( ▁) ▁— ▁ҡазан ▁мәсете ▁, ... (+18 more)` | 28 |
115
- | 16k | `▁ап ан ай ▁мәсете ▁( ▁) ▁— ▁ҡазан ▁мәсете ▁, ... (+16 more)` | 26 |
116
- | 32k | `▁ап ан ай ▁мәсете ▁( ▁) ▁— ▁ҡазан ▁мәсете ▁, ... (+15 more)` | 25 |
117
- | 64k | `▁ап ан ай ▁мәсете ▁( ▁) ▁— ▁ҡазан ▁мәсете ▁, ... (+14 more)` | 24 |
118
 
119
 
120
  ### Key Findings
121
 
122
- - **Best Compression:** 64k achieves 4.673x compression
123
- - **Lowest UNK Rate:** 8k with 0.3956% unknown tokens
124
  - **Trade-off:** Larger vocabularies improve compression but increase model size
125
  - **Recommendation:** 32k vocabulary provides optimal balance for production use
126
 
@@ -137,12 +147,14 @@ Below are sample sentences tokenized with each vocabulary size:
137
 
138
  | N-gram | Variant | Perplexity | Entropy | Unique N-grams | Top-100 Coverage | Top-1000 Coverage |
139
  |--------|---------|------------|---------|----------------|------------------|-------------------|
140
- | **2-gram** | Word | 56,525 | 15.79 | 433,408 | 13.8% | 30.4% |
141
- | **2-gram** | Subword | 489 🏆 | 8.93 | 13,769 | 52.3% | 96.8% |
142
- | **3-gram** | Word | 53,989 | 15.72 | 563,973 | 18.1% | 34.8% |
143
- | **3-gram** | Subword | 4,226 | 12.04 | 117,773 | 18.9% | 58.5% |
144
- | **4-gram** | Word | 61,817 | 15.92 | 883,766 | 19.4% | 36.8% |
145
- | **4-gram** | Subword | 21,528 | 14.39 | 687,383 | 10.2% | 33.2% |
 
 
146
 
147
  ### Top 5 N-grams by Size
148
 
@@ -154,7 +166,7 @@ Below are sample sentences tokenized with each vocabulary size:
154
  | 2 | `һыу реестры` | 40,405 |
155
  | 3 | `дәүләт һыу` | 40,403 |
156
  | 4 | `йылға бассейны` | 40,327 |
157
- | 5 | `рәсәй федерацияһы` | 37,241 |
158
 
159
  **3-grams (Word):**
160
 
@@ -163,55 +175,75 @@ Below are sample sentences tokenized with each vocabulary size:
163
  | 1 | `һыу реестры мәғлүмәттәре` | 20,323 |
164
  | 2 | `дәүләт һыу реестры` | 20,208 |
165
  | 3 | `рәсәй дәүләт һыу` | 20,202 |
166
- | 4 | `дәүләт һыу реестрында` | 20,168 |
167
- | 5 | `реестры мәғлүмәттәре рәсәй` | 20,167 |
168
 
169
  **4-grams (Word):**
170
 
171
  | Rank | N-gram | Count |
172
  |------|--------|-------|
173
  | 1 | `рәсәй дәүләт һыу реестры` | 20,195 |
174
- | 2 | `реестры мәғлүмәттәре рәсәй дәүләт` | 20,167 |
175
- | 3 | `мәғлүмәттәре рәсәй дәүләт һыу` | 20,167 |
176
- | 4 | `һыу реестры мәғлүмәттәре рәсәй` | 20,164 |
177
  | 5 | `дәүләт һыу реестрында һыу` | 20,160 |
178
 
 
 
 
 
 
 
 
 
 
 
179
  **2-grams (Subword):**
180
 
181
  | Rank | N-gram | Count |
182
  |------|--------|-------|
183
- | 1 | `а _` | 2,396,936 |
184
- | 2 | `а р` | 2,197,072 |
185
- | 3 | `ы _` | 2,104,654 |
186
- | 4 | `_ б` | 2,010,552 |
187
- | 5 | `а н` | 1,869,683 |
188
 
189
  **3-grams (Subword):**
190
 
191
  | Rank | N-gram | Count |
192
  |------|--------|-------|
193
- | 1 | `_ й ы` | 756,503 |
194
- | 2 | `й ы л` | 745,794 |
195
- | 3 | `н д а` | 679,041 |
196
- | 4 | `а н _` | 653,833 |
197
- | 5 | `ы ң _` | 648,174 |
198
 
199
  **4-grams (Subword):**
200
 
201
  | Rank | N-gram | Count |
202
  |------|--------|-------|
203
- | 1 | `_ й ы л` | 708,824 |
204
- | 2 | `ы н д а` | 469,174 |
205
- | 3 | `_ һ ә м` | 442,529 |
206
- | 4 | `һ ә м _` | 440,639 |
207
- | 5 | `н д а _` | 409,349 |
 
 
 
 
 
 
 
 
 
 
208
 
209
 
210
  ### Key Findings
211
 
212
- - **Best Perplexity:** 2-gram (subword) with 489
213
  - **Entropy Trend:** Decreases with larger n-grams (more predictable)
214
- - **Coverage:** Top-1000 patterns cover ~33% of corpus
215
  - **Recommendation:** 4-gram or 5-gram for best predictive performance
216
 
217
  ---
@@ -227,14 +259,14 @@ Below are sample sentences tokenized with each vocabulary size:
227
 
228
  | Context | Variant | Avg Entropy | Perplexity | Branching Factor | Unique Contexts | Predictability |
229
  |---------|---------|-------------|------------|------------------|-----------------|----------------|
230
- | **1** | Word | 0.8998 | 1.866 | 8.99 | 915,102 | 10.0% |
231
- | **1** | Subword | 0.9916 | 1.988 | 7.48 | 5,664 | 0.8% |
232
- | **2** | Word | 0.2745 | 1.210 | 1.74 | 8,225,491 | 72.6% |
233
- | **2** | Subword | 0.8603 | 1.815 | 5.91 | 42,359 | 14.0% |
234
- | **3** | Word | 0.0884 | 1.063 | 1.17 | 14,302,544 | 91.2% |
235
- | **3** | Subword | 0.8235 | 1.770 | 4.71 | 250,210 | 17.6% |
236
- | **4** | Word | 0.0321 🏆 | 1.022 | 1.05 | 16,653,317 | 96.8% |
237
- | **4** | Subword | 0.7025 | 1.627 | 3.37 | 1,177,702 | 29.8% |
238
 
239
  ### Generated Text Samples (Word-based)
240
 
@@ -242,27 +274,27 @@ Below are text samples generated from each word-based Markov chain model:
242
 
243
  **Context Size 1:**
244
 
245
- 1. `һәм инәйҙәре тәрбиәләп үҫтергәндәр улы сәғитов м стрельникова с григорьев а а преображенский верфенд...`
246
- 2. `буйынса ла бүлә көньяҡ диалекты там где плещется форель фильм үҙенең ҡатнашыуын ылыҡтыра йылда саҡыр...`
247
- 3. `һыу реестры мәғлүмәте буйынса асыш кубогын еңә йылдан гидромеханизация горных породах и любовь шевцо...`
248
 
249
  **Context Size 2:**
250
 
251
- 1. `гө буйынса һаны номеры 15 гө буйынса коды бассейн коды гө буйынса һаны номеры 03 гө буйынса`
252
- 2. `һыу реестры мәғлүмәте буйынса дәүләт һыу реестрында һыу объектының коды гидрологик өйрәнеү гө буйынс...`
253
- 3. `дәүләт һыу реестры мәғлүмәттәре рәсәй дәүләт һыу реестрында һыу объектының коды гидрологик өйрәнеү г...`
254
 
255
  **Context Size 3:**
256
 
257
- 1. `һыу реестры мәғлүмәттәре рәсәй дәүләт һыу реестры мәғлүмәте буйынса йылға түбәнге обь һыу бассейны о...`
258
- 2. `дәүләт һыу реестры мәғлүмәте буйынса йылға кама һыу бассейны округында урынлашҡан һыу хужалығы участ...`
259
- 3. `рәсәй дәүләт һыу реестры мәғлүмәте буйынса йылға кама һыу бассейны округында урынлашҡан һыу хужалығы...`
260
 
261
  **Context Size 4:**
262
 
263
- 1. `рәсәй дәүләт һыу реестры мәғлүмәте буйынса йылға кубань һыу бассейны округында урынлашҡан һыу хужалы...`
264
- 2. `реестры мәғлүмәттәре рәсәй дәүләт һыу реестры мәғлүмәте буйынса йылға көнбыйыш каспий һыу бассейны о...`
265
- 3. `мәғлүмәттәре рәсәй дәүләт һыу реестры мәғлүмәте буйынса йылға иртыш һыу бассейны округында урынлашҡа...`
266
 
267
 
268
  ### Generated Text Samples (Subword-based)
@@ -271,34 +303,34 @@ Below are text samples generated from each subword-based Markov chain model:
271
 
272
  **Context Size 1:**
273
 
274
- 1. `_бә_бемиәкәл)_өх`
275
- 2. `ацине_аҡъя_тенан`
276
- 3. `радъеле,_бесеср_`
277
 
278
  **Context Size 2:**
279
 
280
- 1. `а_тетыға_олкәр_ре`
281
- 2. `ар,_двинфүргеҙмәт`
282
- 3. `ы_былдағыный_мәһе`
283
 
284
  **Context Size 3:**
285
 
286
- 1. `_йылға_владионерҙә`
287
- 2. `йылға_бүләт_ил_ажн`
288
- 3. `нда_алек_тамблем,_`
289
 
290
  **Context Size 4:**
291
 
292
- 1. `_йылған_күпкә_ҡушыл`
293
- 2. `ында_ҡаршы_ҡустың_ү`
294
- 3. `_һәм_төрлө_метрында`
295
 
296
 
297
  ### Key Findings
298
 
299
  - **Best Predictability:** Context-4 (word) with 96.8% predictability
300
  - **Branching Factor:** Decreases with context size (more deterministic)
301
- - **Memory Trade-off:** Larger contexts require more storage (1,177,702 contexts)
302
  - **Recommendation:** Context-3 or Context-4 for text generation
303
 
304
  ---
@@ -314,48 +346,48 @@ Below are text samples generated from each subword-based Markov chain model:
314
 
315
  | Metric | Value |
316
  |--------|-------|
317
- | Vocabulary Size | 391,795 |
318
- | Total Tokens | 21,537,937 |
319
- | Mean Frequency | 54.97 |
320
  | Median Frequency | 4 |
321
- | Frequency Std Dev | 1228.27 |
322
 
323
  ### Most Common Words
324
 
325
  | Rank | Word | Frequency |
326
  |------|------|-----------|
327
- | 1 | һәм | 442,727 |
328
- | 2 | буйынса | 199,652 |
329
- | 3 | һыу | 168,369 |
330
- | 4 | менән | 154,690 |
331
- | 5 | йылға | 141,126 |
332
- | 6 | йылда | 136,417 |
333
- | 7 | рәсәй | 107,366 |
334
- | 8 | йыл | 97,537 |
335
- | 9 | йылдың | 89,696 |
336
- | 10 | в | 87,704 |
337
 
338
  ### Least Common Words (from vocabulary)
339
 
340
  | Rank | Word | Frequency |
341
  |------|------|-----------|
342
- | 1 | анкалаевҡа | 2 |
343
- | 2 | куцелаба | 2 |
344
- | 3 | хизарович | 2 |
345
- | 4 | чимаевтың | 2 |
346
- | 5 | уиттакерҙың | 2 |
347
- | 6 | дрикус | 2 |
348
- | 7 | шарабутдин | 2 |
349
- | 8 | rcc | 2 |
350
- | 9 | cosmetics | 2 |
351
- | 10 | kits | 2 |
352
 
353
  ### Zipf's Law Analysis
354
 
355
  | Metric | Value |
356
  |--------|-------|
357
- | Zipf Coefficient | 1.0493 |
358
- | R² (Goodness of Fit) | 0.992213 |
359
  | Adherence Quality | **excellent** |
360
 
361
  ### Coverage Analysis
@@ -365,13 +397,13 @@ Below are text samples generated from each subword-based Markov chain model:
365
  | Top 100 | 23.9% |
366
  | Top 1,000 | 52.3% |
367
  | Top 5,000 | 71.5% |
368
- | Top 10,000 | 78.5% |
369
 
370
  ### Key Findings
371
 
372
  - **Zipf Compliance:** R²=0.9922 indicates excellent adherence to Zipf's law
373
  - **High Frequency Dominance:** Top 100 words cover 23.9% of corpus
374
- - **Long Tail:** 381,795 words needed for remaining 21.5% coverage
375
 
376
  ---
377
  ## 5. Word Embeddings Evaluation
@@ -387,37 +419,40 @@ Below are text samples generated from each subword-based Markov chain model:
387
 
388
  ### 5.1 Cross-Lingual Alignment
389
 
390
- > *Note: Multilingual alignment visualization not available for this language.*
 
 
391
 
392
 
393
  ### 5.2 Model Comparison
394
 
395
  | Model | Dimension | Isotropy | Semantic Density | Alignment R@1 | Alignment R@10 |
396
  |-------|-----------|----------|------------------|---------------|----------------|
397
- | **mono_32d** | 32 | 0.7656 | 0.3637 | N/A | N/A |
398
- | **mono_64d** | 64 | 0.7751 🏆 | 0.2899 | N/A | N/A |
399
- | **mono_128d** | 128 | 0.7586 | 0.2211 | N/A | N/A |
 
 
 
400
 
401
  ### Key Findings
402
 
403
- - **Best Isotropy:** mono_64d with 0.7751 (more uniform distribution)
404
- - **Semantic Density:** Average pairwise similarity of 0.2916. Lower values indicate better semantic separation.
405
- - **Alignment Quality:** No aligned models evaluated in this run.
406
  - **Recommendation:** 128d aligned for best cross-lingual performance
407
 
408
  ---
409
  ## 6. Morphological Analysis (Experimental)
410
 
411
- > ⚠️ **Warning:** This language shows low morphological productivity. The statistical signals used for this analysis may be noisy or less reliable than for morphologically rich languages.
412
-
413
  This section presents an automated morphological analysis derived from the statistical divergence between word-level and subword-level models. By analyzing where subword predictability spikes and where word-level coverage fails, we can infer linguistic structures without supervised data.
414
 
415
  ### 6.1 Productivity & Complexity
416
 
417
  | Metric | Value | Interpretation | Recommendation |
418
  |--------|-------|----------------|----------------|
419
- | Productivity Index | **0.000** | Low morphological productivity | ⚠️ Likely unreliable |
420
- | Idiomaticity Gap | **-1.000** | Low formulaic content | - |
421
 
422
  ### 6.2 Affix Inventory (Productive Units)
423
 
@@ -430,11 +465,14 @@ These are the most productive prefixes and suffixes identified by sampling the v
430
  #### Productive Suffixes
431
  | Suffix | Examples |
432
  |--------|----------|
433
- | `-а` | симпозиумдарында, режица, оффенбаха |
434
- | `-ың` | тамаҡтың, ялкайндың, һуҙаһың |
435
- | `-ан` | ышанмаған, аҡсабан, гарнизондарынан |
436
- | `-ар` | стәрлетамаҡлылар, аныҡлаусылар, яндырылғандар |
437
- | `-ға` | ципрофлоксацинға, һауығырға, ҡыҫырыҡларға |
 
 
 
438
 
439
  ### 6.3 Bound Stems (Lexical Roots)
440
 
@@ -442,18 +480,18 @@ Bound stems are high-frequency subword units that are semantically cohesive but
442
 
443
  | Stem | Cohesion | Substitutability | Examples |
444
  |------|----------|------------------|----------|
445
- | `ассе` | 2.59x | 57 contexts | сассе, массе, гассе |
446
- | `ссей` | 3.05x | 29 contexts | бассей, шоссей, иессей |
447
- | `олог` | 1.87x | 205 contexts | лолог, молог, полог |
448
- | `арҙа` | 1.74x | 267 contexts | дарҙа, арҙан, барҙа |
449
- | `арҙы` | 1.79x | 169 contexts | шарҙы, сарҙы, ҡарҙы |
450
- | `лған` | 1.60x | 230 contexts | алған, ялған, ҡлған |
451
- | `шҡор` | 3.05x | 15 contexts | башҡор, башҡорд, башҡорт |
452
- | `ылға` | 1.57x | 213 contexts | йылға, тылға, ҡылға |
453
- | `йылғ` | 1.88x | 73 contexts | йылға, йылғы, уйылға |
454
- | `әрен` | 1.63x | 140 contexts | йәрен, кәрен, дәрен |
455
- | `дәүл` | 2.80x | 16 contexts | дәүли, дәүлә, дәүләт |
456
- | `әүлә` | 1.99x | 39 contexts | хәүлә, дәүлә, мәүлә |
457
 
458
  ### 6.4 Affix Compatibility (Co-occurrence)
459
 
@@ -468,26 +506,28 @@ Using **Recursive Hierarchical Substitutability**, we decompose complex words in
468
 
469
  | Word | Suggested Split | Confidence | Stem |
470
  |------|-----------------|------------|------|
471
- | биониканың | **`бионик-ан-ың`** | 6.0 | `бионик` |
472
- | худякованың | **`худяков-ан-ың`** | 6.0 | `худяков` |
473
- | воронкованың | **`воронков-ан-ың`** | 6.0 | `воронков` |
474
- | давыдованың | **`давыдов-ан-ың`** | 6.0 | `давыдов` |
475
- | фонеманың | **`фонем-ан-ың`** | 6.0 | `фонем` |
476
- | балаһынан | **`балаһын-ан`** | 4.5 | `балаһын` |
477
- | фламенкоға | **`фламенко-ға`** | 4.5 | `фламенко` |
478
- | топонимияһынан | **`топонимияһын-ан`** | 4.5 | `топонимияһын` |
479
- | баштарының | **`баштарын-ың`** | 4.5 | `баштарын` |
480
- | людмилаға | **`людмила-ға`** | 4.5 | `людмила` |
481
- | мозаикаға | **`мозаика-ға`** | 4.5 | `мозаика` |
482
- | орлеанскийға | **`орлеанский-ға`** | 4.5 | `орлеанский` |
483
- | манараларының | **`манараларын-ың`** | 4.5 | `манараларын` |
484
- | начальнигынан | **`начальнигын-ан`** | 4.5 | `начальнигын` |
485
- | кинофильмының | **`кинофильмын-ың`** | 4.5 | `кинофильмын` |
486
 
487
  ### 6.6 Linguistic Interpretation
488
 
489
  > **Automated Insight:**
490
- The language BA appears to be more isolating or has a highly fixed vocabulary. Word-level models perform nearly as well as subword models, indicating fewer productive morphological processes.
 
 
491
 
492
  ---
493
  ## 7. Summary & Recommendations
@@ -499,7 +539,7 @@ The language BA appears to be more isolating or has a highly fixed vocabulary. W
499
  | Component | Recommended | Rationale |
500
  |-----------|-------------|-----------|
501
  | Tokenizer | **64k BPE** | Best compression (4.67x) |
502
- | N-gram | **2-gram** | Lowest perplexity (489) |
503
  | Markov | **Context-4** | Highest predictability (96.8%) |
504
  | Embeddings | **100d** | Balanced semantic capture and isotropy |
505
 
@@ -714,4 +754,4 @@ MIT License - Free for academic and commercial use.
714
  ---
715
  *Generated by Wikilangs Models Pipeline*
716
 
717
- *Report Date: 2026-01-03 07:03:34*
 
1
  ---
2
  language: ba
3
+ language_name: Bashkir
4
  language_family: turkic_kipchak
5
  tags:
6
  - wikilangs
 
10
  - n-gram
11
  - markov
12
  - wikipedia
13
+ - feature-extraction
14
+ - sentence-similarity
15
+ - tokenization
16
+ - n-grams
17
+ - markov-chain
18
+ - text-mining
19
+ - fasttext
20
+ - babelvec
21
+ - vocabulous
22
+ - vocabulary
23
  - monolingual
24
  - family-turkic_kipchak
25
  license: mit
26
  library_name: wikilangs
27
+ pipeline_tag: text-generation
28
  datasets:
29
  - omarkamali/wikipedia-monthly
30
  dataset_info:
 
33
  metrics:
34
  - name: best_compression_ratio
35
  type: compression
36
+ value: 4.674
37
  - name: best_isotropy
38
  type: isotropy
39
+ value: 0.7711
40
  - name: vocabulary_size
41
  type: vocab
42
  value: 0
43
  generated: 2026-01-03
44
  ---
45
 
46
+ # Bashkir - Wikilangs Models
47
  ## Comprehensive Research Report & Full Ablation Study
48
 
49
+ This repository contains NLP models trained and evaluated by Wikilangs, specifically on **Bashkir** Wikipedia data.
50
  We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and word embeddings.
51
 
52
  ## 📋 Repository Contents
 
70
  - [3. Markov Chain Evaluation](#3-markov-chain-evaluation)
71
  - [4. Vocabulary Analysis](#4-vocabulary-analysis)
72
  - [5. Word Embeddings Evaluation](#5-word-embeddings-evaluation)
73
+ - [6. Morphological Analysis (Experimental)](#6--morphological-analysis-experimental)
74
  - [7. Summary & Recommendations](#7-summary--recommendations)
75
  - [Metrics Glossary](#appendix-metrics-glossary--interpretation-guide)
76
  - [Visualizations Index](#visualizations-index)
 
90
 
91
  | Vocab Size | Compression | Avg Token Len | UNK Rate | Total Tokens |
92
  |------------|-------------|---------------|----------|--------------|
93
+ | **8k** | 3.561x | 3.56 | 0.3982% | 1,530,967 |
94
+ | **16k** | 3.999x | 4.00 | 0.4471% | 1,363,432 |
95
+ | **32k** | 4.374x | 4.38 | 0.4891% | 1,246,440 |
96
+ | **64k** | 4.674x 🏆 | 4.68 | 0.5226% | 1,166,431 |
97
 
98
  ### Tokenization Examples
99
 
100
  Below are sample sentences tokenized with each vocabulary size:
101
 
102
+ **Sample 1:** `Нортленд - (ҡитға исеме) лағы дәүләт. Иҫкәрмәләр Һылтанмалар`
103
 
104
  | Vocab | Tokens | Count |
105
  |-------|--------|-------|
106
+ | 8k | `▁н орт лен д ▁- ▁( ҡит ға ▁исеме ) ... (+6 more)` | 16 |
107
+ | 16k | `▁н орт ленд ▁- ▁( ҡит ға ▁исеме ) ▁лағы ... (+4 more)` | 14 |
108
+ | 32k | `▁н орт ленд ▁- ▁( ҡитға ▁исеме ) ▁лағы ▁дәүләт ... (+3 more)` | 13 |
109
+ | 64k | `▁норт ленд ▁- ▁( ҡитға ▁исеме ) ▁лағы ▁дәүләт . ... (+2 more)` | 12 |
110
 
111
+ **Sample 2:** `АвстралияКөньяҡ ярымшарҙарҙа урынлашҡан дәүләт. Австралия (ҡитға) Көнсығыш ...`
112
 
113
  | Vocab | Tokens | Count |
114
  |-------|--------|-------|
115
+ | 8k | `▁австр алия ▁— ▁көньяҡ ▁ярым шар ҙарҙа ▁урынлашҡан ▁дәүләт . ... (+18 more)` | 28 |
116
+ | 16k | `▁австралия ▁��� ▁көньяҡ ▁ярымшар ҙарҙа ▁урынлашҡан ▁дәүләт . ▁австралия ▁( ... (+13 more)` | 23 |
117
+ | 32k | `▁австралия ▁— ▁көньяҡ ▁ярымшар ҙарҙа ▁урынлашҡан ▁дәүләт . ▁австралия ▁( ... (+11 more)` | 21 |
118
+ | 64k | `▁австралия ▁— ▁көньяҡ ▁ярымшар ҙарҙа ▁урынлашҡан ▁дәүләт . ▁австралия ▁( ... (+11 more)` | 21 |
119
 
120
+ **Sample 3:** `йыл йәкшәмбе көнөнән башланған йыл, кәбисә түгел. Ваҡиғалар Тыуғандар Вафат бу...`
121
 
122
  | Vocab | Tokens | Count |
123
  |-------|--------|-------|
124
+ | 8k | `▁йыл ▁— ▁й әк шәмбе ▁көнөнән ▁башланған ▁йыл , ▁кәбисә ... (+10 more)` | 20 |
125
+ | 16k | `▁йыл ▁— ▁йәкшәмбе ▁көнөнән ▁башланған ▁йыл , ▁кәбисә ▁түгел . ... (+8 more)` | 18 |
126
+ | 32k | `▁йыл ▁— ▁йәкшәмбе ▁көнөнән ▁башланған ▁йыл , ▁кәбисә ▁түгел . ... (+8 more)` | 18 |
127
+ | 64k | `▁йыл ▁— ▁йәкшәмбе ▁көнөнән ▁башланған ▁йыл , ▁кәбисә ▁түгел . ... (+8 more)` | 18 |
128
 
129
 
130
  ### Key Findings
131
 
132
+ - **Best Compression:** 64k achieves 4.674x compression
133
+ - **Lowest UNK Rate:** 8k with 0.3982% unknown tokens
134
  - **Trade-off:** Larger vocabularies improve compression but increase model size
135
  - **Recommendation:** 32k vocabulary provides optimal balance for production use
136
 
 
147
 
148
  | N-gram | Variant | Perplexity | Entropy | Unique N-grams | Top-100 Coverage | Top-1000 Coverage |
149
  |--------|---------|------------|---------|----------------|------------------|-------------------|
150
+ | **2-gram** | Word | 56,272 | 15.78 | 432,191 | 13.8% | 30.4% |
151
+ | **2-gram** | Subword | 488 🏆 | 8.93 | 13,737 | 52.3% | 96.8% |
152
+ | **3-gram** | Word | 53,798 | 15.72 | 562,854 | 18.1% | 34.8% |
153
+ | **3-gram** | Subword | 4,221 | 12.04 | 117,501 | 18.9% | 58.6% |
154
+ | **4-gram** | Word | 61,592 | 15.91 | 881,988 | 19.4% | 36.9% |
155
+ | **4-gram** | Subword | 21,484 | 14.39 | 685,600 | 10.3% | 33.2% |
156
+ | **5-gram** | Word | 37,893 | 15.21 | 658,444 | 21.5% | 41.3% |
157
+ | **5-gram** | Subword | 72,234 | 16.14 | 2,075,140 | 7.0% | 23.5% |
158
 
159
  ### Top 5 N-grams by Size
160
 
 
166
  | 2 | `һыу реестры` | 40,405 |
167
  | 3 | `дәүләт һыу` | 40,403 |
168
  | 4 | `йылға бассейны` | 40,327 |
169
+ | 5 | `рәсәй федерацияһы` | 37,239 |
170
 
171
  **3-grams (Word):**
172
 
 
175
  | 1 | `һыу реестры мәғлүмәттәре` | 20,323 |
176
  | 2 | `дәүләт һыу реестры` | 20,208 |
177
  | 3 | `рәсәй дәүләт һыу` | 20,202 |
178
+ | 4 | `мәғлүмәттәре рәсәй дәүләт` | 20,170 |
179
+ | 5 | `реестры мәғлүмәттәре рәсәй` | 20,170 |
180
 
181
  **4-grams (Word):**
182
 
183
  | Rank | N-gram | Count |
184
  |------|--------|-------|
185
  | 1 | `рәсәй дәүләт һыу реестры` | 20,195 |
186
+ | 2 | `реестры мәғлүмәттәре рәсәй дәүләт` | 20,170 |
187
+ | 3 | `мәғлүмәттәре рәсәй дәүләт һыу` | 20,170 |
188
+ | 4 | `һыу реестры мәғлүмәттәре рәсәй` | 20,167 |
189
  | 5 | `дәүләт һыу реестрында һыу` | 20,160 |
190
 
191
+ **5-grams (Word):**
192
+
193
+ | Rank | N-gram | Count |
194
+ |------|--------|-------|
195
+ | 1 | `реестры мәғлүмәттәре рәсәй дәүләт һыу` | 20,170 |
196
+ | 2 | `һыу реестры мәғлүмәттәре рәсәй дәүләт` | 20,167 |
197
+ | 3 | `мәғлүмәттәре рәсәй дәүләт һыу реестры` | 20,165 |
198
+ | 4 | `һыу реестрында һыу объектының коды` | 20,156 |
199
+ | 5 | `дәүләт һыу реестрында һыу объектының` | 20,156 |
200
+
201
  **2-grams (Subword):**
202
 
203
  | Rank | N-gram | Count |
204
  |------|--------|-------|
205
+ | 1 | `а _` | 2,391,231 |
206
+ | 2 | `а р` | 2,191,202 |
207
+ | 3 | `ы _` | 2,097,776 |
208
+ | 4 | `_ б` | 2,006,204 |
209
+ | 5 | `а н` | 1,864,458 |
210
 
211
  **3-grams (Subword):**
212
 
213
  | Rank | N-gram | Count |
214
  |------|--------|-------|
215
+ | 1 | `_ й ы` | 754,633 |
216
+ | 2 | `й ы л` | 743,969 |
217
+ | 3 | `н д а` | 676,936 |
218
+ | 4 | `а н _` | 651,892 |
219
+ | 5 | `ы ң _` | 646,394 |
220
 
221
  **4-grams (Subword):**
222
 
223
  | Rank | N-gram | Count |
224
  |------|--------|-------|
225
+ | 1 | `_ й ы л` | 707,090 |
226
+ | 2 | `ы н д а` | 467,625 |
227
+ | 3 | `_ һ ә м` | 441,510 |
228
+ | 4 | `һ ә м _` | 439,610 |
229
+ | 5 | `н д а _` | 408,202 |
230
+
231
+ **5-grams (Subword):**
232
+
233
+ | Rank | N-gram | Count |
234
+ |------|--------|-------|
235
+ | 1 | `_ һ ә м _` | 438,718 |
236
+ | 2 | `ы н д а _` | 353,882 |
237
+ | 3 | `_ й ы л д` | 323,522 |
238
+ | 4 | `й ы л ғ а` | 269,201 |
239
+ | 5 | `_ й ы л ғ` | 262,857 |
240
 
241
 
242
  ### Key Findings
243
 
244
+ - **Best Perplexity:** 2-gram (subword) with 488
245
  - **Entropy Trend:** Decreases with larger n-grams (more predictable)
246
+ - **Coverage:** Top-1000 patterns cover ~23% of corpus
247
  - **Recommendation:** 4-gram or 5-gram for best predictive performance
248
 
249
  ---
 
259
 
260
  | Context | Variant | Avg Entropy | Perplexity | Branching Factor | Unique Contexts | Predictability |
261
  |---------|---------|-------------|------------|------------------|-----------------|----------------|
262
+ | **1** | Word | 0.8991 | 1.865 | 8.98 | 912,874 | 10.1% |
263
+ | **1** | Subword | 0.9900 | 1.986 | 7.47 | 5,662 | 1.0% |
264
+ | **2** | Word | 0.2746 | 1.210 | 1.74 | 8,193,331 | 72.5% |
265
+ | **2** | Subword | 0.8598 | 1.815 | 5.90 | 42,271 | 14.0% |
266
+ | **3** | Word | 0.0885 | 1.063 | 1.17 | 14,249,949 | 91.1% |
267
+ | **3** | Subword | 0.8239 | 1.770 | 4.71 | 249,519 | 17.6% |
268
+ | **4** | Word | 0.0321 🏆 | 1.023 | 1.05 | 16,595,241 | 96.8% |
269
+ | **4** | Subword | 0.7025 | 1.627 | 3.37 | 1,174,607 | 29.7% |
270
 
271
  ### Generated Text Samples (Word-based)
272
 
 
274
 
275
  **Context Size 1:**
276
 
277
+ 1. `һәм пәйғәмбәр аша ҡулға алалар диск ҡалын һуҙынҡылы ижеккә төшә көнбайыш конференцияһын әҙерләүҙә ул...`
278
+ 2. `буйынса журналистар үҙҙәрен римляндар өсөн рәссам булараҡ игорь задорожный игорь а сатаров в н г сах...`
279
+ 3. `һыу һәм төрлө биҙәгәндәр был блюдоның консистенцияһында исеменең типовой проект ҡаты алыштарҙа дошма...`
280
 
281
  **Context Size 2:**
282
 
283
+ 1. `гө буйынса сығарылыш 2 фаунаһы йылға мәғлүмәттәр буйынса аҙсылыҡтан император гвардияһы училищеһында...`
284
+ 2. `һыу реестры мәғлүмәттәре рәсәй дәүләт һыу реестры мәғлүмәттәре рәсәй дәүләт һыу реестрында һыу объек...`
285
+ 3. `дәүләт һыу реестры мәғлүмәттәре рәсәй дәүләт өлгөһөндәге диплом осоу аппараттарын ҡулланыуҙы көйләү ...`
286
 
287
  **Context Size 3:**
288
 
289
+ 1. `һыу реестры мәғлүмәттәре рәсәй дәүләт һыу реестры мәғлүмәте буйынса йылға двина печора һыу бассейны ...`
290
+ 2. `дәүләт һыу реестры мәғлүмәте буйынса йылға двина печора һыу бассейны округында урынлашҡан һыу хужалы...`
291
+ 3. `рәсәй дәүләт һыу реестры мәғлүмәте буйынса йылға кама һыу һаклағысы чусов сылвин ҡултығы һул ярына т...`
292
 
293
  **Context Size 4:**
294
 
295
+ 1. `рәсәй дәүләт һыу реестры мәғлүмәте буйынса йылға кама һыу бассейны округында урынлашҡан һыу хужалығы...`
296
+ 2. `мәғлүмәттәре рәсәй дәүләт һыу реестры мәғлүмәте буйынса йылға көнбыйыш каспий һыу бассейны округында...`
297
+ 3. `реестры мәғлүмәттәре рәсәй дәүләт һыу реестры мәғлүмәте буйынса йылға кама һыу бассейны округында ур...`
298
 
299
 
300
  ### Generated Text Samples (Subword-based)
 
303
 
304
  **Context Size 1:**
305
 
306
+ 1. `_хрм_илға._—_брл`
307
+ 2. `атемлашылларулең`
308
+ 3. `ралүеүмәмка_ты_а`
309
 
310
  **Context Size 2:**
311
 
312
+ 1. `а_ра_һуң_съ_идери`
313
+ 2. `ар._энты_хайындат`
314
+ 3. `ы__буягацияһальс`
315
 
316
  **Context Size 3:**
317
 
318
+ 1. `_йыл_17_дек_тип_ик`
319
+ 2. `йылдығыштабыуат_ге`
320
+ 3. `ндағы_мәғилми_хеҙм`
321
 
322
  **Context Size 4:**
323
 
324
+ 1. `_йылдан_булат_ҡулты`
325
+ 2. `ындағы_ҡарағыҙ_барғ`
326
+ 3. `_һәм_бөтә_советы,_п`
327
 
328
 
329
  ### Key Findings
330
 
331
  - **Best Predictability:** Context-4 (word) with 96.8% predictability
332
  - **Branching Factor:** Decreases with context size (more deterministic)
333
+ - **Memory Trade-off:** Larger contexts require more storage (1,174,607 contexts)
334
  - **Recommendation:** Context-3 or Context-4 for text generation
335
 
336
  ---
 
346
 
347
  | Metric | Value |
348
  |--------|-------|
349
+ | Vocabulary Size | 390,661 |
350
+ | Total Tokens | 21,477,387 |
351
+ | Mean Frequency | 54.98 |
352
  | Median Frequency | 4 |
353
+ | Frequency Std Dev | 1227.90 |
354
 
355
  ### Most Common Words
356
 
357
  | Rank | Word | Frequency |
358
  |------|------|-----------|
359
+ | 1 | һәм | 441,701 |
360
+ | 2 | буйынса | 199,502 |
361
+ | 3 | һыу | 168,327 |
362
+ | 4 | менән | 154,212 |
363
+ | 5 | йылға | 141,020 |
364
+ | 6 | йылда | 136,113 |
365
+ | 7 | рәсәй | 107,301 |
366
+ | 8 | йыл | 96,991 |
367
+ | 9 | йылдың | 89,541 |
368
+ | 10 | бассейны | 87,464 |
369
 
370
  ### Least Common Words (from vocabulary)
371
 
372
  | Rank | Word | Frequency |
373
  |------|------|-----------|
374
+ | 1 | совкомбанк | 2 |
375
+ | 2 | маркетплейстың | 2 |
376
+ | 3 | суларға | 2 |
377
+ | 4 | кишлак | 2 |
378
+ | 5 | пацанский | 2 |
379
+ | 6 | мунден | 2 |
380
+ | 7 | гертфордшир | 2 |
381
+ | 8 | кроуға | 2 |
382
+ | 9 | франклоу | 2 |
383
+ | 10 | алтынкүлдән | 2 |
384
 
385
  ### Zipf's Law Analysis
386
 
387
  | Metric | Value |
388
  |--------|-------|
389
+ | Zipf Coefficient | 1.0499 |
390
+ | R² (Goodness of Fit) | 0.992209 |
391
  | Adherence Quality | **excellent** |
392
 
393
  ### Coverage Analysis
 
397
  | Top 100 | 23.9% |
398
  | Top 1,000 | 52.3% |
399
  | Top 5,000 | 71.5% |
400
+ | Top 10,000 | 78.6% |
401
 
402
  ### Key Findings
403
 
404
  - **Zipf Compliance:** R²=0.9922 indicates excellent adherence to Zipf's law
405
  - **High Frequency Dominance:** Top 100 words cover 23.9% of corpus
406
+ - **Long Tail:** 380,661 words needed for remaining 21.4% coverage
407
 
408
  ---
409
  ## 5. Word Embeddings Evaluation
 
419
 
420
  ### 5.1 Cross-Lingual Alignment
421
 
422
+ ![Alignment Quality](visualizations/embedding_alignment_quality.png)
423
+
424
+ ![Multilingual t-SNE](visualizations/embedding_tsne_multilingual.png)
425
 
426
 
427
  ### 5.2 Model Comparison
428
 
429
  | Model | Dimension | Isotropy | Semantic Density | Alignment R@1 | Alignment R@10 |
430
  |-------|-----------|----------|------------------|---------------|----------------|
431
+ | **mono_32d** | 32 | 0.7605 | 0.3607 | N/A | N/A |
432
+ | **mono_64d** | 64 | 0.7711 🏆 | 0.2817 | N/A | N/A |
433
+ | **mono_128d** | 128 | 0.7589 | 0.2238 | N/A | N/A |
434
+ | **aligned_32d** | 32 | 0.7605 | 0.3651 | 0.0420 | 0.2620 |
435
+ | **aligned_64d** | 64 | 0.7711 | 0.2829 | 0.0820 | 0.3600 |
436
+ | **aligned_128d** | 128 | 0.7589 | 0.2231 | 0.1140 | 0.4340 |
437
 
438
  ### Key Findings
439
 
440
+ - **Best Isotropy:** mono_64d with 0.7711 (more uniform distribution)
441
+ - **Semantic Density:** Average pairwise similarity of 0.2896. Lower values indicate better semantic separation.
442
+ - **Alignment Quality:** Aligned models achieve up to 11.4% R@1 in cross-lingual retrieval.
443
  - **Recommendation:** 128d aligned for best cross-lingual performance
444
 
445
  ---
446
  ## 6. Morphological Analysis (Experimental)
447
 
 
 
448
  This section presents an automated morphological analysis derived from the statistical divergence between word-level and subword-level models. By analyzing where subword predictability spikes and where word-level coverage fails, we can infer linguistic structures without supervised data.
449
 
450
  ### 6.1 Productivity & Complexity
451
 
452
  | Metric | Value | Interpretation | Recommendation |
453
  |--------|-------|----------------|----------------|
454
+ | Productivity Index | **5.000** | High morphological productivity | Reliable analysis |
455
+ | Idiomaticity Gap | **0.762** | High formulaic/idiomatic content | - |
456
 
457
  ### 6.2 Affix Inventory (Productive Units)
458
 
 
465
  #### Productive Suffixes
466
  | Suffix | Examples |
467
  |--------|----------|
468
+ | `-а` | менеджерҙарға, клиентела, пегаға |
469
+ | `-ың` | амфитеатрының, лединың, ғәлиәкбәровтың |
470
+ | `-ан` | ҡыҙылдарҙан, саутунан, сығылған |
471
+ | `-ар` | андекстар, имплантаттар, тартыуҙар |
472
+ | `-ға` | менеджерҙарға, пегаға, ҡалыуға |
473
+ | `-ның` | амфитеатрының, лединың, соустарының |
474
+ | `-на` | градина, ағзаһына, катилина |
475
+ | `-ов` | крестов, әбшәрипов, протезов |
476
 
477
  ### 6.3 Bound Stems (Lexical Roots)
478
 
 
480
 
481
  | Stem | Cohesion | Substitutability | Examples |
482
  |------|----------|------------------|----------|
483
+ | `ссей` | 3.12x | 29 contexts | шоссей, иессей, бассей |
484
+ | `олог` | 1.84x | 205 contexts | лолог, полог, молог |
485
+ | `әүлә` | 2.51x | 39 contexts | дәүлә, хәүлә, шәүлә |
486
+ | `ассе` | 2.28x | 57 contexts | массе, хассе, гассе |
487
+ | `шҡор` | 3.03x | 15 contexts | башҡор, башҡорт, башҡорд |
488
+ | `лған` | 1.54x | 230 contexts | ялған, ҡлған, алған |
489
+ | `арҙы` | 1.62x | 168 contexts | парҙы, сарҙы, барҙы |
490
+ | `арҙа` | 1.48x | 266 contexts | барҙа, арҙан, арҙат |
491
+ | `аһын` | 1.35x | 378 contexts | шаһын, анаһын, яһаһын |
492
+ | `ттар` | 1.37x | 344 contexts | аттар, юттар, ттары |
493
+ | `ылға` | 1.49x | 213 contexts | йылға, ҡылға, ылғал |
494
+ | `лдар` | 1.45x | 236 contexts | алдар, ялдар, улдар |
495
 
496
  ### 6.4 Affix Compatibility (Co-occurrence)
497
 
 
506
 
507
  | Word | Suggested Split | Confidence | Stem |
508
  |------|-----------------|------------|------|
509
+ | александровна | **`александр-ов-на`** | 6.0 | `александр` |
510
+ | мессинаның | **`месси-на-ның`** | 6.0 | `месси` |
511
+ | салаватовна | **`салават-ов-на`** | 6.0 | `салават` |
512
+ | терракотанан | **`терракот-ан-ан`** | 6.0 | `терракот` |
513
+ | моденаның | **`моде-на-ның`** | 6.0 | `моде` |
514
+ | доломанов | **`долом-ан-ов`** | 6.0 | `долом` |
515
+ | склонениеһына | **`склонениеһы-на`** | 4.5 | `склонениеһы` |
516
+ | характеров | **`характер-ов`** | 4.5 | `характер` |
517
+ | ваҡытының | **`ваҡыты-ның`** | 4.5 | `ваҡыты` |
518
+ | кейекбайға | **`кейекбай-ға`** | 4.5 | `кейекбай` |
519
+ | фомичёваның | **`фомичёва-ның`** | 4.5 | `фомичёва` |
520
+ | никаноров | **`никанор-ов`** | 4.5 | `никанор` |
521
+ | терапияһынан | **`терапияһын-ан`** | 4.5 | `терапияһын` |
522
+ | телевидениеһынан | **`телевидениеһын-ан`** | 4.5 | `телевидениеһын` |
523
+ | сепаратизмына | **`сепаратизмы-на`** | 4.5 | `сепаратизмы` |
524
 
525
  ### 6.6 Linguistic Interpretation
526
 
527
  > **Automated Insight:**
528
+ The language Bashkir shows high morphological productivity. The subword models are significantly more efficient than word models, suggesting a rich system of affixation or compounding.
529
+
530
+ > **Note on Idiomaticity:** The high Idiomaticity Gap suggests a large number of frequent multi-word expressions or formulaic sequences that are statistically distinct from their component parts.
531
 
532
  ---
533
  ## 7. Summary & Recommendations
 
539
  | Component | Recommended | Rationale |
540
  |-----------|-------------|-----------|
541
  | Tokenizer | **64k BPE** | Best compression (4.67x) |
542
+ | N-gram | **2-gram** | Lowest perplexity (488) |
543
  | Markov | **Context-4** | Highest predictability (96.8%) |
544
  | Embeddings | **100d** | Balanced semantic capture and isotropy |
545
 
 
754
  ---
755
  *Generated by Wikilangs Models Pipeline*
756
 
757
+ *Report Date: 2026-01-03 20:08:48*
models/embeddings/aligned/ba_128d.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:12a29b348d60360d3be96399c050ef0d573fa9c305afe29ec2ba99c0b25e86d7
3
+ size 1266971997
models/embeddings/aligned/ba_128d.meta.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lang": "ba", "dim": 128, "max_seq_len": 512, "is_aligned": true}
models/embeddings/aligned/ba_128d.projection.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af0641c54d752ef68dc525ce684e8836ff9b10f2d1609adfef93a82936b57a09
3
+ size 65664
models/embeddings/aligned/ba_128d_metadata.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "language": "ba",
3
+ "dimension": 128,
4
+ "version": "aligned",
5
+ "hub_language": "en",
6
+ "seed_vocab_size": 15145,
7
+ "vocab_size": 231493
8
+ }
models/embeddings/aligned/ba_32d.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d0d97157e963bf05a6754dff717f8f828cdd9be148c364c808f7fe899456ea8e
3
+ size 321185373
models/embeddings/aligned/ba_32d.meta.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lang": "ba", "dim": 32, "max_seq_len": 512, "is_aligned": true}
models/embeddings/aligned/ba_32d.projection.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b9c02343822dac63c89696460b3e6beb5f6299333acf7d38e9eaf3c10fce943
3
+ size 4224
models/embeddings/aligned/ba_32d_metadata.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "language": "ba",
3
+ "dimension": 32,
4
+ "version": "aligned",
5
+ "hub_language": "en",
6
+ "seed_vocab_size": 15145,
7
+ "vocab_size": 231493
8
+ }
models/embeddings/aligned/ba_64d.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a0094447a82217960151cdcc0188313727cbed1ce26b4e49cce3e82acb2434b
3
+ size 636447581
models/embeddings/aligned/ba_64d.meta.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lang": "ba", "dim": 64, "max_seq_len": 512, "is_aligned": true}
models/embeddings/aligned/ba_64d.projection.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18442ca0c847bcc3101e4a2e0536324f2844255bdce426a9095209ca4c8b3869
3
+ size 16512
models/embeddings/aligned/ba_64d_metadata.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "language": "ba",
3
+ "dimension": 64,
4
+ "version": "aligned",
5
+ "hub_language": "en",
6
+ "seed_vocab_size": 15145,
7
+ "vocab_size": 231493
8
+ }
models/embeddings/monolingual/ba_128d.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1d0506e69046f5b9eeed00571257333963c93985c08aaff04359a63f9e18966d
3
- size 1267773201
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:12a29b348d60360d3be96399c050ef0d573fa9c305afe29ec2ba99c0b25e86d7
3
+ size 1266971997
models/embeddings/monolingual/ba_128d_metadata.json CHANGED
@@ -11,5 +11,5 @@
11
  "encoding_method": "rope",
12
  "dim": 128
13
  },
14
- "vocab_size": 232257
15
  }
 
11
  "encoding_method": "rope",
12
  "dim": 128
13
  },
14
+ "vocab_size": 231493
15
  }
models/embeddings/monolingual/ba_32d.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:857e0d1d2801efad1ea0ad2269a5064780058e0457d0dd04fc3e747f514bb814
3
- size 321399825
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d0d97157e963bf05a6754dff717f8f828cdd9be148c364c808f7fe899456ea8e
3
+ size 321185373
models/embeddings/monolingual/ba_32d_metadata.json CHANGED
@@ -11,5 +11,5 @@
11
  "encoding_method": "rope",
12
  "dim": 32
13
  },
14
- "vocab_size": 232257
15
  }
 
11
  "encoding_method": "rope",
12
  "dim": 32
13
  },
14
+ "vocab_size": 231493
15
  }
models/embeddings/monolingual/ba_64d.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a68b56475da9d52e23c1e8987aa18959fc1800f3fc2373b5aa026f7005dc4bd0
3
- size 636857617
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a0094447a82217960151cdcc0188313727cbed1ce26b4e49cce3e82acb2434b
3
+ size 636447581
models/embeddings/monolingual/ba_64d_metadata.json CHANGED
@@ -11,5 +11,5 @@
11
  "encoding_method": "rope",
12
  "dim": 64
13
  },
14
- "vocab_size": 232257
15
  }
 
11
  "encoding_method": "rope",
12
  "dim": 64
13
  },
14
+ "vocab_size": 231493
15
  }
models/subword_markov/ba_markov_ctx1_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:72ab68634a6f78c810b9cb22d356ab586e0001dca90d89cea677632b627a61ff
3
- size 321684
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05780301a06d04162048cbd9c5a2a6395d23596df7dca2fa2d56cfc1b0702d0e
3
+ size 321688
models/subword_markov/ba_markov_ctx1_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 1,
3
  "variant": "subword",
4
  "language": "ba",
5
- "unique_contexts": 5664,
6
- "total_transitions": 166306994
7
  }
 
2
  "context_size": 1,
3
  "variant": "subword",
4
  "language": "ba",
5
+ "unique_contexts": 5662,
6
+ "total_transitions": 165847857
7
  }
models/subword_markov/ba_markov_ctx2_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:64f356772a4c2f2fe17d0b3ced4a871083815f3306cc44a66e5ae43c45a11b68
3
- size 2038058
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6311fb380fbf48483186aa21a01b6a8c6c2d3500e66dded64a04cddf440446c
3
+ size 2056814
models/subword_markov/ba_markov_ctx2_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 2,
3
  "variant": "subword",
4
  "language": "ba",
5
- "unique_contexts": 42359,
6
- "total_transitions": 166243136
7
  }
 
2
  "context_size": 2,
3
  "variant": "subword",
4
  "language": "ba",
5
+ "unique_contexts": 42271,
6
+ "total_transitions": 165783993
7
  }
models/subword_markov/ba_markov_ctx3_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:034412812ae35db516b6f169d1eaad99c3412d09b2b6b86dd34d6c091ce4cc35
3
- size 10033851
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9668dbdc03969a68b32456e92f4601e53e3dd5f7352901212864f6224d20606c
3
+ size 9997997
models/subword_markov/ba_markov_ctx3_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 3,
3
  "variant": "subword",
4
  "language": "ba",
5
- "unique_contexts": 250210,
6
- "total_transitions": 166179278
7
  }
 
2
  "context_size": 3,
3
  "variant": "subword",
4
  "language": "ba",
5
+ "unique_contexts": 249519,
6
+ "total_transitions": 165720129
7
  }
models/subword_markov/ba_markov_ctx4_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:de9d856639518fda8614b7147be3a51f07dc47b782f251d16d66f550cd28ed7e
3
- size 34335118
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3d06eba3c693311b5366241731cc4966871a6c3d4c58908428352caa8dba517
3
+ size 34310629
models/subword_markov/ba_markov_ctx4_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 4,
3
  "variant": "subword",
4
  "language": "ba",
5
- "unique_contexts": 1177702,
6
- "total_transitions": 166115420
7
  }
 
2
  "context_size": 4,
3
  "variant": "subword",
4
  "language": "ba",
5
+ "unique_contexts": 1174607,
6
+ "total_transitions": 165656265
7
  }
models/subword_ngram/ba_2gram_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d8bc5caa63739eb8bd546faf8fd400a3cfed2102cf28c5d178636fd18c4a8540
3
- size 193332
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a93f1b12f868221be998258f1e52c5b7b8405d8d4f993ff205e853c0b5ef530d
3
+ size 192976
models/subword_ngram/ba_2gram_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "n": 2,
3
  "variant": "subword",
4
  "language": "ba",
5
- "unique_ngrams": 13769,
6
- "total_ngrams": 166306994
7
  }
 
2
  "n": 2,
3
  "variant": "subword",
4
  "language": "ba",
5
+ "unique_ngrams": 13737,
6
+ "total_ngrams": 165847857
7
  }
models/subword_ngram/ba_3gram_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:92068c289061bb2e93b829c142f737236f826c376deac540afd5cf5264596713
3
- size 1542627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5427aac4ff333f8f5a4f28484b7317d987f7a3937ec69023d125b8904c873c46
3
+ size 1525757
models/subword_ngram/ba_3gram_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "n": 3,
3
  "variant": "subword",
4
  "language": "ba",
5
- "unique_ngrams": 117773,
6
- "total_ngrams": 166243136
7
  }
 
2
  "n": 3,
3
  "variant": "subword",
4
  "language": "ba",
5
+ "unique_ngrams": 117501,
6
+ "total_ngrams": 165783993
7
  }
models/subword_ngram/ba_4gram_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:829baeb29467446ba03ac3c2e0fd8e4ce93634a4b5989afa2f4cbad92beba6d0
3
- size 9034722
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2bf52e0ecfe5c48dec7032cfbbaa99b84f3f01a3301dad1d6ae5fe0eb6a8340a
3
+ size 9017225
models/subword_ngram/ba_4gram_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "n": 4,
3
  "variant": "subword",
4
  "language": "ba",
5
- "unique_ngrams": 687383,
6
- "total_ngrams": 166179278
7
  }
 
2
  "n": 4,
3
  "variant": "subword",
4
  "language": "ba",
5
+ "unique_ngrams": 685600,
6
+ "total_ngrams": 165720129
7
  }
models/subword_ngram/ba_5gram_subword.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aab260ec4837b34b3579c5030290b4706666ed96ef44f2518b9aefa979d9195b
3
+ size 28891181
models/subword_ngram/ba_5gram_subword_metadata.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "n": 5,
3
+ "variant": "subword",
4
+ "language": "ba",
5
+ "unique_ngrams": 2075140,
6
+ "total_ngrams": 165656265
7
+ }
models/tokenizer/ba_tokenizer_16k.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a303c0d026bb9c2d55bf88aea6c23a3e9ee5479dbe177e187ac24e298d72853f
3
- size 587702
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2d289f5cd313f9621d08c6ffaccc42568a25a90548378a3deff45c01d188d41
3
+ size 587842
models/tokenizer/ba_tokenizer_16k.vocab CHANGED
The diff for this file is too large to render. See raw diff
 
models/tokenizer/ba_tokenizer_32k.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c53247b474ba1fb2bc81a96123dbc8edfeb1a2aa07072d220e4422014f9aa0a7
3
- size 956108
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0975ba054b867292078821676618b968c848836ae76a2358bee4c8cedc1706e6
3
+ size 956601
models/tokenizer/ba_tokenizer_32k.vocab CHANGED
The diff for this file is too large to render. See raw diff
 
models/tokenizer/ba_tokenizer_64k.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b92b6043e85b598f61bfc865f4cea415a1ba6735c830a01264fe91b74338e8a0
3
- size 1712379
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9674a907e50f3fa78f571f6c44e1791be0d3a5dffe071dda527390b641f15bd4
3
+ size 1713363
models/tokenizer/ba_tokenizer_64k.vocab CHANGED
The diff for this file is too large to render. See raw diff
 
models/tokenizer/ba_tokenizer_8k.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ebdbb6ac25e572b302f21c6cd6328ec1606feb430f390d602f0e1d7add179ee0
3
- size 408777
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c4236b291567cfc0b86344e376dedc15dd457adb983af1294a4896a6a7a80ae
3
+ size 408799
models/tokenizer/ba_tokenizer_8k.vocab CHANGED
The diff for this file is too large to render. See raw diff
 
models/vocabulary/ba_vocabulary.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:65de254f1bc43c0356834a3df469079bf03def47f1cebad5fc6f9eda7540b462
3
- size 6781677
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea55b63e567481e3fa0344c160f30692dc8f8c9cbb71d8ab93b1172cc63e3fb4
3
+ size 6780710
models/vocabulary/ba_vocabulary_metadata.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "language": "ba",
3
- "vocabulary_size": 391795,
4
  "variant": "full",
5
  "statistics": {
6
- "type_token_ratio": 0.041486328965339804,
7
  "coverage": {
8
- "top_100": 0.23295039278580296,
9
- "top_1000": 0.510335569094746,
10
- "top_5000": 0.6976257341559833,
11
- "top_10000": 0.7667897867532179
12
  },
13
- "hapax_count": 523451,
14
- "hapax_ratio": 0.5719238324996777,
15
- "total_documents": 63858
16
  }
17
  }
 
1
  {
2
  "language": "ba",
3
+ "vocabulary_size": 390661,
4
  "variant": "full",
5
  "statistics": {
6
+ "type_token_ratio": 0.041501301106049236,
7
  "coverage": {
8
+ "top_100": 0.23327066896778434,
9
+ "top_1000": 0.5108850357531433,
10
+ "top_5000": 0.6981018960947909,
11
+ "top_10000": 0.7672025183565773
12
  },
13
+ "hapax_count": 522357,
14
+ "hapax_ratio": 0.5721212506215649,
15
+ "total_documents": 63864
16
  }
17
  }
models/word_markov/ba_markov_ctx1_word.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:92fd33e481646024de79a8020c805420b7e263733616508172265ae2ae3ffd6d
3
- size 103200739
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a300e6e5b623405b8247e2b76e99381dfe507d0a11b1a63d3a9520509692076
3
+ size 102828495
models/word_markov/ba_markov_ctx1_word_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 1,
3
  "variant": "word",
4
  "language": "ba",
5
- "unique_contexts": 915102,
6
- "total_transitions": 21997530
7
  }
 
2
  "context_size": 1,
3
  "variant": "word",
4
  "language": "ba",
5
+ "unique_contexts": 912874,
6
+ "total_transitions": 21935880
7
  }
models/word_markov/ba_markov_ctx2_word.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6e3439ca9b5e1619b07b22f4f1b78774fe37bef5ff88ba91a9f9bb6d367c1dba
3
- size 335420766
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a584f9eecc95240d8362ca9fad4d1df0552829c6a5fbb565fe89f102f454bf9
3
+ size 334074318
models/word_markov/ba_markov_ctx2_word_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 2,
3
  "variant": "word",
4
  "language": "ba",
5
- "unique_contexts": 8225491,
6
- "total_transitions": 21933672
7
  }
 
2
  "context_size": 2,
3
  "variant": "word",
4
  "language": "ba",
5
+ "unique_contexts": 8193331,
6
+ "total_transitions": 21872016
7
  }