omarkamali commited on
Commit
90539ec
·
verified ·
1 Parent(s): b82db98

Upload all models and assets for be (latest)

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +1 -0
  2. README.md +218 -181
  3. models/embeddings/aligned/be_128d.bin +3 -0
  4. models/embeddings/aligned/be_128d.meta.json +1 -0
  5. models/embeddings/aligned/be_128d.projection.npy +3 -0
  6. models/embeddings/aligned/be_128d_metadata.json +8 -0
  7. models/embeddings/aligned/be_32d.bin +3 -0
  8. models/embeddings/aligned/be_32d.meta.json +1 -0
  9. models/embeddings/aligned/be_32d.projection.npy +3 -0
  10. models/embeddings/aligned/be_32d_metadata.json +8 -0
  11. models/embeddings/aligned/be_64d.bin +3 -0
  12. models/embeddings/aligned/be_64d.meta.json +1 -0
  13. models/embeddings/aligned/be_64d.projection.npy +3 -0
  14. models/embeddings/aligned/be_64d_metadata.json +8 -0
  15. models/embeddings/monolingual/be_128d.bin +2 -2
  16. models/embeddings/monolingual/be_128d_metadata.json +1 -1
  17. models/embeddings/monolingual/be_32d.bin +2 -2
  18. models/embeddings/monolingual/be_32d_metadata.json +1 -1
  19. models/embeddings/monolingual/be_64d.bin +2 -2
  20. models/embeddings/monolingual/be_64d_metadata.json +1 -1
  21. models/subword_markov/be_markov_ctx1_subword.parquet +2 -2
  22. models/subword_markov/be_markov_ctx1_subword_metadata.json +2 -2
  23. models/subword_markov/be_markov_ctx2_subword.parquet +2 -2
  24. models/subword_markov/be_markov_ctx2_subword_metadata.json +2 -2
  25. models/subword_markov/be_markov_ctx3_subword.parquet +2 -2
  26. models/subword_markov/be_markov_ctx3_subword_metadata.json +2 -2
  27. models/subword_markov/be_markov_ctx4_subword.parquet +2 -2
  28. models/subword_markov/be_markov_ctx4_subword_metadata.json +2 -2
  29. models/subword_ngram/be_2gram_subword.parquet +2 -2
  30. models/subword_ngram/be_2gram_subword_metadata.json +2 -2
  31. models/subword_ngram/be_3gram_subword.parquet +2 -2
  32. models/subword_ngram/be_3gram_subword_metadata.json +2 -2
  33. models/subword_ngram/be_4gram_subword.parquet +2 -2
  34. models/subword_ngram/be_4gram_subword_metadata.json +2 -2
  35. models/subword_ngram/be_5gram_subword.parquet +3 -0
  36. models/subword_ngram/be_5gram_subword_metadata.json +7 -0
  37. models/tokenizer/be_tokenizer_16k.model +2 -2
  38. models/tokenizer/be_tokenizer_16k.vocab +0 -0
  39. models/tokenizer/be_tokenizer_32k.model +2 -2
  40. models/tokenizer/be_tokenizer_32k.vocab +0 -0
  41. models/tokenizer/be_tokenizer_64k.model +2 -2
  42. models/tokenizer/be_tokenizer_64k.vocab +0 -0
  43. models/tokenizer/be_tokenizer_8k.model +2 -2
  44. models/tokenizer/be_tokenizer_8k.vocab +0 -0
  45. models/vocabulary/be_vocabulary.parquet +2 -2
  46. models/vocabulary/be_vocabulary_metadata.json +9 -9
  47. models/word_markov/be_markov_ctx1_word.parquet +2 -2
  48. models/word_markov/be_markov_ctx1_word_metadata.json +2 -2
  49. models/word_markov/be_markov_ctx2_word.parquet +2 -2
  50. models/word_markov/be_markov_ctx2_word_metadata.json +2 -2
.gitattributes CHANGED
@@ -39,3 +39,4 @@ visualizations/position_encoding_comparison.png filter=lfs diff=lfs merge=lfs -t
39
  visualizations/tsne_sentences.png filter=lfs diff=lfs merge=lfs -text
40
  visualizations/tsne_words.png filter=lfs diff=lfs merge=lfs -text
41
  visualizations/zipf_law.png filter=lfs diff=lfs merge=lfs -text
 
 
39
  visualizations/tsne_sentences.png filter=lfs diff=lfs merge=lfs -text
40
  visualizations/tsne_words.png filter=lfs diff=lfs merge=lfs -text
41
  visualizations/zipf_law.png filter=lfs diff=lfs merge=lfs -text
42
+ visualizations/embedding_tsne_multilingual.png filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  language: be
3
- language_name: BE
4
  language_family: slavic_east
5
  tags:
6
  - wikilangs
@@ -10,11 +10,21 @@ tags:
10
  - n-gram
11
  - markov
12
  - wikipedia
 
 
 
 
 
 
 
 
 
 
13
  - monolingual
14
  - family-slavic_east
15
  license: mit
16
  library_name: wikilangs
17
- pipeline_tag: feature-extraction
18
  datasets:
19
  - omarkamali/wikipedia-monthly
20
  dataset_info:
@@ -23,20 +33,20 @@ dataset_info:
23
  metrics:
24
  - name: best_compression_ratio
25
  type: compression
26
- value: 4.769
27
  - name: best_isotropy
28
  type: isotropy
29
- value: 0.6512
30
  - name: vocabulary_size
31
  type: vocab
32
  value: 0
33
- generated: 2026-01-03
34
  ---
35
 
36
- # BE - Wikilangs Models
37
  ## Comprehensive Research Report & Full Ablation Study
38
 
39
- This repository contains NLP models trained and evaluated by Wikilangs, specifically on **BE** Wikipedia data.
40
  We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and word embeddings.
41
 
42
  ## 📋 Repository Contents
@@ -60,7 +70,7 @@ We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and
60
  - [3. Markov Chain Evaluation](#3-markov-chain-evaluation)
61
  - [4. Vocabulary Analysis](#4-vocabulary-analysis)
62
  - [5. Word Embeddings Evaluation](#5-word-embeddings-evaluation)
63
- - [6. Morphological Analysis (Experimental)](#6-morphological-analysis)
64
  - [7. Summary & Recommendations](#7-summary--recommendations)
65
  - [Metrics Glossary](#appendix-metrics-glossary--interpretation-guide)
66
  - [Visualizations Index](#visualizations-index)
@@ -80,47 +90,47 @@ We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and
80
 
81
  | Vocab Size | Compression | Avg Token Len | UNK Rate | Total Tokens |
82
  |------------|-------------|---------------|----------|--------------|
83
- | **8k** | 3.593x | 3.60 | 0.0487% | 287,700 |
84
- | **16k** | 4.036x | 4.04 | 0.0547% | 256,163 |
85
- | **32k** | 4.451x | 4.46 | 0.0603% | 232,280 |
86
- | **64k** | 4.769x 🏆 | 4.77 | 0.0646% | 216,795 |
87
 
88
  ### Tokenization Examples
89
 
90
  Below are sample sentences tokenized with each vocabulary size:
91
 
92
- **Sample 1:** `Грынчэнкавэ () — вёска ў Ахтырскім раёне Сумскай вобласці Украіны. Уваходзіць у ...`
93
 
94
  | Vocab | Tokens | Count |
95
  |-------|--------|-------|
96
- | 8k | `▁гры н чэн ка вэ ▁() ▁— ▁вёска ▁ў ▁ах ... (+23 more)` | 33 |
97
- | 16k | `▁грын чэнка вэ ▁() ▁— ▁вёска ▁ў ▁ах ты рскім ... (+21 more)` | 31 |
98
- | 32k | `▁грын чэнка вэ ▁() ▁— ▁вёска ▁ў ▁ахты рскім ▁раёне ... (+19 more)` | 29 |
99
- | 64k | `▁грын чэнка вэ ▁() ▁— ▁вёска ▁ў ▁ахтырскім ▁раёне ▁сумскай ... (+17 more)` | 27 |
100
 
101
- **Sample 2:** `Лугавэ () — вёска ў Бродыўскім раёне Львоўскай вобласці Украіны. Крыніцы пункты ...`
102
 
103
  | Vocab | Tokens | Count |
104
  |-------|--------|-------|
105
- | 8k | `▁луга вэ ▁() ▁— ▁вёска ▁ў ▁б роды ўскім ▁раёне ... (+15 more)` | 25 |
106
- | 16k | `▁луга вэ ▁() ▁— ▁вёска ▁ў ▁б роды ўскім ▁раёне ... (+15 more)` | 25 |
107
- | 32k | `▁луга вэ ▁() ▁— ▁вёска ▁ў ▁броды ўскім ▁раёне ▁львоўскай ... (+13 more)` | 23 |
108
- | 64k | `▁луга вэ ▁() ▁— ▁вёска ▁ў ▁бродыўскім ▁раёне ▁львоўскай ▁вобласці ... (+11 more)` | 21 |
109
 
110
- **Sample 3:** `Косарэвэ () — вёска ў Млыніўскім раёне Ровенскай вобласці Украіны. Уваходзіць у ...`
111
 
112
  | Vocab | Tokens | Count |
113
  |-------|--------|-------|
114
- | 8k | `▁ко са рэ вэ ▁() ▁— ▁вёска ▁ў ▁млы ніў ... (+21 more)` | 31 |
115
- | 16k | `▁ко са рэ вэ ▁() ▁— ▁вёска ▁ў ▁млы ніўскім ... (+19 more)` | 29 |
116
- | 32k | `▁коса рэ вэ ▁() ▁— ▁вёска ▁ў ▁млы ніўскім ▁раёне ... (+17 more)` | 27 |
117
- | 64k | `▁коса рэ вэ ▁() ▁— ▁вёска ▁ў ▁млыніўскім ▁раёне ▁ровенскай ... (+15 more)` | 25 |
118
 
119
 
120
  ### Key Findings
121
 
122
- - **Best Compression:** 64k achieves 4.769x compression
123
- - **Lowest UNK Rate:** 8k with 0.0487% unknown tokens
124
  - **Trade-off:** Larger vocabularies improve compression but increase model size
125
  - **Recommendation:** 32k vocabulary provides optimal balance for production use
126
 
@@ -137,12 +147,14 @@ Below are sample sentences tokenized with each vocabulary size:
137
 
138
  | N-gram | Variant | Perplexity | Entropy | Unique N-grams | Top-100 Coverage | Top-1000 Coverage |
139
  |--------|---------|------------|---------|----------------|------------------|-------------------|
140
- | **2-gram** | Word | 114,899 | 16.81 | 1,095,876 | 11.4% | 25.2% |
141
- | **2-gram** | Subword | 453 🏆 | 8.82 | 15,607 | 55.9% | 96.8% |
142
- | **3-gram** | Word | 176,550 | 17.43 | 1,682,544 | 11.7% | 25.2% |
143
- | **3-gram** | Subword | 4,192 | 12.03 | 145,836 | 18.7% | 59.5% |
144
- | **4-gram** | Word | 286,677 | 18.13 | 2,809,290 | 9.5% | 25.0% |
145
- | **4-gram** | Subword | 25,337 | 14.63 | 930,596 | 8.0% | 29.4% |
 
 
146
 
147
  ### Top 5 N-grams by Size
148
 
@@ -151,19 +163,19 @@ Below are sample sentences tokenized with each vocabulary size:
151
  | Rank | N-gram | Count |
152
  |------|--------|-------|
153
  | 1 | `0 10` | 188,589 |
154
- | 2 | `10 0` | 184,433 |
155
- | 3 | `0 09` | 178,218 |
156
- | 4 | `09 0` | 172,686 |
157
- | 5 | `у годзе` | 140,117 |
158
 
159
  **3-grams (Word):**
160
 
161
  | Rank | N-gram | Count |
162
  |------|--------|-------|
163
- | 1 | `0 10 0` | 183,056 |
164
- | 2 | `0 09 0` | 171,686 |
165
- | 3 | `0 11 0` | 133,046 |
166
- | 4 | `0 08 0` | 125,664 |
167
  | 5 | `0 07 0` | 84,761 |
168
 
169
  **4-grams (Word):**
@@ -176,42 +188,62 @@ Below are sample sentences tokenized with each vocabulary size:
176
  | 4 | `47 0 10 0` | 26,709 |
177
  | 5 | `0 50 0 10` | 26,628 |
178
 
 
 
 
 
 
 
 
 
 
 
179
  **2-grams (Subword):**
180
 
181
  | Rank | N-gram | Count |
182
  |------|--------|-------|
183
- | 1 | `а _` | 7,375,676 |
184
- | 2 | `н а` | 5,829,339 |
185
- | 3 | `р а` | 5,735,773 |
186
- | 4 | `к а` | 4,959,811 |
187
- | 5 | `_ п` | 4,750,427 |
188
 
189
  **3-grams (Subword):**
190
 
191
  | Rank | N-gram | Count |
192
  |------|--------|-------|
193
- | 1 | `_ п а` | 2,102,007 |
194
- | 2 | `_ 0 ,` | 1,872,298 |
195
- | 3 | `_ н а` | 1,670,363 |
196
- | 4 | `н а _` | 1,424,587 |
197
- | 5 | `_ п р` | 1,341,590 |
198
 
199
  **4-grams (Subword):**
200
 
201
  | Rank | N-gram | Count |
202
  |------|--------|-------|
203
- | 1 | `а г а _` | 980,628 |
204
- | 2 | `_ п р а` | 746,402 |
205
- | 3 | `_ г о д` | 708,921 |
206
- | 4 | `_ н а _` | 692,237 |
207
- | 5 | `к а й _` | 545,902 |
 
 
 
 
 
 
 
 
 
 
208
 
209
 
210
  ### Key Findings
211
 
212
  - **Best Perplexity:** 2-gram (subword) with 453
213
  - **Entropy Trend:** Decreases with larger n-grams (more predictable)
214
- - **Coverage:** Top-1000 patterns cover ~29% of corpus
215
  - **Recommendation:** 4-gram or 5-gram for best predictive performance
216
 
217
  ---
@@ -227,14 +259,14 @@ Below are sample sentences tokenized with each vocabulary size:
227
 
228
  | Context | Variant | Avg Entropy | Perplexity | Branching Factor | Unique Contexts | Predictability |
229
  |---------|---------|-------------|------------|------------------|-----------------|----------------|
230
- | **1** | Word | 0.9806 | 1.973 | 10.65 | 1,594,726 | 1.9% |
231
- | **1** | Subword | 0.4731 | 1.388 | 3.96 | 16,459 | 52.7% |
232
- | **2** | Word | 0.3129 | 1.242 | 1.94 | 16,955,773 | 68.7% |
233
- | **2** | Subword | 0.6387 | 1.557 | 4.81 | 65,143 | 36.1% |
234
- | **3** | Word | 0.1126 | 1.081 | 1.23 | 32,878,014 | 88.7% |
235
- | **3** | Subword | 0.8192 | 1.764 | 4.91 | 313,186 | 18.1% |
236
- | **4** | Word | 0.0455 🏆 | 1.032 | 1.08 | 40,250,681 | 95.5% |
237
- | **4** | Subword | 0.7603 | 1.694 | 3.75 | 1,537,647 | 24.0% |
238
 
239
  ### Generated Text Samples (Word-based)
240
 
@@ -242,27 +274,27 @@ Below are text samples generated from each word-based Markov chain model:
242
 
243
  **Context Size 1:**
244
 
245
- 1. `0 57 0 09 0 67 0 07 0 58 км на 1 20 лютага жэнева`
246
- 2. `і стаўшы першым урадзе і гітарыст разам з поўдня сутыкненні прыпыніліся на кіргізскай сср 10 0`
247
- 3. `у годзе гэтыя эксперыменты па год 11 0 56 0 75 0 50 0 08 0`
248
 
249
  **Context Size 2:**
250
 
251
- 1. `0 10 0 50 0 10 0 39 0 11 0 36 0 12 0 54 0`
252
- 2. `10 0 68 0 25 0 6 1 52 1 25 джэсіка пегула эна сібахара 7 6`
253
- 3. `0 09 0 46 0 10 0 35 0 12 0 37 0 12 0 д2 прамень`
254
 
255
  **Context Size 3:**
256
 
257
- 1. `0 10 0 37 0 12 0 35 0 48 0 10 0 56 0 09 0 51`
258
- 2. `0 09 0 37 0 12 0 57 0 09 0 41 0 11 0 45 0 10`
259
- 3. `0 11 0 42 0 11 0 61 0 08 0 51 0 09 0 37 0 12`
260
 
261
  **Context Size 4:**
262
 
263
- 1. `0 44 0 10 0 52 0 09 0 43 0 11 0 76 0 07 0 37 0`
264
- 2. `44 0 10 0 51 0 09 0 51 0 09 0 42 0 11 0 60 0 08`
265
- 3. `0 47 0 10 0 54 0 09 0 65 0 08 0 38 0 11 0 46 0`
266
 
267
 
268
  ### Generated Text Samples (Subword-based)
@@ -271,34 +303,34 @@ Below are text samples generated from each subword-based Markov chain model:
271
 
272
  **Context Size 1:**
273
 
274
- 1. `_irone_саджырода`
275
- 2. `аса._бетвекаенсы`
276
- 3. `ных_г._тэні_09_��`
277
 
278
  **Context Size 2:**
279
 
280
- 1. `а_абто_чальны,_пр`
281
- 2. `наяны_нькімпіныма`
282
- 3. `раён_з_10),_якаге`
283
 
284
  **Context Size 3:**
285
 
286
- 1. `_паднакадэміі_пало`
287
- 2. `_0,40_0,56_0,50_0,`
288
- 3. `_на_паданні._перац`
289
 
290
  **Context Size 4:**
291
 
292
- 1. `ага_адсек_нацыя_4_т`
293
- 2. `_пра_ў_сваюць_62-я_`
294
- 3. `_годзе._жывяць_дызе`
295
 
296
 
297
  ### Key Findings
298
 
299
- - **Best Predictability:** Context-4 (word) with 95.5% predictability
300
  - **Branching Factor:** Decreases with context size (more deterministic)
301
- - **Memory Trade-off:** Larger contexts require more storage (1,537,647 contexts)
302
  - **Recommendation:** Context-3 or Context-4 for text generation
303
 
304
  ---
@@ -314,48 +346,48 @@ Below are text samples generated from each subword-based Markov chain model:
314
 
315
  | Metric | Value |
316
  |--------|-------|
317
- | Vocabulary Size | 739,605 |
318
- | Total Tokens | 54,963,738 |
319
- | Mean Frequency | 74.31 |
320
  | Median Frequency | 4 |
321
- | Frequency Std Dev | 3865.57 |
322
 
323
  ### Most Common Words
324
 
325
  | Rank | Word | Frequency |
326
  |------|------|-----------|
327
- | 1 | 0 | 1,944,698 |
328
- | 2 | і | 1,322,186 |
329
- | 3 | у | 1,231,156 |
330
- | 4 | ў | 1,155,870 |
331
- | 5 | з | 858,124 |
332
- | 6 | на | 705,989 |
333
- | 7 | года | 365,156 |
334
- | 8 | да | 288,350 |
335
- | 9 | годзе | 255,744 |
336
- | 10 | 10 | 239,762 |
337
 
338
  ### Least Common Words (from vocabulary)
339
 
340
  | Rank | Word | Frequency |
341
  |------|------|-----------|
342
- | 1 | іцуно | 2 |
343
- | 2 | міурай | 2 |
344
- | 3 | kodanshas | 2 |
345
- | 4 | llb | 2 |
346
- | 5 | давы́даўскае | 2 |
347
- | 6 | эльханон | 2 |
348
- | 7 | vilner | 2 |
349
- | 8 | emes | 2 |
350
- | 9 | folkstsaytung | 2 |
351
- | 10 | dertseyln | 2 |
352
 
353
  ### Zipf's Law Analysis
354
 
355
  | Metric | Value |
356
  |--------|-------|
357
  | Zipf Coefficient | 0.9714 |
358
- | R² (Goodness of Fit) | 0.997385 |
359
  | Adherence Quality | **excellent** |
360
 
361
  ### Coverage Analysis
@@ -371,7 +403,7 @@ Below are text samples generated from each subword-based Markov chain model:
371
 
372
  - **Zipf Compliance:** R²=0.9974 indicates excellent adherence to Zipf's law
373
  - **High Frequency Dominance:** Top 100 words cover 29.3% of corpus
374
- - **Long Tail:** 729,605 words needed for remaining 25.5% coverage
375
 
376
  ---
377
  ## 5. Word Embeddings Evaluation
@@ -387,37 +419,40 @@ Below are text samples generated from each subword-based Markov chain model:
387
 
388
  ### 5.1 Cross-Lingual Alignment
389
 
390
- > *Note: Multilingual alignment visualization not available for this language.*
 
 
391
 
392
 
393
  ### 5.2 Model Comparison
394
 
395
  | Model | Dimension | Isotropy | Semantic Density | Alignment R@1 | Alignment R@10 |
396
  |-------|-----------|----------|------------------|---------------|----------------|
397
- | **mono_32d** | 32 | 0.6148 | 0.3550 | N/A | N/A |
398
- | **mono_64d** | 64 | 0.6479 | 0.2915 | N/A | N/A |
399
- | **mono_128d** | 128 | 0.6512 🏆 | 0.2220 | N/A | N/A |
 
 
 
400
 
401
  ### Key Findings
402
 
403
- - **Best Isotropy:** mono_128d with 0.6512 (more uniform distribution)
404
- - **Semantic Density:** Average pairwise similarity of 0.2895. Lower values indicate better semantic separation.
405
- - **Alignment Quality:** No aligned models evaluated in this run.
406
  - **Recommendation:** 128d aligned for best cross-lingual performance
407
 
408
  ---
409
  ## 6. Morphological Analysis (Experimental)
410
 
411
- > ⚠️ **Warning:** This language shows low morphological productivity. The statistical signals used for this analysis may be noisy or less reliable than for morphologically rich languages.
412
-
413
  This section presents an automated morphological analysis derived from the statistical divergence between word-level and subword-level models. By analyzing where subword predictability spikes and where word-level coverage fails, we can infer linguistic structures without supervised data.
414
 
415
  ### 6.1 Productivity & Complexity
416
 
417
  | Metric | Value | Interpretation | Recommendation |
418
  |--------|-------|----------------|----------------|
419
- | Productivity Index | **0.000** | Low morphological productivity | ⚠️ Likely unreliable |
420
- | Idiomaticity Gap | **-1.000** | Low formulaic content | - |
421
 
422
  ### 6.2 Affix Inventory (Productive Units)
423
 
@@ -426,21 +461,21 @@ These are the most productive prefixes and suffixes identified by sampling the v
426
  #### Productive Prefixes
427
  | Prefix | Examples |
428
  |--------|----------|
429
- | `-ка` | каганаў, кайлі, карэлятыўных |
430
- | `-па` | пасуэлу, падую, паліцыянтаў |
431
- | `-пр` | протестантами, провозглашении, принципу |
432
 
433
  #### Productive Suffixes
434
  | Suffix | Examples |
435
  |--------|----------|
436
- | `-а` | кішскага, краснасельскага, апельсіна |
437
- | `-кі` | ліпнякі, чарашкі, вярцінскі |
438
- | `-га` | кішскага, краснасельскага, луэнга |
439
- | `-ай` | абнаўленчай, пустэльніцай, факталагічнай |
440
- | `-ага` | кішскага, краснасельскага, найбагацейшага |
441
- | `-мі` | неадмоўнымі, контурамі, абрамі |
442
- | `-ая` | наватухінская, загорская, чакаўская |
443
- | `-ыя` | шматбаковыя, перанятыя, узбагачаныя |
444
 
445
  ### 6.3 Bound Stems (Lexical Roots)
446
 
@@ -448,18 +483,18 @@ Bound stems are high-frequency subword units that are semantically cohesive but
448
 
449
  | Stem | Cohesion | Substitutability | Examples |
450
  |------|----------|------------------|----------|
451
- | `насц` | 1.82x | 190 contexts | насцю, насць, насці |
452
- | `елар` | 2.47x | 46 contexts | белар, гелар, келар |
453
- | `анск` | 1.35x | 1021 contexts | ганск, данск, канск |
454
- | `асел` | 2.07x | 87 contexts | расел, насел, асель |
455
- | `нскі` | 1.43x | 414 contexts | янскі, енскі, інскі |
456
- | `ання` | 1.67x | 173 contexts | рання, вання, ранняе |
457
- | `аецц` | 2.21x | 48 contexts | ваецца, каецца, лаецца |
458
- | `нска` | 1.35x | 500 contexts | унска, янска, минска |
459
- | `ўска` | 1.52x | 236 contexts | еўска, іўска, еўская |
460
- | `ленн` | 1.48x | 234 contexts | гленн, ленны, ленная |
461
- | `йска` | 1.59x | 149 contexts | йская, ейска, войска |
462
- | `уска` | 1.36x | 263 contexts | буска, гуска, ускат |
463
 
464
  ### 6.4 Affix Compatibility (Co-occurrence)
465
 
@@ -467,16 +502,16 @@ This table shows which prefixes and suffixes most frequently co-occur on the sam
467
 
468
  | Prefix | Suffix | Frequency | Examples |
469
  |--------|--------|-----------|----------|
470
- | `-ка` | `-а` | 66 words | каміна, камунізма |
471
- | `-па` | `-а` | 55 words | паступаленка, панінскага |
472
- | `-пр` | `-а` | 28 words | прыкладвацца, прынада |
473
- | `-па` | `-ай` | 21 words | паплаўковай, пастаяннай |
474
- | `-па` | `-мі` | 17 words | пасіўнымі, паказнікамі |
475
- | `-па` | `-кі` | 16 words | палінскі, падзьячаскі |
476
- | `-ка` | `-га` | 16 words | какамега, калобжагскага |
477
- | `-ка` | `-ага` | 15 words | калобжагскага, каламойскага |
478
- | `-ка` | `-кі` | 14 words | кадомскі, каўхаёкі |
479
- | `-ка` | `-аў` | 12 words | карыбаў, катэрынычаў |
480
 
481
  ### 6.5 Recursive Morpheme Segmentation
482
 
@@ -484,26 +519,28 @@ Using **Recursive Hierarchical Substitutability**, we decompose complex words in
484
 
485
  | Word | Suggested Split | Confidence | Stem |
486
  |------|-----------------|------------|------|
487
- | барыёнамі | **`барыё-на-мі`** | 6.0 | `барыё` |
488
- | курапаткіна | **`курапат-кі-на`** | 6.0 | `курапат` |
489
- | хакеістаў | **`хакеіст-аў`** | 4.5 | `хакеіст` |
490
- | навасібірская | **`навасібірск-ая`** | 4.5 | `навасібірск` |
491
- | пірамідаў | **`пірамід-аў`** | 4.5 | `пірамід` |
492
- | трансфарматараў | **`трансфарматар-аў`** | 4.5 | `трансфарматар` |
493
- | участковыя | **`участков-ыя`** | 4.5 | `участков` |
494
- | вузельчыкамі | **`вузельчыка-мі`** | 4.5 | `вузельчыка` |
495
- | мікрараёнаў | **`мікрараён-аў`** | 4.5 | `мікрараён` |
496
- | патраціць | **`па-траціць`** | 4.5 | `траціць` |
497
- | папоўніцца | **`па-поўніцца`** | 4.5 | `поўніцца` |
498
- | капашчэўскі | **`ка-па-шчэўс-кі`** | 4.5 | `шчэўс` |
499
- | накрыўкамі | **`накрыўка-мі`** | 4.5 | `накрыўка` |
500
- | наведвальніцкі | **`наведвальніц-кі`** | 4.5 | `наведвальніц` |
501
- | беспартыйнымі | **`беспартыйны-мі`** | 4.5 | `беспартыйны` |
502
 
503
  ### 6.6 Linguistic Interpretation
504
 
505
  > **Automated Insight:**
506
- The language BE appears to be more isolating or has a highly fixed vocabulary. Word-level models perform nearly as well as subword models, indicating fewer productive morphological processes.
 
 
507
 
508
  ---
509
  ## 7. Summary & Recommendations
@@ -516,7 +553,7 @@ The language BE appears to be more isolating or has a highly fixed vocabulary. W
516
  |-----------|-------------|-----------|
517
  | Tokenizer | **64k BPE** | Best compression (4.77x) |
518
  | N-gram | **2-gram** | Lowest perplexity (453) |
519
- | Markov | **Context-4** | Highest predictability (95.5%) |
520
  | Embeddings | **100d** | Balanced semantic capture and isotropy |
521
 
522
 
@@ -730,4 +767,4 @@ MIT License - Free for academic and commercial use.
730
  ---
731
  *Generated by Wikilangs Models Pipeline*
732
 
733
- *Report Date: 2026-01-03 11:32:17*
 
1
  ---
2
  language: be
3
+ language_name: Belarusian
4
  language_family: slavic_east
5
  tags:
6
  - wikilangs
 
10
  - n-gram
11
  - markov
12
  - wikipedia
13
+ - feature-extraction
14
+ - sentence-similarity
15
+ - tokenization
16
+ - n-grams
17
+ - markov-chain
18
+ - text-mining
19
+ - fasttext
20
+ - babelvec
21
+ - vocabulous
22
+ - vocabulary
23
  - monolingual
24
  - family-slavic_east
25
  license: mit
26
  library_name: wikilangs
27
+ pipeline_tag: text-generation
28
  datasets:
29
  - omarkamali/wikipedia-monthly
30
  dataset_info:
 
33
  metrics:
34
  - name: best_compression_ratio
35
  type: compression
36
+ value: 4.771
37
  - name: best_isotropy
38
  type: isotropy
39
+ value: 0.6444
40
  - name: vocabulary_size
41
  type: vocab
42
  value: 0
43
+ generated: 2026-01-06
44
  ---
45
 
46
+ # Belarusian - Wikilangs Models
47
  ## Comprehensive Research Report & Full Ablation Study
48
 
49
+ This repository contains NLP models trained and evaluated by Wikilangs, specifically on **Belarusian** Wikipedia data.
50
  We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and word embeddings.
51
 
52
  ## 📋 Repository Contents
 
70
  - [3. Markov Chain Evaluation](#3-markov-chain-evaluation)
71
  - [4. Vocabulary Analysis](#4-vocabulary-analysis)
72
  - [5. Word Embeddings Evaluation](#5-word-embeddings-evaluation)
73
+ - [6. Morphological Analysis (Experimental)](#6--morphological-analysis-experimental)
74
  - [7. Summary & Recommendations](#7-summary--recommendations)
75
  - [Metrics Glossary](#appendix-metrics-glossary--interpretation-guide)
76
  - [Visualizations Index](#visualizations-index)
 
90
 
91
  | Vocab Size | Compression | Avg Token Len | UNK Rate | Total Tokens |
92
  |------------|-------------|---------------|----------|--------------|
93
+ | **8k** | 3.599x | 3.60 | 0.0489% | 286,335 |
94
+ | **16k** | 4.042x | 4.05 | 0.0549% | 254,965 |
95
+ | **32k** | 4.455x | 4.46 | 0.0605% | 231,292 |
96
+ | **64k** | 4.771x 🏆 | 4.78 | 0.0648% | 215,975 |
97
 
98
  ### Tokenization Examples
99
 
100
  Below are sample sentences tokenized with each vocabulary size:
101
 
102
+ **Sample 1:** `Ланавычы () — вёска ў Самбірскім раёне Львоўскай вобласці Украіны. Крыніцы пункт...`
103
 
104
  | Vocab | Tokens | Count |
105
  |-------|--------|-------|
106
+ | 8k | `▁ла на вы чы ▁() ▁— ▁вёска ▁ў ▁сам бі ... (+12 more)` | 22 |
107
+ | 16k | `▁ла на вы чы ▁() ▁— ▁вёска ▁ў ▁сам бі ... (+12 more)` | 22 |
108
+ | 32k | `▁ла на вычы ▁() ▁— ▁вёска ▁ў ▁самбі рскім ▁раёне ... (+9 more)` | 19 |
109
+ | 64k | `▁лана вычы ▁() ▁— ▁вёска ▁ў ▁самбірскім ▁раёне ▁львоўскай ▁вобласці ... (+6 more)` | 16 |
110
 
111
+ **Sample 2:** `Марсо () — французскае прозвішча. Вядомыя носьбіты Марсель Марсо, французскі арт...`
112
 
113
  | Vocab | Tokens | Count |
114
  |-------|--------|-------|
115
+ | 8k | `▁мар со ▁() ���— ▁француз скае ▁прозвішча . ▁вядомыя ▁носьбіты ... (+17 more)` | 27 |
116
+ | 16k | `▁мар со ▁() ▁— ▁француз скае ▁прозвішча . ▁вядомыя ▁носьбіты ... (+16 more)` | 26 |
117
+ | 32k | `▁мар со ▁() ▁— ▁француз скае ▁прозвішча . ▁вядомыя ▁носьбіты ... (+15 more)` | 25 |
118
+ | 64k | `▁мар со ▁() ▁— ▁французскае ▁прозвішча . ▁вядомыя ▁носьбіты ▁марсель ... (+14 more)` | 24 |
119
 
120
+ **Sample 3:** `Вораніў () — вёска ў Гарадэнкіўскім раёне Івана-Франкоўскай вобласці Украіны. Кр...`
121
 
122
  | Vocab | Tokens | Count |
123
  |-------|--------|-------|
124
+ | 8k | `▁вора ніў ▁() ▁— ▁вёска ▁ў ▁гарад эн кі ўскім ... (+21 more)` | 31 |
125
+ | 16k | `▁вора ніў ▁() ▁— ▁вёска ▁ў ▁гарад эн кіўскім ▁раёне ... (+18 more)` | 28 |
126
+ | 32k | `▁вора ніў ▁() ▁— ▁вёска ▁ў ▁гарад эн кіўскім ▁раёне ... (+17 more)` | 27 |
127
+ | 64k | `▁вора ніў ▁() ▁— ▁вёска ▁ў ▁гарад эн кіўскім ▁раёне ... (+17 more)` | 27 |
128
 
129
 
130
  ### Key Findings
131
 
132
+ - **Best Compression:** 64k achieves 4.771x compression
133
+ - **Lowest UNK Rate:** 8k with 0.0489% unknown tokens
134
  - **Trade-off:** Larger vocabularies improve compression but increase model size
135
  - **Recommendation:** 32k vocabulary provides optimal balance for production use
136
 
 
147
 
148
  | N-gram | Variant | Perplexity | Entropy | Unique N-grams | Top-100 Coverage | Top-1000 Coverage |
149
  |--------|---------|------------|---------|----------------|------------------|-------------------|
150
+ | **2-gram** | Word | 115,602 | 16.82 | 1,101,685 | 11.4% | 25.2% |
151
+ | **2-gram** | Subword | 453 🏆 | 8.82 | 15,623 | 55.9% | 96.8% |
152
+ | **3-gram** | Word | 178,210 | 17.44 | 1,692,602 | 11.7% | 25.1% |
153
+ | **3-gram** | Subword | 4,191 | 12.03 | 146,010 | 18.7% | 59.5% |
154
+ | **4-gram** | Word | 289,150 | 18.14 | 2,823,610 | 9.4% | 24.9% |
155
+ | **4-gram** | Subword | 25,327 | 14.63 | 932,448 | 8.0% | 29.4% |
156
+ | **5-gram** | Word | 212,986 | 17.70 | 2,118,708 | 8.7% | 25.2% |
157
+ | **5-gram** | Subword | 104,621 | 16.67 | 3,234,164 | 4.5% | 17.2% |
158
 
159
  ### Top 5 N-grams by Size
160
 
 
163
  | Rank | N-gram | Count |
164
  |------|--------|-------|
165
  | 1 | `0 10` | 188,589 |
166
+ | 2 | `10 0` | 184,434 |
167
+ | 3 | `0 09` | 178,217 |
168
+ | 4 | `09 0` | 172,685 |
169
+ | 5 | `у годзе` | 141,829 |
170
 
171
  **3-grams (Word):**
172
 
173
  | Rank | N-gram | Count |
174
  |------|--------|-------|
175
+ | 1 | `0 10 0` | 183,055 |
176
+ | 2 | `0 09 0` | 171,685 |
177
+ | 3 | `0 11 0` | 133,047 |
178
+ | 4 | `0 08 0` | 125,665 |
179
  | 5 | `0 07 0` | 84,761 |
180
 
181
  **4-grams (Word):**
 
188
  | 4 | `47 0 10 0` | 26,709 |
189
  | 5 | `0 50 0 10` | 26,628 |
190
 
191
+ **5-grams (Word):**
192
+
193
+ | Rank | N-gram | Count |
194
+ |------|--------|-------|
195
+ | 1 | `0 44 0 10 0` | 27,892 |
196
+ | 2 | `0 47 0 10 0` | 26,707 |
197
+ | 3 | `0 50 0 10 0` | 26,249 |
198
+ | 4 | `0 45 0 10 0` | 25,524 |
199
+ | 5 | `0 49 0 10 0` | 24,716 |
200
+
201
  **2-grams (Subword):**
202
 
203
  | Rank | N-gram | Count |
204
  |------|--------|-------|
205
+ | 1 | `а _` | 7,411,164 |
206
+ | 2 | `н а` | 5,858,867 |
207
+ | 3 | `р а` | 5,764,007 |
208
+ | 4 | `к а` | 4,983,576 |
209
+ | 5 | `_ п` | 4,779,657 |
210
 
211
  **3-grams (Subword):**
212
 
213
  | Rank | N-gram | Count |
214
  |------|--------|-------|
215
+ | 1 | `_ п а` | 2,113,963 |
216
+ | 2 | `_ 0 ,` | 1,872,411 |
217
+ | 3 | `_ н а` | 1,678,358 |
218
+ | 4 | `н а _` | 1,430,853 |
219
+ | 5 | `_ п р` | 1,351,115 |
220
 
221
  **4-grams (Subword):**
222
 
223
  | Rank | N-gram | Count |
224
  |------|--------|-------|
225
+ | 1 | `а г а _` | 985,197 |
226
+ | 2 | `_ п р а` | 752,091 |
227
+ | 3 | `_ г о д` | 714,067 |
228
+ | 4 | `_ н а _` | 694,537 |
229
+ | 5 | `к а й _` | 548,513 |
230
+
231
+ **5-grams (Subword):**
232
+
233
+ | Rank | N-gram | Count |
234
+ |------|--------|-------|
235
+ | 1 | `к а г а _` | 467,479 |
236
+ | 2 | `с к а й _` | 409,977 |
237
+ | 3 | `с к а г а` | 393,058 |
238
+ | 4 | `б е л а р` | 392,561 |
239
+ | 5 | `е л а р у` | 392,043 |
240
 
241
 
242
  ### Key Findings
243
 
244
  - **Best Perplexity:** 2-gram (subword) with 453
245
  - **Entropy Trend:** Decreases with larger n-grams (more predictable)
246
+ - **Coverage:** Top-1000 patterns cover ~17% of corpus
247
  - **Recommendation:** 4-gram or 5-gram for best predictive performance
248
 
249
  ---
 
259
 
260
  | Context | Variant | Avg Entropy | Perplexity | Branching Factor | Unique Contexts | Predictability |
261
  |---------|---------|-------------|------------|------------------|-----------------|----------------|
262
+ | **1** | Word | 0.9802 | 1.973 | 10.66 | 1,600,794 | 2.0% |
263
+ | **1** | Subword | 0.4743 | 1.389 | 3.96 | 16,475 | 52.6% |
264
+ | **2** | Word | 0.3132 | 1.242 | 1.95 | 17,028,048 | 68.7% |
265
+ | **2** | Subword | 0.6391 | 1.557 | 4.81 | 65,298 | 36.1% |
266
+ | **3** | Word | 0.1128 | 1.081 | 1.23 | 33,045,925 | 88.7% |
267
+ | **3** | Subword | 0.8191 | 1.764 | 4.91 | 313,830 | 18.1% |
268
+ | **4** | Word | 0.0455 🏆 | 1.032 | 1.08 | 40,473,004 | 95.4% |
269
+ | **4** | Subword | 0.7606 | 1.694 | 3.75 | 1,541,159 | 23.9% |
270
 
271
  ### Generated Text Samples (Word-based)
272
 
 
274
 
275
  **Context Size 1:**
276
 
277
+ 1. `0 06 0 1 мінскай вобласці беларусі ў раёне віцебскай губерні земскага самакіравання якая выказалася ...`
278
+ 2. `і дзіцячы сад каралевы якія выменьвалі ў эджбастане бірмінгем сіці манчэстэр юнайтэд дзе адносна нев...`
279
+ 3. `у годзе стала ўскосным выглядзе шоу consecința istorică sibiu mitropolitul andrei yahorau alena маё ...`
280
 
281
  **Context Size 2:**
282
 
283
+ 1. `0 10 0 34 0 12 0 38 0 11 0 53 0 09 0 41 0`
284
+ 2. `10 0 55 0 09 0 46 0 10 0 63 0 08 0 75 0 07`
285
+ 3. `0 09 0 54 0 09 0 47 0 10 0 48 0 10 0 45 0`
286
 
287
  **Context Size 3:**
288
 
289
+ 1. `0 10 0 37 0 12 0 45 0 10 0 60 0 08 0 58 0 09`
290
+ 2. `0 09 0 54 0 09 0 50 0 09 so a 0 67 0 08 0 79`
291
+ 3. `0 11 0 47 0 10 0 54 0 09 0 48 0 10 0 43 0 11`
292
 
293
  **Context Size 4:**
294
 
295
+ 1. `0 44 0 10 0 40 0 11 0 54 0 32 0 45 0 32 0 56 0`
296
+ 2. `44 0 10 0 47 0 10 0 48 0 10 0 48 0 10 0 57 0 06`
297
+ 3. `0 47 0 10 0 54 0 09 0 87 0 06 sbbc 0 78 0 07 0 47`
298
 
299
 
300
  ### Generated Text Samples (Subword-based)
 
303
 
304
  **Context Size 1:**
305
 
306
+ 1. `_бек»_мано_szk._`
307
+ 2. `аёрларныкльбеніц`
308
+ 3. `нагркаў_вай_stol`
309
 
310
  **Context Size 2:**
311
 
312
+ 1. `а_вылкі_ў_парышша`
313
+ 2. `на_апілік_вы,_які`
314
+ 3. `раў_звагарскаў_вы`
315
 
316
  **Context Size 3:**
317
 
318
+ 1. `_памка:_ю._тайскаг`
319
+ 2. `_0,53_0,42_0,43_0,`
320
+ 3. `_насцю_і_тавіч_см.`
321
 
322
  **Context Size 4:**
323
 
324
+ 1. `ага_заняў_і_паведа,`
325
+ 2. `_прасійскаў_супольс`
326
+ 3. `_годзе_прыезда_філь`
327
 
328
 
329
  ### Key Findings
330
 
331
+ - **Best Predictability:** Context-4 (word) with 95.4% predictability
332
  - **Branching Factor:** Decreases with context size (more deterministic)
333
+ - **Memory Trade-off:** Larger contexts require more storage (1,541,159 contexts)
334
  - **Recommendation:** Context-3 or Context-4 for text generation
335
 
336
  ---
 
346
 
347
  | Metric | Value |
348
  |--------|-------|
349
+ | Vocabulary Size | 741,819 |
350
+ | Total Tokens | 55,243,342 |
351
+ | Mean Frequency | 74.47 |
352
  | Median Frequency | 4 |
353
+ | Frequency Std Dev | 3873.91 |
354
 
355
  ### Most Common Words
356
 
357
  | Rank | Word | Frequency |
358
  |------|------|-----------|
359
+ | 1 | 0 | 1,944,910 |
360
+ | 2 | і | 1,331,350 |
361
+ | 3 | у | 1,238,468 |
362
+ | 4 | ў | 1,161,043 |
363
+ | 5 | з | 862,221 |
364
+ | 6 | на | 708,262 |
365
+ | 7 | года | 367,568 |
366
+ | 8 | да | 290,434 |
367
+ | 9 | годзе | 258,378 |
368
+ | 10 | 10 | 239,964 |
369
 
370
  ### Least Common Words (from vocabulary)
371
 
372
  | Rank | Word | Frequency |
373
  |------|------|-----------|
374
+ | 1 | девятке | 2 |
375
+ | 2 | дэкунаў | 2 |
376
+ | 3 | iovine | 2 |
377
+ | 4 | іавін | 2 |
378
+ | 5 | аёвіну | 2 |
379
+ | 6 | джэніка | 2 |
380
+ | 7 | мэрылінам | 2 |
381
+ | 8 | сардэшная | 2 |
382
+ | 9 | івасю | 2 |
383
+ | 10 | стеценко | 2 |
384
 
385
  ### Zipf's Law Analysis
386
 
387
  | Metric | Value |
388
  |--------|-------|
389
  | Zipf Coefficient | 0.9714 |
390
+ | R² (Goodness of Fit) | 0.997383 |
391
  | Adherence Quality | **excellent** |
392
 
393
  ### Coverage Analysis
 
403
 
404
  - **Zipf Compliance:** R²=0.9974 indicates excellent adherence to Zipf's law
405
  - **High Frequency Dominance:** Top 100 words cover 29.3% of corpus
406
+ - **Long Tail:** 731,819 words needed for remaining 25.5% coverage
407
 
408
  ---
409
  ## 5. Word Embeddings Evaluation
 
419
 
420
  ### 5.1 Cross-Lingual Alignment
421
 
422
+ ![Alignment Quality](visualizations/embedding_alignment_quality.png)
423
+
424
+ ![Multilingual t-SNE](visualizations/embedding_tsne_multilingual.png)
425
 
426
 
427
  ### 5.2 Model Comparison
428
 
429
  | Model | Dimension | Isotropy | Semantic Density | Alignment R@1 | Alignment R@10 |
430
  |-------|-----------|----------|------------------|---------------|----------------|
431
+ | **mono_32d** | 32 | 0.6096 | 0.3533 | N/A | N/A |
432
+ | **mono_64d** | 64 | 0.6408 | 0.2859 | N/A | N/A |
433
+ | **mono_128d** | 128 | 0.6444 | 0.2271 | N/A | N/A |
434
+ | **aligned_32d** | 32 | 0.6096 | 0.3568 | 0.0440 | 0.3040 |
435
+ | **aligned_64d** | 64 | 0.6408 | 0.2908 | 0.1380 | 0.5080 |
436
+ | **aligned_128d** | 128 | 0.6444 🏆 | 0.2362 | 0.2300 | 0.6220 |
437
 
438
  ### Key Findings
439
 
440
+ - **Best Isotropy:** aligned_128d with 0.6444 (more uniform distribution)
441
+ - **Semantic Density:** Average pairwise similarity of 0.2917. Lower values indicate better semantic separation.
442
+ - **Alignment Quality:** Aligned models achieve up to 23.0% R@1 in cross-lingual retrieval.
443
  - **Recommendation:** 128d aligned for best cross-lingual performance
444
 
445
  ---
446
  ## 6. Morphological Analysis (Experimental)
447
 
 
 
448
  This section presents an automated morphological analysis derived from the statistical divergence between word-level and subword-level models. By analyzing where subword predictability spikes and where word-level coverage fails, we can infer linguistic structures without supervised data.
449
 
450
  ### 6.1 Productivity & Complexity
451
 
452
  | Metric | Value | Interpretation | Recommendation |
453
  |--------|-------|----------------|----------------|
454
+ | Productivity Index | **5.000** | High morphological productivity | Reliable analysis |
455
+ | Idiomaticity Gap | **0.467** | High formulaic/idiomatic content | - |
456
 
457
  ### 6.2 Affix Inventory (Productive Units)
458
 
 
461
  #### Productive Prefixes
462
  | Prefix | Examples |
463
  |--------|----------|
464
+ | `-па` | параллельной, падаплёка, падкіданні |
465
+ | `-ка` | канавалава, кафедрамі, калеснікава |
466
+ | `-пр` | прышчэпаўшчына, прыпяцкі, прапіткі |
467
 
468
  #### Productive Suffixes
469
  | Suffix | Examples |
470
  |--------|----------|
471
+ | `-а` | гароха, прышчэпаўшчына, падаплёка |
472
+ | `-га` | паўднёвага, іпацеўскага, міжазёрнага |
473
+ | `-кі` | леанінскі, прыпяцкі, прапіткі |
474
+ | `-ай` | кіянкай, ольстэрскай, найноўшай |
475
+ | `-ага` | паўднёвага, іпацеўскага, міжазёрнага |
476
+ | `-ая` | рудэральная, прымененая, свальбардская |
477
+ | `-аў` | шакіраваў, вігаў, шукальнікаў |
478
+ | `-на` | прышчэпаўшчына, непэсрэдна, скампанавана |
479
 
480
  ### 6.3 Bound Stems (Lexical Roots)
481
 
 
483
 
484
  | Stem | Cohesion | Substitutability | Examples |
485
  |------|----------|------------------|----------|
486
+ | `анск` | 1.51x | 1027 contexts | ганск, данск, канск |
487
+ | `нска` | 1.55x | 503 contexts | унска, янска, інская |
488
+ | `насц` | 1.79x | 190 contexts | насце, насця, насцю |
489
+ | `асел` | 2.08x | 87 contexts | асель, аселі, расел |
490
+ | `елар` | 2.39x | 47 contexts | белар, селар, гелар |
491
+ | `ўска` | 1.58x | 236 contexts | еўска, іўска, ёўскае |
492
+ | `аецц` | 2.20x | 48 contexts | маецца, каецца, лаецца |
493
+ | `тычн` | 1.49x | 233 contexts | этычны, стычня, этычна |
494
+ | `нскі` | 1.34x | 416 contexts | енскі, янс��і, інскі |
495
+ | `ельн` | 1.32x | 342 contexts | ельню, ельна, ельні |
496
+ | `ходз` | 1.47x | 182 contexts | ходзі, ходза, ходзь |
497
+ | `ання` | 1.47x | 174 contexts | рання, вання, арання |
498
 
499
  ### 6.4 Affix Compatibility (Co-occurrence)
500
 
 
502
 
503
  | Prefix | Suffix | Frequency | Examples |
504
  |--------|--------|-----------|----------|
505
+ | `-па` | `-а` | 57 words | падлічваюцца, павета |
506
+ | `-ка` | `-а` | 51 words | карахана, каралькова |
507
+ | `-пр` | `-а` | 33 words | прынцэса, працягваюцца |
508
+ | `-па` | `-ыя` | 14 words | падпружныя, пасярэбраныя |
509
+ | `-па` | `-ай` | 14 words | паўлавіцкай, пагібельнай |
510
+ | `-ка` | `-ая` | 14 words | карнуая, карэспандэнцкая |
511
+ | `-ка` | `-на` | 13 words | карахана, кадрына |
512
+ | `-ка` | `-га` | 13 words | калевальскага, каларадскага |
513
+ | `-па` | `-кі` | 13 words | пакупкі, палачанкі |
514
+ | `-па` | `-га` | 13 words | папаленага, палаткавага |
515
 
516
  ### 6.5 Recursive Morpheme Segmentation
517
 
 
519
 
520
  | Word | Suggested Split | Confidence | Stem |
521
  |------|-----------------|------------|------|
522
+ | галіцынаўка | **`галіцын-аў-ка`** | 6.0 | `галіцын` |
523
+ | перакладчыкаў | **`перакладчык-аў`** | 4.5 | `перакладчык` |
524
+ | зікуратаў | **`зікурат-аў`** | 4.5 | `зікурат` |
525
+ | астраблемай | **`астраблем-ай`** | 4.5 | `астраблем` |
526
+ | авіяатрадаў | **`авіяатрад-аў`** | 4.5 | `авіяатрад` |
527
+ | гукарадаў | **`гукарад-аў`** | 4.5 | `гукарад` |
528
+ | цырульнікаў | **`цырульнік-аў`** | 4.5 | `цырульнік` |
529
+ | адпраўшчыкаў | **`адпраўшчык-аў`** | 4.5 | `адпраўшчык` |
530
+ | рэдэмптарыстаў | **`рэдэмптарыст-аў`** | 4.5 | `рэдэмптарыст` |
531
+ | кулінараў | **`кулінар-аў`** | 4.5 | `кулінар` |
532
+ | іньігесаў | **`іньігес-аў`** | 4.5 | `іньігес` |
533
+ | гэлтахтаў | **`гэлтахт-аў`** | 4.5 | `гэлтахт` |
534
+ | рэгістрацыйна | **`рэгістрацый-на`** | 4.5 | `рэгістрацый` |
535
+ | чапаеўскага | **`чапаеўск-ага`** | 4.5 | `чапаеўск` |
536
+ | грунтоўка | **`грунтоў-ка`** | 4.5 | `грунтоў` |
537
 
538
  ### 6.6 Linguistic Interpretation
539
 
540
  > **Automated Insight:**
541
+ The language Belarusian shows high morphological productivity. The subword models are significantly more efficient than word models, suggesting a rich system of affixation or compounding.
542
+
543
+ > **Note on Idiomaticity:** The high Idiomaticity Gap suggests a large number of frequent multi-word expressions or formulaic sequences that are statistically distinct from their component parts.
544
 
545
  ---
546
  ## 7. Summary & Recommendations
 
553
  |-----------|-------------|-----------|
554
  | Tokenizer | **64k BPE** | Best compression (4.77x) |
555
  | N-gram | **2-gram** | Lowest perplexity (453) |
556
+ | Markov | **Context-4** | Highest predictability (95.4%) |
557
  | Embeddings | **100d** | Balanced semantic capture and isotropy |
558
 
559
 
 
767
  ---
768
  *Generated by Wikilangs Models Pipeline*
769
 
770
+ *Report Date: 2026-01-06 15:57:39*
models/embeddings/aligned/be_128d.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2398fdf9124672a6f7ed2d7a9dba453f1c6fc1da62df8fd1b78517f026d18c39
3
+ size 1569698641
models/embeddings/aligned/be_128d.meta.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lang": "be", "dim": 128, "max_seq_len": 512, "is_aligned": true}
models/embeddings/aligned/be_128d.projection.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96fcf5038c28eaead63c420c94055bca962719cdbb23ca63661cdc0becf7130d
3
+ size 65664
models/embeddings/aligned/be_128d_metadata.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "language": "be",
3
+ "dimension": 128,
4
+ "version": "aligned",
5
+ "hub_language": "en",
6
+ "seed_vocab_size": 42773,
7
+ "vocab_size": 519797
8
+ }
models/embeddings/aligned/be_32d.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4755e61bf97fcc4ab93a50344a113542111b8613ffa53c5507175f2b8dae6dd
3
+ size 402494545
models/embeddings/aligned/be_32d.meta.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lang": "be", "dim": 32, "max_seq_len": 512, "is_aligned": true}
models/embeddings/aligned/be_32d.projection.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e7612eb2b5ad863c05f6f99e62aa672d4fc162776296f2a9d41a0d77a46c676
3
+ size 4224
models/embeddings/aligned/be_32d_metadata.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "language": "be",
3
+ "dimension": 32,
4
+ "version": "aligned",
5
+ "hub_language": "en",
6
+ "seed_vocab_size": 42773,
7
+ "vocab_size": 519797
8
+ }
models/embeddings/aligned/be_64d.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ded6f659a5f25bd1e4c2cf7f0634e7fc41467c0b8fffc1cfabec6463e11ecc8d
3
+ size 791562577
models/embeddings/aligned/be_64d.meta.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lang": "be", "dim": 64, "max_seq_len": 512, "is_aligned": true}
models/embeddings/aligned/be_64d.projection.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3958212057a2f8723f332186599eb6d9450132cf5b6366dc313f0fd9c84760c2
3
+ size 16512
models/embeddings/aligned/be_64d_metadata.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "language": "be",
3
+ "dimension": 64,
4
+ "version": "aligned",
5
+ "hub_language": "en",
6
+ "seed_vocab_size": 42773,
7
+ "vocab_size": 519797
8
+ }
models/embeddings/monolingual/be_128d.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e80dd83fd9b000c473bacdfc520317bc08c8e6232f6acc8ddf47a4dc636212b7
3
- size 1567868138
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2398fdf9124672a6f7ed2d7a9dba453f1c6fc1da62df8fd1b78517f026d18c39
3
+ size 1569698641
models/embeddings/monolingual/be_128d_metadata.json CHANGED
@@ -11,5 +11,5 @@
11
  "encoding_method": "rope",
12
  "dim": 128
13
  },
14
- "vocab_size": 518052
15
  }
 
11
  "encoding_method": "rope",
12
  "dim": 128
13
  },
14
+ "vocab_size": 519797
15
  }
models/embeddings/monolingual/be_32d.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e15ec6617f84546d2951de84ffe80fbfa2280da80a7135e996e30747c163a575
3
- size 402004202
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4755e61bf97fcc4ab93a50344a113542111b8613ffa53c5507175f2b8dae6dd
3
+ size 402494545
models/embeddings/monolingual/be_32d_metadata.json CHANGED
@@ -11,5 +11,5 @@
11
  "encoding_method": "rope",
12
  "dim": 32
13
  },
14
- "vocab_size": 518052
15
  }
 
11
  "encoding_method": "rope",
12
  "dim": 32
13
  },
14
+ "vocab_size": 519797
15
  }
models/embeddings/monolingual/be_64d.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0dca4824861fd94e6b9de472d555ae08662bf04a8795cab1ac77097e32c191f3
3
- size 790625514
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ded6f659a5f25bd1e4c2cf7f0634e7fc41467c0b8fffc1cfabec6463e11ecc8d
3
+ size 791562577
models/embeddings/monolingual/be_64d_metadata.json CHANGED
@@ -11,5 +11,5 @@
11
  "encoding_method": "rope",
12
  "dim": 64
13
  },
14
- "vocab_size": 518052
15
  }
 
11
  "encoding_method": "rope",
12
  "dim": 64
13
  },
14
+ "vocab_size": 519797
15
  }
models/subword_markov/be_markov_ctx1_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:16ec89ccbf7b33b419dff7091cc3396dd6b9c2f2d9e7b4aaa101c1f6dc261e98
3
- size 528755
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3942a714b5a38d5d96ee158d63c1aa919d9ab86f0e43931415d4af822ef1069e
3
+ size 534939
models/subword_markov/be_markov_ctx1_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 1,
3
  "variant": "subword",
4
  "language": "be",
5
- "unique_contexts": 16459,
6
- "total_transitions": 384276543
7
  }
 
2
  "context_size": 1,
3
  "variant": "subword",
4
  "language": "be",
5
+ "unique_contexts": 16475,
6
+ "total_transitions": 386334702
7
  }
models/subword_markov/be_markov_ctx2_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:75a72bc43ff9fcb1e07421d9900ef838856a31dd2e997a85da1ec51c2da7313f
3
- size 2698586
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dac602100171ff68ef9af0bed9e4c36f887a6ef52a25684636d67ff1ca55a61e
3
+ size 2719683
models/subword_markov/be_markov_ctx2_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 2,
3
  "variant": "subword",
4
  "language": "be",
5
- "unique_contexts": 65143,
6
- "total_transitions": 384021043
7
  }
 
2
  "context_size": 2,
3
  "variant": "subword",
4
  "language": "be",
5
+ "unique_contexts": 65298,
6
+ "total_transitions": 386077966
7
  }
models/subword_markov/be_markov_ctx3_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3cc5bbbf80158973cace1739a5b2da93ae4aba1805dddfad45e13be87b4dd5b4
3
- size 12779069
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b861f2a419cdf3ef03894e8873567b72ea96075efe4764bb551f15037bff314
3
+ size 12801191
models/subword_markov/be_markov_ctx3_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 3,
3
  "variant": "subword",
4
  "language": "be",
5
- "unique_contexts": 313186,
6
- "total_transitions": 383765543
7
  }
 
2
  "context_size": 3,
3
  "variant": "subword",
4
  "language": "be",
5
+ "unique_contexts": 313830,
6
+ "total_transitions": 385821230
7
  }
models/subword_markov/be_markov_ctx4_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:78d8dd2621f613c8c1d06109067bcf9cbac4f41f2929f949639de78672590907
3
- size 48720729
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:53d211d50f2ccdea6ee36e7c38dca0f43324033cc38b6518b3a16d49c5e9c97a
3
+ size 48789277
models/subword_markov/be_markov_ctx4_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 4,
3
  "variant": "subword",
4
  "language": "be",
5
- "unique_contexts": 1537647,
6
- "total_transitions": 383510043
7
  }
 
2
  "context_size": 4,
3
  "variant": "subword",
4
  "language": "be",
5
+ "unique_contexts": 1541159,
6
+ "total_transitions": 385564494
7
  }
models/subword_ngram/be_2gram_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aafc4ee9f69f303f6f198618f5bee0cac66a99dacae147499dc0cae12854a772
3
- size 221209
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a20ed88f731298f393820cc8e6a49e7e9ad366b29d99b05efe77c0d29a8897b
3
+ size 221285
models/subword_ngram/be_2gram_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "n": 2,
3
  "variant": "subword",
4
  "language": "be",
5
- "unique_ngrams": 15607,
6
- "total_ngrams": 384276543
7
  }
 
2
  "n": 2,
3
  "variant": "subword",
4
  "language": "be",
5
+ "unique_ngrams": 15623,
6
+ "total_ngrams": 386334702
7
  }
models/subword_ngram/be_3gram_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:87fb73101845b2c8cdea801fcdcd4465df82baa9bba94ed1aefa8c506c088840
3
- size 1907996
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3f393ed68e52b9087a46bac8620c06fcc6c26109ce49b55725df279ef40727b
3
+ size 1898998
models/subword_ngram/be_3gram_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "n": 3,
3
  "variant": "subword",
4
  "language": "be",
5
- "unique_ngrams": 145836,
6
- "total_ngrams": 384021043
7
  }
 
2
  "n": 3,
3
  "variant": "subword",
4
  "language": "be",
5
+ "unique_ngrams": 146010,
6
+ "total_ngrams": 386077966
7
  }
models/subword_ngram/be_4gram_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0a1d30f2ce0acd57d7abf371e1916b606dd8f958d5fc479f3dc5736b5bb18b10
3
- size 12274905
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4830ea433f90560d2d7544d8e762d5cfc269a9c2d43405ae2bfc9b424f8022e0
3
+ size 12284822
models/subword_ngram/be_4gram_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "n": 4,
3
  "variant": "subword",
4
  "language": "be",
5
- "unique_ngrams": 930596,
6
- "total_ngrams": 383765543
7
  }
 
2
  "n": 4,
3
  "variant": "subword",
4
  "language": "be",
5
+ "unique_ngrams": 932448,
6
+ "total_ngrams": 385821230
7
  }
models/subword_ngram/be_5gram_subword.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62972cf284ad315d93677c8b7f92a406ee75b38d2a82ff40e134a09c75f18e04
3
+ size 45152293
models/subword_ngram/be_5gram_subword_metadata.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "n": 5,
3
+ "variant": "subword",
4
+ "language": "be",
5
+ "unique_ngrams": 3234164,
6
+ "total_ngrams": 385564494
7
+ }
models/tokenizer/be_tokenizer_16k.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:008fe4df9c07918b817613d49143c9d406e08cd7c95f2c94d7e35e4d7af0322f
3
- size 592885
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f3de32b0a7b3b3a3b69d24f6b697d9794ee59c08d0ffc70f7d561ebad1d439f
3
+ size 592882
models/tokenizer/be_tokenizer_16k.vocab CHANGED
The diff for this file is too large to render. See raw diff
 
models/tokenizer/be_tokenizer_32k.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f2db34459f167d40ce24759a3730279bf398faad2bcfe0de422d5a1ec7a70ffc
3
- size 969782
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ecb23f7b5f2f82a8fd7aea00b3054d7ecd036fe45503b17c3ad800cc12d9bb2
3
+ size 969548
models/tokenizer/be_tokenizer_32k.vocab CHANGED
The diff for this file is too large to render. See raw diff
 
models/tokenizer/be_tokenizer_64k.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:df2ee1b2850c4e4bd93d09aa2f1f4c06b4fd62dd623170b145a36f61154961b9
3
- size 1751650
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4dfed6e10895ec9f14ab639f9cbc12b5c079d85b4ab209949758712be587f3fe
3
+ size 1749733
models/tokenizer/be_tokenizer_64k.vocab CHANGED
The diff for this file is too large to render. See raw diff
 
models/tokenizer/be_tokenizer_8k.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e07b5ee32211d68f303eb0ca2473ef5a3e47cf3d435dbe20a3f50b5e40747119
3
- size 410385
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d29ff98b89bf26ec846ef03f8b5039985fda77c537f34a01eb209ee2abcb87d
3
+ size 410417
models/tokenizer/be_tokenizer_8k.vocab CHANGED
The diff for this file is too large to render. See raw diff
 
models/vocabulary/be_vocabulary.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eaef8e90391cc62be2430106a5b0b4c67cc2dfacdb35f517f62c46107295d042
3
- size 12490294
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:60f54584eb998e8e4cec858bc3bc846ac14f0cc3cb7cb2b11da43a339eb5bfda
3
+ size 12528391
models/vocabulary/be_vocabulary_metadata.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "language": "be",
3
- "vocabulary_size": 739605,
4
  "variant": "full",
5
  "statistics": {
6
- "type_token_ratio": 0.028584751562556937,
7
  "coverage": {
8
- "top_100": 0.28834518105660356,
9
- "top_1000": 0.49802155961854777,
10
- "top_5000": 0.6639092244917146,
11
- "top_10000": 0.7333083111156797
12
  },
13
- "hapax_count": 855988,
14
- "hapax_ratio": 0.5364701399417019,
15
- "total_documents": 255500
16
  }
17
  }
 
1
  {
2
  "language": "be",
3
+ "vocabulary_size": 741819,
4
  "variant": "full",
5
  "statistics": {
6
+ "type_token_ratio": 0.028548400082640594,
7
  "coverage": {
8
+ "top_100": 0.2882919700503959,
9
+ "top_1000": 0.4979523531099726,
10
+ "top_5000": 0.663883930712732,
11
+ "top_10000": 0.7333078754770741
12
  },
13
+ "hapax_count": 859837,
14
+ "hapax_ratio": 0.5368424930197245,
15
+ "total_documents": 256736
16
  }
17
  }
models/word_markov/be_markov_ctx1_word.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e8935add6e30b042b611c05c62b5e95de82abb4595dfd2e015226bf394cfb1f0
3
- size 207227789
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e032b1d50d56ef25717d00fa1d93c643346f946c30281462d04acb7143fee7bf
3
+ size 207708356
models/word_markov/be_markov_ctx1_word_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 1,
3
  "variant": "word",
4
  "language": "be",
5
- "unique_contexts": 1594726,
6
- "total_transitions": 55564226
7
  }
 
2
  "context_size": 1,
3
  "variant": "word",
4
  "language": "be",
5
+ "unique_contexts": 1600794,
6
+ "total_transitions": 55846443
7
  }
models/word_markov/be_markov_ctx2_word.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ecade528e4ed410d8e502020cb4476eb3034bb468a559acee35d2d25b0b413e1
3
- size 740234356
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1ae989a56b7f61469acbe2340dda964b8474df26de7089f3a1ff8a2ff624b38
3
+ size 743701483
models/word_markov/be_markov_ctx2_word_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 2,
3
  "variant": "word",
4
  "language": "be",
5
- "unique_contexts": 16955773,
6
- "total_transitions": 55308726
7
  }
 
2
  "context_size": 2,
3
  "variant": "word",
4
  "language": "be",
5
+ "unique_contexts": 17028048,
6
+ "total_transitions": 55589707
7
  }