omarkamali commited on
Commit
63c361e
·
verified ·
1 Parent(s): e423d75

Upload all models and assets for cv (latest)

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +1 -0
  2. README.md +207 -170
  3. models/embeddings/aligned/cv_128d.bin +3 -0
  4. models/embeddings/aligned/cv_128d.meta.json +1 -0
  5. models/embeddings/aligned/cv_128d.projection.npy +3 -0
  6. models/embeddings/aligned/cv_128d_metadata.json +8 -0
  7. models/embeddings/aligned/cv_32d.bin +3 -0
  8. models/embeddings/aligned/cv_32d.meta.json +1 -0
  9. models/embeddings/aligned/cv_32d.projection.npy +3 -0
  10. models/embeddings/aligned/cv_32d_metadata.json +8 -0
  11. models/embeddings/aligned/cv_64d.bin +3 -0
  12. models/embeddings/aligned/cv_64d.meta.json +1 -0
  13. models/embeddings/aligned/cv_64d.projection.npy +3 -0
  14. models/embeddings/aligned/cv_64d_metadata.json +8 -0
  15. models/embeddings/monolingual/cv_128d.bin +2 -2
  16. models/embeddings/monolingual/cv_128d_metadata.json +1 -1
  17. models/embeddings/monolingual/cv_32d.bin +2 -2
  18. models/embeddings/monolingual/cv_32d_metadata.json +1 -1
  19. models/embeddings/monolingual/cv_64d.bin +2 -2
  20. models/embeddings/monolingual/cv_64d_metadata.json +1 -1
  21. models/subword_markov/cv_markov_ctx1_subword.parquet +2 -2
  22. models/subword_markov/cv_markov_ctx1_subword_metadata.json +2 -2
  23. models/subword_markov/cv_markov_ctx2_subword.parquet +2 -2
  24. models/subword_markov/cv_markov_ctx2_subword_metadata.json +2 -2
  25. models/subword_markov/cv_markov_ctx3_subword.parquet +2 -2
  26. models/subword_markov/cv_markov_ctx3_subword_metadata.json +2 -2
  27. models/subword_markov/cv_markov_ctx4_subword.parquet +2 -2
  28. models/subword_markov/cv_markov_ctx4_subword_metadata.json +2 -2
  29. models/subword_ngram/cv_2gram_subword.parquet +2 -2
  30. models/subword_ngram/cv_2gram_subword_metadata.json +2 -2
  31. models/subword_ngram/cv_3gram_subword.parquet +2 -2
  32. models/subword_ngram/cv_3gram_subword_metadata.json +2 -2
  33. models/subword_ngram/cv_4gram_subword.parquet +2 -2
  34. models/subword_ngram/cv_4gram_subword_metadata.json +2 -2
  35. models/subword_ngram/cv_5gram_subword.parquet +3 -0
  36. models/subword_ngram/cv_5gram_subword_metadata.json +7 -0
  37. models/tokenizer/cv_tokenizer_16k.model +2 -2
  38. models/tokenizer/cv_tokenizer_16k.vocab +0 -0
  39. models/tokenizer/cv_tokenizer_32k.model +2 -2
  40. models/tokenizer/cv_tokenizer_32k.vocab +0 -0
  41. models/tokenizer/cv_tokenizer_64k.model +2 -2
  42. models/tokenizer/cv_tokenizer_64k.vocab +0 -0
  43. models/tokenizer/cv_tokenizer_8k.model +2 -2
  44. models/tokenizer/cv_tokenizer_8k.vocab +0 -0
  45. models/vocabulary/cv_vocabulary.parquet +2 -2
  46. models/vocabulary/cv_vocabulary_metadata.json +9 -9
  47. models/word_markov/cv_markov_ctx1_word.parquet +2 -2
  48. models/word_markov/cv_markov_ctx1_word_metadata.json +2 -2
  49. models/word_markov/cv_markov_ctx2_word.parquet +2 -2
  50. models/word_markov/cv_markov_ctx2_word_metadata.json +2 -2
.gitattributes CHANGED
@@ -39,3 +39,4 @@ visualizations/position_encoding_comparison.png filter=lfs diff=lfs merge=lfs -t
39
  visualizations/tsne_sentences.png filter=lfs diff=lfs merge=lfs -text
40
  visualizations/tsne_words.png filter=lfs diff=lfs merge=lfs -text
41
  visualizations/zipf_law.png filter=lfs diff=lfs merge=lfs -text
 
 
39
  visualizations/tsne_sentences.png filter=lfs diff=lfs merge=lfs -text
40
  visualizations/tsne_words.png filter=lfs diff=lfs merge=lfs -text
41
  visualizations/zipf_law.png filter=lfs diff=lfs merge=lfs -text
42
+ visualizations/embedding_tsne_multilingual.png filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  language: cv
3
- language_name: CV
4
  language_family: turkic_other
5
  tags:
6
  - wikilangs
@@ -10,11 +10,21 @@ tags:
10
  - n-gram
11
  - markov
12
  - wikipedia
 
 
 
 
 
 
 
 
 
 
13
  - monolingual
14
  - family-turkic_other
15
  license: mit
16
  library_name: wikilangs
17
- pipeline_tag: feature-extraction
18
  datasets:
19
  - omarkamali/wikipedia-monthly
20
  dataset_info:
@@ -23,20 +33,20 @@ dataset_info:
23
  metrics:
24
  - name: best_compression_ratio
25
  type: compression
26
- value: 3.792
27
  - name: best_isotropy
28
  type: isotropy
29
- value: 0.8332
30
  - name: vocabulary_size
31
  type: vocab
32
  value: 0
33
  generated: 2026-01-03
34
  ---
35
 
36
- # CV - Wikilangs Models
37
  ## Comprehensive Research Report & Full Ablation Study
38
 
39
- This repository contains NLP models trained and evaluated by Wikilangs, specifically on **CV** Wikipedia data.
40
  We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and word embeddings.
41
 
42
  ## 📋 Repository Contents
@@ -60,7 +70,7 @@ We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and
60
  - [3. Markov Chain Evaluation](#3-markov-chain-evaluation)
61
  - [4. Vocabulary Analysis](#4-vocabulary-analysis)
62
  - [5. Word Embeddings Evaluation](#5-word-embeddings-evaluation)
63
- - [6. Morphological Analysis (Experimental)](#6-morphological-analysis)
64
  - [7. Summary & Recommendations](#7-summary--recommendations)
65
  - [Metrics Glossary](#appendix-metrics-glossary--interpretation-guide)
66
  - [Visualizations Index](#visualizations-index)
@@ -80,47 +90,47 @@ We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and
80
 
81
  | Vocab Size | Compression | Avg Token Len | UNK Rate | Total Tokens |
82
  |------------|-------------|---------------|----------|--------------|
83
- | **8k** | 3.084x | 3.09 | 0.2455% | 244,836 |
84
- | **16k** | 3.356x | 3.36 | 0.2672% | 224,964 |
85
- | **32k** | 3.590x | 3.60 | 0.2857% | 210,324 |
86
- | **64k** | 3.792x 🏆 | 3.80 | 0.3018% | 199,115 |
87
 
88
  ### Tokenization Examples
89
 
90
  Below are sample sentences tokenized with each vocabulary size:
91
 
92
- **Sample 1:** `— Ăна хуплашкасене, тата пекех япаласене каясран сыхлама усă тата шалти Ĕç тата ...`
93
 
94
  | Vocab | Tokens | Count |
95
  |-------|--------|-------|
96
- | 8k | `▁— ▁ă на ▁хуп лаш ка сене , ▁тата ▁пекех ... (+24 more)` | 34 |
97
- | 16k | `▁— ▁ă на ▁хуп лаш ка сене , ▁тата ▁пекех ... (+24 more)` | 34 |
98
- | 32k | `▁— ▁ă на ▁хуп лашка сене , ▁тата ▁пекех ▁япаласене ... (+21 more)` | 31 |
99
- | 64k | `▁— ▁ă на ▁хуп лашка сене , ▁тата ▁пекех ▁япаласене ... (+20 more)` | 30 |
100
 
101
- **Sample 2:** `Шатдорф() коммуна, Швейцарири варринче Ури кантонне çын кодсайт хулисем ...`
102
 
103
  | Vocab | Tokens | Count |
104
  |-------|--------|-------|
105
- | 8k | `▁ш ат дорф ▁— ▁() ▁— ▁коммуна , ▁швейцарири ▁варринче ... (+9 more)` | 19 |
106
- | 16k | `▁шат дорф ▁— ▁() ▁— ▁коммуна , ▁швейцарири ▁варринче ▁ури ... (+8 more)` | 18 |
107
- | 32k | `▁шат дорф ▁— ▁() ▁— ▁коммуна , ▁швейцарири ▁варринче ▁ури ... (+8 more)` | 18 |
108
- | 64k | `▁шат дорф ▁— ▁() ▁— ▁коммуна , ▁швейцарири ▁варринче ▁ури ... (+8 more)` | 18 |
109
 
110
- **Sample 3:** `Çĕр йышшисем пуклак Малти урисем кайри урисем ытларах çĕр иртет. тата пурӑнать...`
111
 
112
  | Vocab | Tokens | Count |
113
  |-------|--------|-------|
114
- | 8k | `▁çĕ р ▁йышшисем ▁– ▁пу к лак ▁малти ▁у рисем ... (+30 more)` | 40 |
115
- | 16k | `▁çĕ р ▁йышшисем ▁– ▁пу к лак ▁малти ▁урисем ▁кайри ... (+27 more)` | 37 |
116
- | 32k | `▁çĕ р ▁йышшисем ▁– ▁пук лак ▁малти ▁урисем ▁кайри ▁урисем ... (+26 more)` | 36 |
117
- | 64k | `▁çĕ р ▁йышшисем ▁– ▁пуклак ▁малти ▁урисем ▁кайри ▁урисем ▁ытларах ... (+25 more)` | 35 |
118
 
119
 
120
  ### Key Findings
121
 
122
- - **Best Compression:** 64k achieves 3.792x compression
123
- - **Lowest UNK Rate:** 8k with 0.2455% unknown tokens
124
  - **Trade-off:** Larger vocabularies improve compression but increase model size
125
  - **Recommendation:** 32k vocabulary provides optimal balance for production use
126
 
@@ -137,12 +147,14 @@ Below are sample sentences tokenized with each vocabulary size:
137
 
138
  | N-gram | Variant | Perplexity | Entropy | Unique N-grams | Top-100 Coverage | Top-1000 Coverage |
139
  |--------|---------|------------|---------|----------------|------------------|-------------------|
140
- | **2-gram** | Word | 9,368 | 13.19 | 70,721 | 26.8% | 48.1% |
141
- | **2-gram** | Subword | 533 🏆 | 9.06 | 7,915 | 52.7% | 95.2% |
142
- | **3-gram** | Word | 8,221 | 13.01 | 88,916 | 30.4% | 52.4% |
143
- | **3-gram** | Subword | 4,932 | 12.27 | 69,443 | 17.2% | 56.3% |
144
- | **4-gram** | Word | 14,425 | 13.82 | 168,587 | 26.5% | 47.7% |
145
- | **4-gram** | Subword | 26,358 | 14.69 | 379,120 | 10.1% | 32.1% |
 
 
146
 
147
  ### Top 5 N-grams by Size
148
 
@@ -150,7 +162,7 @@ Below are sample sentences tokenized with each vocabulary size:
150
 
151
  | Rank | N-gram | Count |
152
  |------|--------|-------|
153
- | 1 | `шыв шыв` | 22,909 |
154
  | 2 | `территоринчи юханшыв` | 14,353 |
155
  | 3 | `территорипе юхать` | 13,579 |
156
  | 4 | `юхса юханшыв` | 13,517 |
@@ -162,56 +174,76 @@ Below are sample sentences tokenized with each vocabulary size:
162
  |------|--------|-------|
163
  | 1 | `рф экологи министерстви` | 11,700 |
164
  | 2 | `территорин шыв геоинформаци` | 11,389 |
165
- | 3 | `агентстви рф территорин` | 11,389 |
166
- | 4 | `рф территорин шыв` | 11,389 |
167
- | 5 | `федераци агентстви рф` | 11,389 |
168
 
169
  **4-grams (Word):**
170
 
171
  | Rank | N-gram | Count |
172
  |------|--------|-------|
173
- | 1 | `шыв геоинформаци системин шыв` | 11,389 |
174
- | 2 | `федераци агентстви рф территорин` | 11,389 |
175
  | 3 | `агентстви рф территорин шыв` | 11,389 |
176
- | 4 | `территорин шыв геоинформаци системин` | 11,389 |
177
- | 5 | `рф территорин шыв геоинформаци` | 11,389 |
 
 
 
 
 
 
 
 
 
 
178
 
179
  **2-grams (Subword):**
180
 
181
  | Rank | N-gram | Count |
182
  |------|--------|-------|
183
- | 1 | `. _` | 462,300 |
184
- | 2 | `а _` | 400,655 |
185
- | 3 | `и _` | 362,557 |
186
- | 4 | `— _` | 343,684 |
187
- | 5 | `_ —` | 341,328 |
188
 
189
  **3-grams (Subword):**
190
 
191
  | Rank | N-gram | Count |
192
  |------|--------|-------|
193
- | 1 | `_ — _` | 340,403 |
194
- | 2 | `ш ы в` | 149,607 |
195
- | 3 | `ы в _` | 121,960 |
196
- | 4 | `_ ю х` | 94,722 |
197
- | 5 | `т е р` | 86,265 |
198
 
199
  **4-grams (Subword):**
200
 
201
  | Rank | N-gram | Count |
202
  |------|--------|-------|
203
- | 1 | `ш ы в _` | 121,866 |
204
- | 2 | `_ ш ы в` | 85,504 |
205
- | 3 | `_ ю х а` | 76,923 |
206
- | 4 | `ю х а н` | 63,389 |
207
- | 5 | `х а н ш` | 63,293 |
 
 
 
 
 
 
 
 
 
 
208
 
209
 
210
  ### Key Findings
211
 
212
- - **Best Perplexity:** 2-gram (subword) with 533
213
  - **Entropy Trend:** Decreases with larger n-grams (more predictable)
214
- - **Coverage:** Top-1000 patterns cover ~32% of corpus
215
  - **Recommendation:** 4-gram or 5-gram for best predictive performance
216
 
217
  ---
@@ -227,14 +259,14 @@ Below are sample sentences tokenized with each vocabulary size:
227
 
228
  | Context | Variant | Avg Entropy | Perplexity | Branching Factor | Unique Contexts | Predictability |
229
  |---------|---------|-------------|------------|------------------|-----------------|----------------|
230
- | **1** | Word | 0.7797 | 1.717 | 5.33 | 352,008 | 22.0% |
231
- | **1** | Subword | 0.6180 | 1.535 | 6.05 | 3,627 | 38.2% |
232
- | **2** | Word | 0.1825 | 1.135 | 1.40 | 1,864,001 | 81.8% |
233
- | **2** | Subword | 0.9070 | 1.875 | 6.21 | 21,896 | 9.3% |
234
- | **3** | Word | 0.0523 | 1.037 | 1.09 | 2,580,848 | 94.8% |
235
- | **3** | Subword | 0.8725 | 1.831 | 4.70 | 135,774 | 12.7% |
236
- | **4** | Word | 0.0222 🏆 | 1.016 | 1.04 | 2,780,564 | 97.8% |
237
- | **4** | Subword | 0.7090 | 1.635 | 3.14 | 637,656 | 29.1% |
238
 
239
  ### Generated Text Samples (Word-based)
240
 
@@ -242,27 +274,27 @@ Below are text samples generated from each word-based Markov chain model:
242
 
243
  **Context Size 1:**
244
 
245
- 1. `шыв бассейн км2 экономика культура ял акимов анатолий тимофеевич разин остин пауэрс голдмембер найдж...`
246
- 2. `юханшыв хура kara online куликовский п в сталин рузвельт черчилль в в карьере сампраса на экраны`
247
- 3. `в о герард доу джонс в иван карабиц валерий валерьевич милици ричард iii вăл 15 pp`
248
 
249
  **Context Size 2:**
250
 
251
- 1. `шыв шыв иртыш юханшыв бассейн шыв кара таз чиккинчен енисей чиккичен бассейн çук юханшыв кара обь ху...`
252
- 2. `территоринчи юханшыв республики ростов ставрополь ен территорипе юхать юханшыва юхса юханшыв 33 км ю...`
253
- 3. `территорипе юхать лелен еган юханшыва юхса юханшыв 7 500 км ытла b 15 калса çу çур çу`
254
 
255
  **Context Size 3:**
256
 
257
- 1. `федераци агентстви рф территорин шыв геоинформаци системин шыв шыв гидрологи гт бассейн том гт 08 гт...`
258
- 2. `шыв шыв гидрологи гт бассейн том гт 11 гт 1 рф экологи министерстви республикин ао юпписем`
259
- 3. `шыв федераци агентстви рф территорин шыв геоинформаци системин шыв шыв гидрологи бассейн том гт 15 г...`
260
 
261
  **Context Size 4:**
262
 
263
- 1. `шыв геоинформаци системин шыв шыв гидрологи бассейн том гт 15 гт 3 рф экологи министерстви автономи ...`
264
- 2. `рф территорин шыв геоинформаци системин шыв шыв гидрологи бассейн том 15 3 рф экологи министерстви а...`
265
- 3. `федераци агентстви рф территорин шыв геоинформаци системин шыв шыв гидрологи гт бассейн том гт 03 гт...`
266
 
267
 
268
  ### Generated Text Samples (Subword-based)
@@ -271,34 +303,34 @@ Below are text samples generated from each subword-based Markov chain model:
271
 
272
  **Context Size 1:**
273
 
274
- 1. `_kruk=_neesmblot`
275
- 2. `адов_ин_пчнфи_ое`
276
- 3. `изра_,_ацин_—_je`
277
 
278
  **Context Size 2:**
279
 
280
- 1. `._кий,_фрах_(3_г.`
281
- 2. `а__перрипе_влеор`
282
- 3. `и_э.в._афизм_—_де`
283
 
284
  **Context Size 3:**
285
 
286
- 1. `_—_дев_анчах_тата_`
287
- 2. `шыва_сайтра_нингсе`
288
- 3. `ыв_5_-6_км._шыв_чу`
289
 
290
  **Context Size 4:**
291
 
292
- 1. `шыв_федераци_систер`
293
- 2. `_шыв_65_çын__--_гр`
294
- 3. `_юханшыв_шыв_-_ката`
295
 
296
 
297
  ### Key Findings
298
 
299
  - **Best Predictability:** Context-4 (word) with 97.8% predictability
300
  - **Branching Factor:** Decreases with context size (more deterministic)
301
- - **Memory Trade-off:** Larger contexts require more storage (637,656 contexts)
302
  - **Recommendation:** Context-3 or Context-4 for text generation
303
 
304
  ---
@@ -314,64 +346,64 @@ Below are text samples generated from each subword-based Markov chain model:
314
 
315
  | Metric | Value |
316
  |--------|-------|
317
- | Vocabulary Size | 148,629 |
318
- | Total Tokens | 3,880,417 |
319
- | Mean Frequency | 26.11 |
320
  | Median Frequency | 4 |
321
- | Frequency Std Dev | 438.75 |
322
 
323
  ### Most Common Words
324
 
325
  | Rank | Word | Frequency |
326
  |------|------|-----------|
327
- | 1 | шыв | 84,182 |
328
- | 2 | юханшыв | 53,747 |
329
- | 3 | в | 44,759 |
330
- | 4 | и | 40,900 |
331
- | 5 | с | 37,083 |
332
- | 6 | тата | 34,644 |
333
- | 7 | бассейн | 28,458 |
334
  | 8 | км | 25,026 |
335
- | 9 | м | 24,683 |
336
- | 10 | рф | 24,443 |
337
 
338
  ### Least Common Words (from vocabulary)
339
 
340
  | Rank | Word | Frequency |
341
  |------|------|-----------|
342
- | 1 | тӑтӑм | 2 |
343
- | 2 | мероприятире | 2 |
344
- | 3 | амфистрофпа | 2 |
345
- | 4 | ыйтрӗҫ | 2 |
346
- | 5 | поэмӑна | 2 |
347
- | 6 | хуламӑрта | 2 |
348
- | 7 | маршаленко | 2 |
349
- | 8 | шахмаметьев | 2 |
350
- | 9 | топилкин | 2 |
351
- | 10 | lupulella | 2 |
352
 
353
  ### Zipf's Law Analysis
354
 
355
  | Metric | Value |
356
  |--------|-------|
357
- | Zipf Coefficient | 1.0392 |
358
- | R² (Goodness of Fit) | 0.997768 |
359
  | Adherence Quality | **excellent** |
360
 
361
  ### Coverage Analysis
362
 
363
  | Top N Words | Coverage |
364
  |-------------|----------|
365
- | Top 100 | 30.1% |
366
  | Top 1,000 | 56.1% |
367
  | Top 5,000 | 72.5% |
368
- | Top 10,000 | 79.1% |
369
 
370
  ### Key Findings
371
 
372
- - **Zipf Compliance:** R²=0.9978 indicates excellent adherence to Zipf's law
373
- - **High Frequency Dominance:** Top 100 words cover 30.1% of corpus
374
- - **Long Tail:** 138,629 words needed for remaining 20.9% coverage
375
 
376
  ---
377
  ## 5. Word Embeddings Evaluation
@@ -387,37 +419,40 @@ Below are text samples generated from each subword-based Markov chain model:
387
 
388
  ### 5.1 Cross-Lingual Alignment
389
 
390
- > *Note: Multilingual alignment visualization not available for this language.*
 
 
391
 
392
 
393
  ### 5.2 Model Comparison
394
 
395
  | Model | Dimension | Isotropy | Semantic Density | Alignment R@1 | Alignment R@10 |
396
  |-------|-----------|----------|------------------|---------------|----------------|
397
- | **mono_32d** | 32 | 0.8332 🏆 | 0.3551 | N/A | N/A |
398
- | **mono_64d** | 64 | 0.8307 | 0.2707 | N/A | N/A |
399
- | **mono_128d** | 128 | 0.8029 | 0.2175 | N/A | N/A |
 
 
 
400
 
401
  ### Key Findings
402
 
403
- - **Best Isotropy:** mono_32d with 0.8332 (more uniform distribution)
404
- - **Semantic Density:** Average pairwise similarity of 0.2811. Lower values indicate better semantic separation.
405
- - **Alignment Quality:** No aligned models evaluated in this run.
406
  - **Recommendation:** 128d aligned for best cross-lingual performance
407
 
408
  ---
409
  ## 6. Morphological Analysis (Experimental)
410
 
411
- > ⚠️ **Warning:** This language shows low morphological productivity. The statistical signals used for this analysis may be noisy or less reliable than for morphologically rich languages.
412
-
413
  This section presents an automated morphological analysis derived from the statistical divergence between word-level and subword-level models. By analyzing where subword predictability spikes and where word-level coverage fails, we can infer linguistic structures without supervised data.
414
 
415
  ### 6.1 Productivity & Complexity
416
 
417
  | Metric | Value | Interpretation | Recommendation |
418
  |--------|-------|----------------|----------------|
419
- | Productivity Index | **0.000** | Low morphological productivity | ⚠️ Likely unreliable |
420
- | Idiomaticity Gap | **-1.000** | Low formulaic content | - |
421
 
422
  ### 6.2 Affix Inventory (Productive Units)
423
 
@@ -430,12 +465,12 @@ These are the most productive prefixes and suffixes identified by sampling the v
430
  #### Productive Suffixes
431
  | Suffix | Examples |
432
  |--------|----------|
433
- | `-а` | ыйтса, родства, мещерякова |
434
- | `-ен` | килнисен, центрӗнчен, теппермен |
435
- | `-ов` | камышлов, ярдов, горбов |
436
- | `-не` | мартинсоне, кёльне, характеристикине |
437
- | `-ем` | автомагистральсем, сооруженисем, алкалоидсем |
438
- | `-сем` | автомагистральсем, сооруженисем, алкалоидсем |
439
 
440
  ### 6.3 Bound Stems (Lexical Roots)
441
 
@@ -443,18 +478,18 @@ Bound stems are high-frequency subword units that are semantically cohesive but
443
 
444
  | Stem | Cohesion | Substitutability | Examples |
445
  |------|----------|------------------|----------|
446
- | `олог` | 1.78x | 172 contexts | зоолог, эколог, геолог |
447
- | `сейн` | 2.68x | 24 contexts | хасейн, хусейн, басейн |
448
- | `рито` | 2.40x | 26 contexts | крито, ритон, барито |
449
- | `огра` | 1.72x | 94 contexts | богра, ноград, ограды |
450
- | `ссей` | 2.68x | 17 contexts | эссей, ессей, рассей |
451
- | `ншыв` | 2.63x | 17 contexts | юханшыв, юшаншыв, юханшыве |
452
- | `ерри` | 2.35x | 22 contexts | черри, дерри, шерри |
453
- | `исте` | 1.74x | 58 contexts | листе, истеми, листер |
454
- | `орин` | 1.61x | 74 contexts | шорин, горин, борин |
455
- | `аншы` | 2.63x | 13 contexts | юханшыв, юшаншыв, юханшыве |
456
- | `блик` | 2.15x | 17 contexts | облик, облике, облика |
457
- | `нист` | 1.75x | 30 contexts | финист, горнист, хронист |
458
 
459
  ### 6.4 Affix Compatibility (Co-occurrence)
460
 
@@ -469,26 +504,28 @@ Using **Recursive Hierarchical Substitutability**, we decompose complex words in
469
 
470
  | Word | Suggested Split | Confidence | Stem |
471
  |------|-----------------|------------|------|
472
- | арӑслансем | **`арӑслан-сем`** | 4.5 | `арӑслан` |
473
- | минералов | **`минерал-ов`** | 4.5 | `минерал` |
474
- | эскимосов | **`эскимос-ов`** | 4.5 | `эскимос` |
475
- | динамикине | **`динамики-не`** | 4.5 | `динамики` |
476
- | учрежденийӗсем | **`учрежденийӗ-сем`** | 4.5 | `учрежденийӗ` |
477
- | председательне | **`председатель-не`** | 4.5 | `председатель` |
478
- | флоренцине | **`флоренци-не`** | 4.5 | `флоренци` |
479
- | вестготсем | **`вестгот-сем`** | 4.5 | `вестгот` |
480
- | королевине | **`королеви-не`** | 4.5 | `королеви` |
481
- | паракансем | **`паракан-сем`** | 4.5 | `паракан` |
482
- | приоритетне | **`приоритет-не`** | 4.5 | `приоритет` |
483
- | юханшывен | **`юханшыв-ен`** | 4.5 | `юханшыв` |
484
- | йӑмпӑлчӑксем | **`йӑмпӑлчӑк-сем`** | 4.5 | `йӑмпӑлчӑк` |
485
- | пайланусем | **`пайлану-сем`** | 4.5 | `пайлану` |
486
- | асапланнине | **`асапланни-не`** | 4.5 | `асапланни` |
487
 
488
  ### 6.6 Linguistic Interpretation
489
 
490
  > **Automated Insight:**
491
- The language CV appears to be more isolating or has a highly fixed vocabulary. Word-level models perform nearly as well as subword models, indicating fewer productive morphological processes.
 
 
492
 
493
  ---
494
  ## 7. Summary & Recommendations
@@ -499,8 +536,8 @@ The language CV appears to be more isolating or has a highly fixed vocabulary. W
499
 
500
  | Component | Recommended | Rationale |
501
  |-----------|-------------|-----------|
502
- | Tokenizer | **64k BPE** | Best compression (3.79x) |
503
- | N-gram | **2-gram** | Lowest perplexity (533) |
504
  | Markov | **Context-4** | Highest predictability (97.8%) |
505
  | Embeddings | **100d** | Balanced semantic capture and isotropy |
506
 
@@ -715,4 +752,4 @@ MIT License - Free for academic and commercial use.
715
  ---
716
  *Generated by Wikilangs Models Pipeline*
717
 
718
- *Report Date: 2026-01-03 11:00:29*
 
1
  ---
2
  language: cv
3
+ language_name: Chuvash
4
  language_family: turkic_other
5
  tags:
6
  - wikilangs
 
10
  - n-gram
11
  - markov
12
  - wikipedia
13
+ - feature-extraction
14
+ - sentence-similarity
15
+ - tokenization
16
+ - n-grams
17
+ - markov-chain
18
+ - text-mining
19
+ - fasttext
20
+ - babelvec
21
+ - vocabulous
22
+ - vocabulary
23
  - monolingual
24
  - family-turkic_other
25
  license: mit
26
  library_name: wikilangs
27
+ pipeline_tag: text-generation
28
  datasets:
29
  - omarkamali/wikipedia-monthly
30
  dataset_info:
 
33
  metrics:
34
  - name: best_compression_ratio
35
  type: compression
36
+ value: 3.778
37
  - name: best_isotropy
38
  type: isotropy
39
+ value: 0.8326
40
  - name: vocabulary_size
41
  type: vocab
42
  value: 0
43
  generated: 2026-01-03
44
  ---
45
 
46
+ # Chuvash - Wikilangs Models
47
  ## Comprehensive Research Report & Full Ablation Study
48
 
49
+ This repository contains NLP models trained and evaluated by Wikilangs, specifically on **Chuvash** Wikipedia data.
50
  We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and word embeddings.
51
 
52
  ## 📋 Repository Contents
 
70
  - [3. Markov Chain Evaluation](#3-markov-chain-evaluation)
71
  - [4. Vocabulary Analysis](#4-vocabulary-analysis)
72
  - [5. Word Embeddings Evaluation](#5-word-embeddings-evaluation)
73
+ - [6. Morphological Analysis (Experimental)](#6--morphological-analysis-experimental)
74
  - [7. Summary & Recommendations](#7-summary--recommendations)
75
  - [Metrics Glossary](#appendix-metrics-glossary--interpretation-guide)
76
  - [Visualizations Index](#visualizations-index)
 
90
 
91
  | Vocab Size | Compression | Avg Token Len | UNK Rate | Total Tokens |
92
  |------------|-------------|---------------|----------|--------------|
93
+ | **8k** | 3.075x | 3.08 | 0.2413% | 246,622 |
94
+ | **16k** | 3.345x | 3.35 | 0.2625% | 226,699 |
95
+ | **32k** | 3.576x | 3.58 | 0.2806% | 212,069 |
96
+ | **64k** | 3.778x 🏆 | 3.78 | 0.2964% | 200,734 |
97
 
98
  ### Tokenization Examples
99
 
100
  Below are sample sentences tokenized with each vocabulary size:
101
 
102
+ **Sample 1:** `Вики: Вики Wiki Wiki WIKI (FM) Wiki wiki dollar Wiki Wiki Shuttle WikiWikiWeb Ви...`
103
 
104
  | Vocab | Tokens | Count |
105
  |-------|--------|-------|
106
+ | 8k | `▁вики : ▁вики ▁wik i ▁wik i ▁wik i ▁( ... (+41 more)` | 51 |
107
+ | 16k | `▁вики : ▁вики ▁wiki ▁wiki ▁wiki ▁( f m ) ... (+28 more)` | 38 |
108
+ | 32k | `▁вики : ▁вики ▁wiki ▁wiki ▁wiki ▁( fm ) ▁wiki ... (+25 more)` | 35 |
109
+ | 64k | `▁вики : ▁вики ▁wiki ▁wiki ▁wiki ▁( fm ) ▁wiki ... (+23 more)` | 33 |
110
 
111
+ **Sample 2:** `Хро́мпикят е мар ят. Хромпик калий Топоним Хромпикçул Первоуральск (стан...`
112
 
113
  | Vocab | Tokens | Count |
114
  |-------|--------|-------|
115
+ | 8k | `▁х ро ́м п ик ▁— ▁ят ▁е ▁мар ▁ят ... (+51 more)` | 61 |
116
+ | 16k | `▁х ро ́м п ик ▁— ▁ят ▁е ▁мар ▁ят ... (+43 more)` | 53 |
117
+ | 32k | `▁х ро ́м пик ▁— ▁ят ▁е ▁мар ▁ят . ... (+36 more)` | 46 |
118
+ | 64k | `▁х ро ́м пик ▁— ▁ят ▁е ▁мар ▁ят . ... (+32 more)` | 42 |
119
 
120
+ **Sample 3:** `Мушар Республикин Куславкка ял. ял Коричев АССР Халах Вуламалли алфавитпа`
121
 
122
  | Vocab | Tokens | Count |
123
  |-------|--------|-------|
124
+ | 8k | `▁му шар ▁— ▁республикин ▁куславкка ▁ял . ▁ял ▁кори чев ... (+4 more)` | 14 |
125
+ | 16k | `▁му шар ▁— ▁республикин ▁куславкка ▁ял . ▁ял ▁кори чев ... (+4 more)` | 14 |
126
+ | 32k | `▁му шар ▁— ▁республикин ▁куславкка ▁ял . ▁ял ▁коричев ▁асср ... (+3 more)` | 13 |
127
+ | 64k | `▁му шар ▁— ▁республикин ▁куславкка ▁ял . ▁ял ▁коричев ▁асср ... (+3 more)` | 13 |
128
 
129
 
130
  ### Key Findings
131
 
132
+ - **Best Compression:** 64k achieves 3.778x compression
133
+ - **Lowest UNK Rate:** 8k with 0.2413% unknown tokens
134
  - **Trade-off:** Larger vocabularies improve compression but increase model size
135
  - **Recommendation:** 32k vocabulary provides optimal balance for production use
136
 
 
147
 
148
  | N-gram | Variant | Perplexity | Entropy | Unique N-grams | Top-100 Coverage | Top-1000 Coverage |
149
  |--------|---------|------------|---------|----------------|------------------|-------------------|
150
+ | **2-gram** | Word | 9,473 | 13.21 | 71,211 | 26.6% | 47.9% |
151
+ | **2-gram** | Subword | 532 🏆 | 9.06 | 7,908 | 52.7% | 95.2% |
152
+ | **3-gram** | Word | 8,325 | 13.02 | 89,585 | 30.3% | 52.2% |
153
+ | **3-gram** | Subword | 4,929 | 12.27 | 69,351 | 17.2% | 56.3% |
154
+ | **4-gram** | Word | 14,593 | 13.83 | 169,630 | 26.4% | 47.5% |
155
+ | **4-gram** | Subword | 26,364 | 14.69 | 378,926 | 10.1% | 32.1% |
156
+ | **5-gram** | Word | 12,306 | 13.59 | 144,170 | 27.1% | 49.1% |
157
+ | **5-gram** | Subword | 81,182 | 16.31 | 1,007,721 | 7.9% | 24.5% |
158
 
159
  ### Top 5 N-grams by Size
160
 
 
162
 
163
  | Rank | N-gram | Count |
164
  |------|--------|-------|
165
+ | 1 | `шыв шыв` | 22,911 |
166
  | 2 | `территоринчи юханшыв` | 14,353 |
167
  | 3 | `территорипе юхать` | 13,579 |
168
  | 4 | `юхса юханшыв` | 13,517 |
 
174
  |------|--------|-------|
175
  | 1 | `рф экологи министерстви` | 11,700 |
176
  | 2 | `территорин шыв геоинформаци` | 11,389 |
177
+ | 3 | `геоинформаци системин шыв` | 11,389 |
178
+ | 4 | `федераци агентстви рф` | 11,389 |
179
+ | 5 | `шыв федераци агентстви` | 11,389 |
180
 
181
  **4-grams (Word):**
182
 
183
  | Rank | N-gram | Count |
184
  |------|--------|-------|
185
+ | 1 | `геоинформаци системин шыв шыв` | 11,389 |
186
+ | 2 | `рф территорин шыв геоинформаци` | 11,389 |
187
  | 3 | `агентстви рф территорин шыв` | 11,389 |
188
+ | 4 | `федераци агентстви рф территорин` | 11,389 |
189
+ | 5 | `территорин шыв геоинформаци сис��емин` | 11,389 |
190
+
191
+ **5-grams (Word):**
192
+
193
+ | Rank | N-gram | Count |
194
+ |------|--------|-------|
195
+ | 1 | `агентстви рф территорин шыв геоинформаци` | 11,389 |
196
+ | 2 | `федераци агентстви рф территорин шыв` | 11,389 |
197
+ | 3 | `шыв геоинформаци системин шыв шыв` | 11,389 |
198
+ | 4 | `территорин шыв геоинформаци системин шыв` | 11,389 |
199
+ | 5 | `шыв федераци агентстви рф территорин` | 11,389 |
200
 
201
  **2-grams (Subword):**
202
 
203
  | Rank | N-gram | Count |
204
  |------|--------|-------|
205
+ | 1 | `. _` | 465,426 |
206
+ | 2 | `а _` | 402,164 |
207
+ | 3 | `и _` | 363,006 |
208
+ | 4 | `— _` | 346,175 |
209
+ | 5 | `_ —` | 343,660 |
210
 
211
  **3-grams (Subword):**
212
 
213
  | Rank | N-gram | Count |
214
  |------|--------|-------|
215
+ | 1 | `_ — _` | 342,728 |
216
+ | 2 | `ш ы в` | 149,577 |
217
+ | 3 | `ы в _` | 121,922 |
218
+ | 4 | `_ ю х` | 94,718 |
219
+ | 5 | `т е р` | 86,508 |
220
 
221
  **4-grams (Subword):**
222
 
223
  | Rank | N-gram | Count |
224
  |------|--------|-------|
225
+ | 1 | `ш ы в _` | 121,828 |
226
+ | 2 | `_ ш ы в` | 85,484 |
227
+ | 3 | `_ ю х а` | 76,914 |
228
+ | 4 | `ю х а н` | 63,379 |
229
+ | 5 | `х а н ш` | 63,281 |
230
+
231
+ **5-grams (Subword):**
232
+
233
+ | Rank | N-gram | Count |
234
+ |------|--------|-------|
235
+ | 1 | `_ ш ы в _` | 83,923 |
236
+ | 2 | `ю х а н ш` | 63,268 |
237
+ | 3 | `х а н ш ы` | 63,265 |
238
+ | 4 | `а н ш ы в` | 63,263 |
239
+ | 5 | `_ ю х а н` | 62,475 |
240
 
241
 
242
  ### Key Findings
243
 
244
+ - **Best Perplexity:** 2-gram (subword) with 532
245
  - **Entropy Trend:** Decreases with larger n-grams (more predictable)
246
+ - **Coverage:** Top-1000 patterns cover ~25% of corpus
247
  - **Recommendation:** 4-gram or 5-gram for best predictive performance
248
 
249
  ---
 
259
 
260
  | Context | Variant | Avg Entropy | Perplexity | Branching Factor | Unique Contexts | Predictability |
261
  |---------|---------|-------------|------------|------------------|-----------------|----------------|
262
+ | **1** | Word | 0.7800 | 1.717 | 5.34 | 352,836 | 22.0% |
263
+ | **1** | Subword | 0.6157 | 1.532 | 6.03 | 3,635 | 38.4% |
264
+ | **2** | Word | 0.1829 | 1.135 | 1.40 | 1,869,675 | 81.7% |
265
+ | **2** | Subword | 0.9040 | 1.871 | 6.19 | 21,903 | 9.6% |
266
+ | **3** | Word | 0.0525 | 1.037 | 1.09 | 2,591,084 | 94.7% |
267
+ | **3** | Subword | 0.8721 | 1.830 | 4.70 | 135,543 | 12.8% |
268
+ | **4** | Word | 0.0223 🏆 | 1.016 | 1.04 | 2,792,400 | 97.8% |
269
+ | **4** | Subword | 0.7095 | 1.635 | 3.14 | 636,890 | 29.1% |
270
 
271
  ### Generated Text Samples (Word-based)
272
 
 
274
 
275
  **Context Size 1:**
276
 
277
+ 1. `шыв гидрологи бассейн шыв шыв геоинформаци системин шыв федераци агентстви рф территорин шыв геоинфо...`
278
+ 2. `юханшыв двина печора шыв федераци агентстви рф экологи министерстви республикин ао коми республики т...`
279
+ 3. `в цене чем предпочитают вспоминать и дефекты зрения м советская энциклопедия в унисон с любашей леро...`
280
 
281
  **Context Size 2:**
282
 
283
+ 1. `шыв шыв тури обь иртыш шыв федераци агентстви рф территорин шыв геоинформаци системин шыв шыв тури о...`
284
+ 2. `территоринчи юханшыв рейн вестфали территорипе юхать юханшыв негус ях сулахай 13 км шыв шыв тури бас...`
285
+ 3. `территорипе юхать юханшыв мăн салым сулахай 220 км юхса юханшыв 12 км шыв шыв гидрологи бассейн том`
286
 
287
  **Context Size 3:**
288
 
289
+ 1. `федераци агентстви рф территорин шыв геоинформаци системин шыв шыв гидрологи гт бассейн том гт 15 гт...`
290
+ 2. `шыв федераци агентстви рф территорин шыв геоинформаци системин шыв шыв гидрологи бассейн том 15 3 рф...`
291
+ 3. `шыв геоинформаци системин шыв шыв гидрологи гт бассейн том гт 11 гт 1 рф экологи министерстви респуб...`
292
 
293
  **Context Size 4:**
294
 
295
+ 1. `шыв геоинформаци системин шыв шыв гидрологи гт бассейн том гт 03 гт 0 рф экологи министерстви ао рес...`
296
+ 2. `территорин шыв геоинформаци системин шыв шыв гидрологи бассейн том 15 3 рф экологи министерстви авто...`
297
+ 3. `геоинформаци системин шыв шыв гидрологи гт бассейн том гт 03 гт 0 рф экологи министерстви ао республ...`
298
 
299
 
300
  ### Generated Text Samples (Subword-based)
 
303
 
304
  **Context Size 1:**
305
 
306
+ 1. `_—_фикулигинци_в`
307
+ 2. `а,_;_улслаки_пид`
308
+ 3. _каспалименияни`
309
 
310
  **Context Size 2:**
311
 
312
+ 1. `.__торф_тыслана_`
313
+ 2. `а_медилостви_тута`
314
+ 3. `и_йышши_баллина_з`
315
 
316
  **Context Size 3:**
317
 
318
+ 1. `_—_теминисем_астар`
319
+ 2. `шыв__мар_монтовол`
320
+ 3. `ыв_шыв._команицы:_`
321
 
322
  **Context Size 4:**
323
 
324
+ 1. `шыв_шыв_—_венгрла._`
325
+ 2. `_шыв_федераци_агент`
326
+ 3. `_юханшыв_шыв_геоинф`
327
 
328
 
329
  ### Key Findings
330
 
331
  - **Best Predictability:** Context-4 (word) with 97.8% predictability
332
  - **Branching Factor:** Decreases with context size (more deterministic)
333
+ - **Memory Trade-off:** Larger contexts require more storage (636,890 contexts)
334
  - **Recommendation:** Context-3 or Context-4 for text generation
335
 
336
  ---
 
346
 
347
  | Metric | Value |
348
  |--------|-------|
349
+ | Vocabulary Size | 149,054 |
350
+ | Total Tokens | 3,895,916 |
351
+ | Mean Frequency | 26.14 |
352
  | Median Frequency | 4 |
353
+ | Frequency Std Dev | 439.39 |
354
 
355
  ### Most Common Words
356
 
357
  | Rank | Word | Frequency |
358
  |------|------|-----------|
359
+ | 1 | шыв | 84,160 |
360
+ | 2 | юханшыв | 53,731 |
361
+ | 3 | в | 45,242 |
362
+ | 4 | и | 41,204 |
363
+ | 5 | с | 37,543 |
364
+ | 6 | тата | 34,625 |
365
+ | 7 | бассейн | 28,455 |
366
  | 8 | км | 25,026 |
367
+ | 9 | м | 24,932 |
368
+ | 10 | рф | 24,450 |
369
 
370
  ### Least Common Words (from vocabulary)
371
 
372
  | Rank | Word | Frequency |
373
  |------|------|-----------|
374
+ | 1 | дустлик | 2 |
375
+ | 2 | галляарал | 2 |
376
+ | 3 | зарбдар | 2 |
377
+ | 4 | джизакской | 2 |
378
+ | 5 | сардоба | 2 |
379
+ | 6 | баяут | 2 |
380
+ | 7 | хаваст | 2 |
381
+ | 8 | сырдарьинской | 2 |
382
+ | 9 | пайт | 2 |
383
+ | 10 | клинов | 2 |
384
 
385
  ### Zipf's Law Analysis
386
 
387
  | Metric | Value |
388
  |--------|-------|
389
+ | Zipf Coefficient | 1.0393 |
390
+ | R² (Goodness of Fit) | 0.997747 |
391
  | Adherence Quality | **excellent** |
392
 
393
  ### Coverage Analysis
394
 
395
  | Top N Words | Coverage |
396
  |-------------|----------|
397
+ | Top 100 | 30.0% |
398
  | Top 1,000 | 56.1% |
399
  | Top 5,000 | 72.5% |
400
+ | Top 10,000 | 79.0% |
401
 
402
  ### Key Findings
403
 
404
+ - **Zipf Compliance:** R²=0.9977 indicates excellent adherence to Zipf's law
405
+ - **High Frequency Dominance:** Top 100 words cover 30.0% of corpus
406
+ - **Long Tail:** 139,054 words needed for remaining 21.0% coverage
407
 
408
  ---
409
  ## 5. Word Embeddings Evaluation
 
419
 
420
  ### 5.1 Cross-Lingual Alignment
421
 
422
+ ![Alignment Quality](visualizations/embedding_alignment_quality.png)
423
+
424
+ ![Multilingual t-SNE](visualizations/embedding_tsne_multilingual.png)
425
 
426
 
427
  ### 5.2 Model Comparison
428
 
429
  | Model | Dimension | Isotropy | Semantic Density | Alignment R@1 | Alignment R@10 |
430
  |-------|-----------|----------|------------------|---------------|----------------|
431
+ | **mono_32d** | 32 | 0.8326 🏆 | 0.3463 | N/A | N/A |
432
+ | **mono_64d** | 64 | 0.8301 | 0.2835 | N/A | N/A |
433
+ | **mono_128d** | 128 | 0.7992 | 0.2278 | N/A | N/A |
434
+ | **aligned_32d** | 32 | 0.8326 | 0.3575 | 0.0120 | 0.1340 |
435
+ | **aligned_64d** | 64 | 0.8301 | 0.2722 | 0.0400 | 0.2360 |
436
+ | **aligned_128d** | 128 | 0.7992 | 0.2219 | 0.0680 | 0.3000 |
437
 
438
  ### Key Findings
439
 
440
+ - **Best Isotropy:** mono_32d with 0.8326 (more uniform distribution)
441
+ - **Semantic Density:** Average pairwise similarity of 0.2849. Lower values indicate better semantic separation.
442
+ - **Alignment Quality:** Aligned models achieve up to 6.8% R@1 in cross-lingual retrieval.
443
  - **Recommendation:** 128d aligned for best cross-lingual performance
444
 
445
  ---
446
  ## 6. Morphological Analysis (Experimental)
447
 
 
 
448
  This section presents an automated morphological analysis derived from the statistical divergence between word-level and subword-level models. By analyzing where subword predictability spikes and where word-level coverage fails, we can infer linguistic structures without supervised data.
449
 
450
  ### 6.1 Productivity & Complexity
451
 
452
  | Metric | Value | Interpretation | Recommendation |
453
  |--------|-------|----------------|----------------|
454
+ | Productivity Index | **5.000** | High morphological productivity | Reliable analysis |
455
+ | Idiomaticity Gap | **1.001** | High formulaic/idiomatic content | - |
456
 
457
  ### 6.2 Affix Inventory (Productive Units)
458
 
 
465
  #### Productive Suffixes
466
  | Suffix | Examples |
467
  |--------|----------|
468
+ | `-а` | курска, никсона, подвига |
469
+ | `-ен` | америкасен, слышен, судьясен |
470
+ | `-не` | взводне, очерксене, болгарине |
471
+ | `-ов` | резюков, коршунов, щенков |
472
+ | `-ем` | сикекенсем, символсем, перуанецсем |
473
+ | `-ий` | выступлений, парфентий, праславянский |
474
 
475
  ### 6.3 Bound Stems (Lexical Roots)
476
 
 
478
 
479
  | Stem | Cohesion | Substitutability | Examples |
480
  |------|----------|------------------|----------|
481
+ | `олог` | 2.08x | 173 contexts | геолог, пологи, эколог |
482
+ | `сейн` | 2.92x | 24 contexts | сейнер, хусейн, хасейн |
483
+ | `ссей` | 2.92x | 17 contexts | ессей, эссей, рассей |
484
+ | `огра` | 1.78x | 95 contexts | богра, ограды, ограда |
485
+ | `рито` | 2.46x | 26 contexts | ритон, крито, приток |
486
+ | `ншыв` | 2.79x | 17 contexts | юшаншыв, юханшыв, юханшыве |
487
+ | `ерри` | 2.45x | 22 contexts | черри, ферри, дерри |
488
+ | `орин` | 1.72x | 74 contexts | дорин, шорин, борин |
489
+ | `аншы` | 2.79x | 13 contexts | юшаншыв, юханшыв, юханшыве |
490
+ | `исте` | 1.81x | 57 contexts | листе, хистет, истерн |
491
+ | `блик` | 2.25x | 17 contexts | облик, облика, коблик |
492
+ | `нист` | 1.86x | 30 contexts | финист, пианист, капнист |
493
 
494
  ### 6.4 Affix Compatibility (Co-occurrence)
495
 
 
504
 
505
  | Word | Suggested Split | Confidence | Stem |
506
  |------|-----------------|------------|------|
507
+ | айсбергов | **`айсберг-ов`** | 4.5 | `айсберг` |
508
+ | фахрутдинов | **`фахрутдин-ов`** | 4.5 | `фахрутдин` |
509
+ | экономикине | **`экономики-не`** | 4.5 | `экономики` |
510
+ | пурнӑҫланине | **`пурнӑҫлани-не`** | 4.5 | `пурнӑҫлани` |
511
+ | ансамбльне | **`ансамбль-не`** | 4.5 | `ансамбль` |
512
+ | хрустальне | **`хрусталь-не`** | 4.5 | `хрусталь` |
513
+ | анатомине | **`анатоми-не`** | 4.5 | `анатоми` |
514
+ | инженеров | **`инженер-ов`** | 4.5 | `инженер` |
515
+ | багдасаров | **`багдасар-ов`** | 4.5 | `багдасар` |
516
+ | фотографий | **`фотограф-ий`** | 4.5 | `фотограф` |
517
+ | ассамблейине | **`ассамблейи-не`** | 4.5 | `ассамблейи` |
518
+ | символикине | **`символики-не`** | 4.5 | `символики` |
519
+ | бриллиантов | **`бриллиант-ов`** | 4.5 | `бриллиант` |
520
+ | кинокритиков | **`кинокритик-ов`** | 4.5 | `кинокритик` |
521
+ | наводнений | **`наводн-ен-ий`** | 3.0 | `наводн` |
522
 
523
  ### 6.6 Linguistic Interpretation
524
 
525
  > **Automated Insight:**
526
+ The language Chuvash shows high morphological productivity. The subword models are significantly more efficient than word models, suggesting a rich system of affixation or compounding.
527
+
528
+ > **Note on Idiomaticity:** The high Idiomaticity Gap suggests a large number of frequent multi-word expressions or formulaic sequences that are statistically distinct from their component parts.
529
 
530
  ---
531
  ## 7. Summary & Recommendations
 
536
 
537
  | Component | Recommended | Rationale |
538
  |-----------|-------------|-----------|
539
+ | Tokenizer | **64k BPE** | Best compression (3.78x) |
540
+ | N-gram | **2-gram** | Lowest perplexity (532) |
541
  | Markov | **Context-4** | Highest predictability (97.8%) |
542
  | Embeddings | **100d** | Balanced semantic capture and isotropy |
543
 
 
752
  ---
753
  *Generated by Wikilangs Models Pipeline*
754
 
755
+ *Report Date: 2026-01-03 23:50:11*
models/embeddings/aligned/cv_128d.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:130680d2248f57ae1c36fc9ab1e6e0eda0a7d2308787a32f6d268a9423f78b02
3
+ size 1096942926
models/embeddings/aligned/cv_128d.meta.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lang": "cv", "dim": 128, "max_seq_len": 512, "is_aligned": true}
models/embeddings/aligned/cv_128d.projection.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:695485e87ea1ae53971a035832321d9c5606a060d44573475d57fc389a4c22ce
3
+ size 65664
models/embeddings/aligned/cv_128d_metadata.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "language": "cv",
3
+ "dimension": 128,
4
+ "version": "aligned",
5
+ "hub_language": "en",
6
+ "seed_vocab_size": 9371,
7
+ "vocab_size": 69656
8
+ }
models/embeddings/aligned/cv_32d.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8cdc753305931cf408774c9aa08526d34b9612f3481c0d05a6e024ad61935fba
3
+ size 275447118
models/embeddings/aligned/cv_32d.meta.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lang": "cv", "dim": 32, "max_seq_len": 512, "is_aligned": true}
models/embeddings/aligned/cv_32d.projection.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4367b3967495e33f979b6c501a98e9c8f8446e8335038b833c232be2fb9adfa8
3
+ size 4224
models/embeddings/aligned/cv_32d_metadata.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "language": "cv",
3
+ "dimension": 32,
4
+ "version": "aligned",
5
+ "hub_language": "en",
6
+ "seed_vocab_size": 9371,
7
+ "vocab_size": 69656
8
+ }
models/embeddings/aligned/cv_64d.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97e63f8b115a259d1af448ee3adf665e39112df91110cabb11dc33e5522f85a6
3
+ size 549279054
models/embeddings/aligned/cv_64d.meta.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lang": "cv", "dim": 64, "max_seq_len": 512, "is_aligned": true}
models/embeddings/aligned/cv_64d.projection.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d36328a9a375aba46f68707b77df6819a099e24dd96a52370a42a484641469cd
3
+ size 16512
models/embeddings/aligned/cv_64d_metadata.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "language": "cv",
3
+ "dimension": 64,
4
+ "version": "aligned",
5
+ "hub_language": "en",
6
+ "seed_vocab_size": 9371,
7
+ "vocab_size": 69656
8
+ }
models/embeddings/monolingual/cv_128d.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6cb5a0b37d9ab0f0662eb5bebfad6713c0f0648b4a42c5b0743b36462becfd73
3
- size 1096617351
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:130680d2248f57ae1c36fc9ab1e6e0eda0a7d2308787a32f6d268a9423f78b02
3
+ size 1096942926
models/embeddings/monolingual/cv_128d_metadata.json CHANGED
@@ -11,5 +11,5 @@
11
  "encoding_method": "rope",
12
  "dim": 128
13
  },
14
- "vocab_size": 69346
15
  }
 
11
  "encoding_method": "rope",
12
  "dim": 128
13
  },
14
+ "vocab_size": 69656
15
  }
models/embeddings/monolingual/cv_32d.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:70007f295deadd64b95415e5970e57d83f6d92216140317dc2902789bf6da1de
3
- size 275359623
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8cdc753305931cf408774c9aa08526d34b9612f3481c0d05a6e024ad61935fba
3
+ size 275447118
models/embeddings/monolingual/cv_32d_metadata.json CHANGED
@@ -11,5 +11,5 @@
11
  "encoding_method": "rope",
12
  "dim": 32
13
  },
14
- "vocab_size": 69346
15
  }
 
11
  "encoding_method": "rope",
12
  "dim": 32
13
  },
14
+ "vocab_size": 69656
15
  }
models/embeddings/monolingual/cv_64d.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a5e77b5b9974a5a7347da624921ef4bd323fb01d62ffb62e50b9d2ca69a948e4
3
- size 549112199
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97e63f8b115a259d1af448ee3adf665e39112df91110cabb11dc33e5522f85a6
3
+ size 549279054
models/embeddings/monolingual/cv_64d_metadata.json CHANGED
@@ -11,5 +11,5 @@
11
  "encoding_method": "rope",
12
  "dim": 64
13
  },
14
- "vocab_size": 69346
15
  }
 
11
  "encoding_method": "rope",
12
  "dim": 64
13
  },
14
+ "vocab_size": 69656
15
  }
models/subword_markov/cv_markov_ctx1_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e40043954b83a926cb69dce26f1f67caf4075d1babdc98a7a3543ca0afb6275b
3
- size 174799
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3346d398d4005b40751df8506c1395d6f041d0b99c81d1c518e557a581ce287
3
+ size 174960
models/subword_markov/cv_markov_ctx1_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 1,
3
  "variant": "subword",
4
  "language": "cv",
5
- "unique_contexts": 3627,
6
- "total_transitions": 29032785
7
  }
 
2
  "context_size": 1,
3
  "variant": "subword",
4
  "language": "cv",
5
+ "unique_contexts": 3635,
6
+ "total_transitions": 29154520
7
  }
models/subword_markov/cv_markov_ctx2_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:da927115de7638fb21f20a2ffe9e1afb50762a8897417d080a3fbfbb98fede29
3
- size 1108384
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96e19271842803fd5df5bc0ebaa3af87e64fd6cbe51dbb84fbf1d48de884da92
3
+ size 1112248
models/subword_markov/cv_markov_ctx2_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 2,
3
  "variant": "subword",
4
  "language": "cv",
5
- "unique_contexts": 21896,
6
- "total_transitions": 28978297
7
  }
 
2
  "context_size": 2,
3
  "variant": "subword",
4
  "language": "cv",
5
+ "unique_contexts": 21903,
6
+ "total_transitions": 29099815
7
  }
models/subword_markov/cv_markov_ctx3_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8cc4344364830518084929051b4ae7fb28c1adcdaa163605965b74168fdbce0e
3
- size 4746572
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51d6967be1a56e0bbb759070107c2ad61a4b36bb6f31f02a1953134b7eaf92e4
3
+ size 4743832
models/subword_markov/cv_markov_ctx3_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 3,
3
  "variant": "subword",
4
  "language": "cv",
5
- "unique_contexts": 135774,
6
- "total_transitions": 28923809
7
  }
 
2
  "context_size": 3,
3
  "variant": "subword",
4
  "language": "cv",
5
+ "unique_contexts": 135543,
6
+ "total_transitions": 29045110
7
  }
models/subword_markov/cv_markov_ctx4_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bbad63be2ef01e2c8ae1ec9ea81d1bd588420f175327541dc01a0c9c80d33097
3
- size 16746581
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0836a4e42204ea786562096268d49a9fad133bda16cdbbad35bc4aa850e6ba44
3
+ size 16755694
models/subword_markov/cv_markov_ctx4_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 4,
3
  "variant": "subword",
4
  "language": "cv",
5
- "unique_contexts": 637656,
6
- "total_transitions": 28869321
7
  }
 
2
  "context_size": 4,
3
  "variant": "subword",
4
  "language": "cv",
5
+ "unique_contexts": 636890,
6
+ "total_transitions": 28990405
7
  }
models/subword_ngram/cv_2gram_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a6dc007ee19af6fa94d26edbb4ded18e4913986e603f56c0eb0a613928d49afa
3
- size 111624
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf921911e3ac02a40136c41c3cc4c4180d17b05f1c96145dc12d2f915ff427ad
3
+ size 111747
models/subword_ngram/cv_2gram_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "n": 2,
3
  "variant": "subword",
4
  "language": "cv",
5
- "unique_ngrams": 7915,
6
- "total_ngrams": 29032785
7
  }
 
2
  "n": 2,
3
  "variant": "subword",
4
  "language": "cv",
5
+ "unique_ngrams": 7908,
6
+ "total_ngrams": 29154520
7
  }
models/subword_ngram/cv_3gram_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5a53062fcd868eabb3bedaef471787c15f34a1e4d7bd64541430cfad78fca8d3
3
- size 879358
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:432c05f064a7e8b8652618f319730cbb81e70c2dfb55083be821ec9115b6dbe7
3
+ size 878040
models/subword_ngram/cv_3gram_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "n": 3,
3
  "variant": "subword",
4
  "language": "cv",
5
- "unique_ngrams": 69443,
6
- "total_ngrams": 28978297
7
  }
 
2
  "n": 3,
3
  "variant": "subword",
4
  "language": "cv",
5
+ "unique_ngrams": 69351,
6
+ "total_ngrams": 29099815
7
  }
models/subword_ngram/cv_4gram_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8ec5a505fdfed8bff6da064dc01c77eb56f326b342a958243b49281b2add80c1
3
- size 4733643
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ab8068997235bcf52d4d43c2a5acd248a6d44be2e34b297dee57ac2a61e2939
3
+ size 4727224
models/subword_ngram/cv_4gram_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "n": 4,
3
  "variant": "subword",
4
  "language": "cv",
5
- "unique_ngrams": 379120,
6
- "total_ngrams": 28923809
7
  }
 
2
  "n": 4,
3
  "variant": "subword",
4
  "language": "cv",
5
+ "unique_ngrams": 378926,
6
+ "total_ngrams": 29045110
7
  }
models/subword_ngram/cv_5gram_subword.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0536222fe93e097cd31d45bcd1240bb299e6063dc26a96c86a46b7ef8e3f541
3
+ size 13563004
models/subword_ngram/cv_5gram_subword_metadata.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "n": 5,
3
+ "variant": "subword",
4
+ "language": "cv",
5
+ "unique_ngrams": 1007721,
6
+ "total_ngrams": 28990405
7
+ }
models/tokenizer/cv_tokenizer_16k.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:47ce826b88b912a0ee22e1317e6f0258e7665b2a8c708987a2abda3b35abceff
3
- size 564983
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:640e17b3c3cd055630d02fcfc5512991276e9037fb2a3e5d9930106f36d3504d
3
+ size 564943
models/tokenizer/cv_tokenizer_16k.vocab CHANGED
The diff for this file is too large to render. See raw diff
 
models/tokenizer/cv_tokenizer_32k.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4f2de0762f3ba8f96fbc31f699f83e9b5dc454f42975ac96462ea30d55c9456f
3
- size 912506
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16b0ac2359544b071959cae6ac454d8543c07b671d234d4901fcaead35fd0b0d
3
+ size 912892
models/tokenizer/cv_tokenizer_32k.vocab CHANGED
The diff for this file is too large to render. See raw diff
 
models/tokenizer/cv_tokenizer_64k.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5877d8c09381815bf1acb251e4535e07c5b0157e1b804e3dd591e8925c8268d9
3
- size 1629877
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54d00d4fc5c53f631ee8370abbab67b321b928713b3479bb02d4d6031446f747
3
+ size 1631182
models/tokenizer/cv_tokenizer_64k.vocab CHANGED
The diff for this file is too large to render. See raw diff
 
models/tokenizer/cv_tokenizer_8k.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:64845d052561c0fd324adfbc8fd0d900a63c92141c555143df24fa74a9d7b46e
3
- size 398474
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97a9c78ba41c4fc4820e32fef12108c285f66f628f6c63d23cd3de9331e9428e
3
+ size 398499
models/tokenizer/cv_tokenizer_8k.vocab CHANGED
The diff for this file is too large to render. See raw diff
 
models/vocabulary/cv_vocabulary.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fc7aca4a1d1d56f053bd0d6a123e4ef044a92c56295fcfe1be7fd13c9ca9e750
3
- size 2661487
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e837c9c3778915d77ff47bb28bd5e206258e5ffd40f04519456428cbd5bc0005
3
+ size 2667305
models/vocabulary/cv_vocabulary_metadata.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "language": "cv",
3
- "vocabulary_size": 148629,
4
  "variant": "full",
5
  "statistics": {
6
- "type_token_ratio": 0.08645308234335648,
7
  "coverage": {
8
- "top_100": 0.28553243324755495,
9
- "top_1000": 0.5333145325860036,
10
- "top_5000": 0.6888109114837315,
11
- "top_10000": 0.7510137225871395
12
  },
13
- "hapax_count": 204527,
14
- "hapax_ratio": 0.5791406630497571,
15
- "total_documents": 54488
16
  }
17
  }
 
1
  {
2
  "language": "cv",
3
+ "vocabulary_size": 149054,
4
  "variant": "full",
5
  "statistics": {
6
+ "type_token_ratio": 0.08632064084275211,
7
  "coverage": {
8
+ "top_100": 0.2853232866356975,
9
+ "top_1000": 0.5330236414401892,
10
+ "top_5000": 0.6887052684199617,
11
+ "top_10000": 0.7508867673774949
12
  },
13
+ "hapax_count": 204934,
14
+ "hapax_ratio": 0.5789292292394093,
15
+ "total_documents": 54705
16
  }
17
  }
models/word_markov/cv_markov_ctx1_word.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c12896d7bb4321044e1d15a1f2c739138431e35a2c716c4bae85b032486c8b29
3
- size 23653393
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c9a42680ec25e46b6e314a284536406b993337cb8b1c4daca1f33227ddec8e6
3
+ size 23816242
models/word_markov/cv_markov_ctx1_word_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 1,
3
  "variant": "word",
4
  "language": "cv",
5
- "unique_contexts": 352008,
6
- "total_transitions": 4030456
7
  }
 
2
  "context_size": 1,
3
  "variant": "word",
4
  "language": "cv",
5
+ "unique_contexts": 352836,
6
+ "total_transitions": 4046145
7
  }
models/word_markov/cv_markov_ctx2_word.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:04538a1a074396c3e2dba751a45663e0683f8dd3242406fc2c55ff9ab7c488ea
3
- size 58306430
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:81e59ba49e08be0159d5763ef1440f0c2fbd0076b4cefac0d569414c9327edc5
3
+ size 58581375
models/word_markov/cv_markov_ctx2_word_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 2,
3
  "variant": "word",
4
  "language": "cv",
5
- "unique_contexts": 1864001,
6
- "total_transitions": 3975968
7
  }
 
2
  "context_size": 2,
3
  "variant": "word",
4
  "language": "cv",
5
+ "unique_contexts": 1869675,
6
+ "total_transitions": 3991440
7
  }