omarkamali commited on
Commit
fd98249
·
verified ·
1 Parent(s): 7a41473

Upload all models and assets for ady (latest)

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +1 -0
  2. README.md +214 -178
  3. models/embeddings/aligned/ady_128d.bin +3 -0
  4. models/embeddings/aligned/ady_128d.meta.json +1 -0
  5. models/embeddings/aligned/ady_128d.projection.npy +3 -0
  6. models/embeddings/aligned/ady_128d_metadata.json +8 -0
  7. models/embeddings/aligned/ady_32d.bin +3 -0
  8. models/embeddings/aligned/ady_32d.meta.json +1 -0
  9. models/embeddings/aligned/ady_32d.projection.npy +3 -0
  10. models/embeddings/aligned/ady_32d_metadata.json +8 -0
  11. models/embeddings/aligned/ady_64d.bin +3 -0
  12. models/embeddings/aligned/ady_64d.meta.json +1 -0
  13. models/embeddings/aligned/ady_64d.projection.npy +3 -0
  14. models/embeddings/aligned/ady_64d_metadata.json +8 -0
  15. models/embeddings/monolingual/ady_128d.bin +2 -2
  16. models/embeddings/monolingual/ady_128d_metadata.json +1 -1
  17. models/embeddings/monolingual/ady_32d.bin +2 -2
  18. models/embeddings/monolingual/ady_32d_metadata.json +1 -1
  19. models/embeddings/monolingual/ady_64d.bin +2 -2
  20. models/embeddings/monolingual/ady_64d_metadata.json +1 -1
  21. models/subword_markov/ady_markov_ctx1_subword.parquet +2 -2
  22. models/subword_markov/ady_markov_ctx1_subword_metadata.json +2 -2
  23. models/subword_markov/ady_markov_ctx2_subword.parquet +2 -2
  24. models/subword_markov/ady_markov_ctx2_subword_metadata.json +2 -2
  25. models/subword_markov/ady_markov_ctx3_subword.parquet +2 -2
  26. models/subword_markov/ady_markov_ctx3_subword_metadata.json +2 -2
  27. models/subword_markov/ady_markov_ctx4_subword.parquet +2 -2
  28. models/subword_markov/ady_markov_ctx4_subword_metadata.json +2 -2
  29. models/subword_ngram/ady_2gram_subword.parquet +2 -2
  30. models/subword_ngram/ady_2gram_subword_metadata.json +2 -2
  31. models/subword_ngram/ady_3gram_subword.parquet +2 -2
  32. models/subword_ngram/ady_3gram_subword_metadata.json +2 -2
  33. models/subword_ngram/ady_4gram_subword.parquet +2 -2
  34. models/subword_ngram/ady_4gram_subword_metadata.json +2 -2
  35. models/subword_ngram/ady_5gram_subword.parquet +3 -0
  36. models/subword_ngram/ady_5gram_subword_metadata.json +7 -0
  37. models/tokenizer/ady_tokenizer_16k.model +2 -2
  38. models/tokenizer/ady_tokenizer_16k.vocab +0 -0
  39. models/tokenizer/ady_tokenizer_32k.model +2 -2
  40. models/tokenizer/ady_tokenizer_32k.vocab +0 -0
  41. models/tokenizer/ady_tokenizer_8k.model +2 -2
  42. models/tokenizer/ady_tokenizer_8k.vocab +0 -0
  43. models/vocabulary/ady_vocabulary.parquet +2 -2
  44. models/vocabulary/ady_vocabulary_metadata.json +9 -9
  45. models/word_markov/ady_markov_ctx1_word.parquet +2 -2
  46. models/word_markov/ady_markov_ctx1_word_metadata.json +2 -2
  47. models/word_markov/ady_markov_ctx2_word.parquet +2 -2
  48. models/word_markov/ady_markov_ctx2_word_metadata.json +2 -2
  49. models/word_markov/ady_markov_ctx3_word.parquet +2 -2
  50. models/word_markov/ady_markov_ctx3_word_metadata.json +2 -2
.gitattributes CHANGED
@@ -40,3 +40,4 @@ visualizations/tsne_sentences.png filter=lfs diff=lfs merge=lfs -text
40
  visualizations/tsne_words.png filter=lfs diff=lfs merge=lfs -text
41
  visualizations/zipf_law.png filter=lfs diff=lfs merge=lfs -text
42
  visualizations/ngram_coverage.png filter=lfs diff=lfs merge=lfs -text
 
 
40
  visualizations/tsne_words.png filter=lfs diff=lfs merge=lfs -text
41
  visualizations/zipf_law.png filter=lfs diff=lfs merge=lfs -text
42
  visualizations/ngram_coverage.png filter=lfs diff=lfs merge=lfs -text
43
+ visualizations/embedding_tsne_multilingual.png filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  language: ady
3
- language_name: ADY
4
  language_family: caucasian_northwest
5
  tags:
6
  - wikilangs
@@ -10,11 +10,21 @@ tags:
10
  - n-gram
11
  - markov
12
  - wikipedia
 
 
 
 
 
 
 
 
 
 
13
  - monolingual
14
  - family-caucasian_northwest
15
  license: mit
16
  library_name: wikilangs
17
- pipeline_tag: feature-extraction
18
  datasets:
19
  - omarkamali/wikipedia-monthly
20
  dataset_info:
@@ -23,20 +33,20 @@ dataset_info:
23
  metrics:
24
  - name: best_compression_ratio
25
  type: compression
26
- value: 4.231
27
  - name: best_isotropy
28
  type: isotropy
29
- value: 0.4850
30
  - name: vocabulary_size
31
  type: vocab
32
  value: 0
33
  generated: 2026-01-03
34
  ---
35
 
36
- # ADY - Wikilangs Models
37
  ## Comprehensive Research Report & Full Ablation Study
38
 
39
- This repository contains NLP models trained and evaluated by Wikilangs, specifically on **ADY** Wikipedia data.
40
  We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and word embeddings.
41
 
42
  ## 📋 Repository Contents
@@ -60,7 +70,7 @@ We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and
60
  - [3. Markov Chain Evaluation](#3-markov-chain-evaluation)
61
  - [4. Vocabulary Analysis](#4-vocabulary-analysis)
62
  - [5. Word Embeddings Evaluation](#5-word-embeddings-evaluation)
63
- - [6. Morphological Analysis (Experimental)](#6-morphological-analysis)
64
  - [7. Summary & Recommendations](#7-summary--recommendations)
65
  - [Metrics Glossary](#appendix-metrics-glossary--interpretation-guide)
66
  - [Visualizations Index](#visualizations-index)
@@ -80,43 +90,43 @@ We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and
80
 
81
  | Vocab Size | Compression | Avg Token Len | UNK Rate | Total Tokens |
82
  |------------|-------------|---------------|----------|--------------|
83
- | **8k** | 3.442x | 3.45 | 0.1638% | 134,283 |
84
- | **16k** | 3.798x | 3.80 | 0.1808% | 121,676 |
85
- | **32k** | 4.231x 🏆 | 4.24 | 0.2014% | 109,215 |
86
 
87
  ### Tokenization Examples
88
 
89
  Below are sample sentences tokenized with each vocabulary size:
90
 
91
- **Sample 1:** `Киев Украинэ и Нэбгырэ млн 2.9 фэдиз дэс. Къалэм и – Кличко Виталий Владимир ы...`
92
 
93
  | Vocab | Tokens | Count |
94
  |-------|--------|-------|
95
- | 8k | `▁киев ▁— ▁украинэ ▁и ▁нэбгырэ ▁млн 2 . 9 ... (+34 more)` | 44 |
96
- | 16k | `▁киев ▁— ▁украинэ ▁и ▁нэбгырэ ▁млн 2 . 9 ... (+25 more)` | 35 |
97
- | 32k | `▁киев ▁— ▁украинэ ▁и ▁нэбгырэ ▁млн 2 . 9 ... (+17 more)` | 27 |
98
 
99
- **Sample 2:** `Пётровице Полшэм и Нэбгырэ 352 фэдиз Bank Danych Lokalnych. ТехьэпӀэхэр Пётров...`
100
 
101
  | Vocab | Tokens | Count |
102
  |-------|--------|-------|
103
- | 8k | `▁пёт ров ице ▁– ▁пол шэм ▁и ▁нэбгырэ 3 ... (+27 more)` | 37 |
104
- | 16k | `▁пётровице ▁– ▁полшэм ▁и ▁нэбгырэ 3 5 2 ▁фэдиз ... (+16 more)` | 26 |
105
- | 32k | `▁пётровице ▁– ▁полшэм ▁и ▁нэбгырэ 3 5 2 ▁фэдиз ... (+13 more)` | 23 |
106
 
107
- **Sample 3:** `пшъэшъэ пшъашъэхэм алъыплъэу, якъэшъон, языгъэпсэфын гъунэ алъызыфэу джэгур пш...`
108
 
109
  | Vocab | Tokens | Count |
110
  |-------|--------|-------|
111
- | 8k | `▁пшъэшъэ ▁– ▁пшъашъэ хэм ▁алъ ып лъэу , ▁я къэ ... (+17 more)` | 27 |
112
- | 16k | `▁пшъэшъэ ▁– ▁пшъашъэ хэм ▁алъыплъэу , ▁я къэ шъо н ... (+10 more)` | 20 |
113
- | 32k | `▁пшъэшъэ ▁– ▁пшъашъэхэм ▁алъыплъэу , ▁якъэшъон , ▁языгъэпсэфын ▁гъунэ ▁алъызыфэу ... (+3 more)` | 13 |
114
 
115
 
116
  ### Key Findings
117
 
118
- - **Best Compression:** 32k achieves 4.231x compression
119
- - **Lowest UNK Rate:** 8k with 0.1638% unknown tokens
120
  - **Trade-off:** Larger vocabularies improve compression but increase model size
121
  - **Recommendation:** 32k vocabulary provides optimal balance for production use
122
 
@@ -133,12 +143,14 @@ Below are sample sentences tokenized with each vocabulary size:
133
 
134
  | N-gram | Variant | Perplexity | Entropy | Unique N-grams | Top-100 Coverage | Top-1000 Coverage |
135
  |--------|---------|------------|---------|----------------|------------------|-------------------|
136
- | **2-gram** | Word | 418 | 8.71 | 593 | 45.3% | 100.0% |
137
- | **2-gram** | Subword | 399 🏆 | 8.64 | 2,072 | 57.0% | 97.4% |
138
- | **3-gram** | Word | 706 | 9.46 | 922 | 33.9% | 100.0% |
139
- | **3-gram** | Subword | 2,788 | 11.44 | 11,614 | 24.5% | 65.1% |
140
- | **4-gram** | Word | 2,848 | 11.48 | 3,264 | 13.1% | 44.3% |
141
- | **4-gram** | Subword | 10,651 | 13.38 | 35,316 | 12.4% | 39.6% |
 
 
142
 
143
  ### Top 5 N-grams by Size
144
 
@@ -146,11 +158,11 @@ Below are sample sentences tokenized with each vocabulary size:
146
 
147
  | Rank | N-gram | Count |
148
  |------|--------|-------|
149
- | 1 | `нэбгырэ млн` | 169 |
150
  | 2 | `къехъу щэпсэу` | 104 |
151
- | 3 | `картым тетэу` | 100 |
152
- | 4 | къехъу` | 89 |
153
- | 5 | `дло м` | 87 |
154
 
155
  **3-grams (Word):**
156
 
@@ -158,7 +170,7 @@ Below are sample sentences tokenized with each vocabulary size:
158
  |------|--------|-------|
159
  | 1 | `м къехъу щэпсэу` | 76 |
160
  | 2 | `къехъу щэпсэу хэгэгум` | 70 |
161
- | 3 | `адыгэ республикэм и` | 48 |
162
  | 4 | `дло м хахьэ` | 44 |
163
  | 5 | `м хахьэ хэгъэгу` | 39 |
164
 
@@ -172,42 +184,62 @@ Below are sample sentences tokenized with each vocabulary size:
172
  | 4 | `америкэм ит къэралыгъу къэлэ` | 19 |
173
  | 5 | `азием ит къэралыгъу къэлэ` | 18 |
174
 
 
 
 
 
 
 
 
 
 
 
175
  **2-grams (Subword):**
176
 
177
  | Rank | N-gram | Count |
178
  |------|--------|-------|
179
- | 1 | `г ъ` | 9,349 |
180
- | 2 | `ъ э` | 9,255 |
181
- | 3 | `э _` | 8,719 |
182
- | 4 | `м _` | 7,823 |
183
- | 5 | `э р` | 6,778 |
184
 
185
  **3-grams (Subword):**
186
 
187
  | Rank | N-gram | Count |
188
  |------|--------|-------|
189
- | 1 | `г ъ э` | 4,967 |
190
- | 2 | `_ к ъ` | 4,149 |
191
- | 3 | `э м _` | 3,582 |
192
- | 4 | `ы г ъ` | 3,357 |
193
- | 5 | `э р _` | 3,016 |
194
 
195
  **4-grams (Subword):**
196
 
197
  | Rank | N-gram | Count |
198
  |------|--------|-------|
199
- | 1 | `ы г ъ э` | 1,903 |
200
- | 2 | `х э р _` | 1,450 |
201
- | 3 | `а г ъ э` | 1,351 |
202
- | 4 | `х э м _` | 1,305 |
203
  | 5 | `_ к ъ э` | 1,289 |
204
 
 
 
 
 
 
 
 
 
 
 
205
 
206
  ### Key Findings
207
 
208
- - **Best Perplexity:** 2-gram (subword) with 399
209
  - **Entropy Trend:** Decreases with larger n-grams (more predictable)
210
- - **Coverage:** Top-1000 patterns cover ~40% of corpus
211
  - **Recommendation:** 4-gram or 5-gram for best predictive performance
212
 
213
  ---
@@ -223,14 +255,14 @@ Below are sample sentences tokenized with each vocabulary size:
223
 
224
  | Context | Variant | Avg Entropy | Perplexity | Branching Factor | Unique Contexts | Predictability |
225
  |---------|---------|-------------|------------|------------------|-----------------|----------------|
226
- | **1** | Word | 0.4365 | 1.353 | 2.10 | 22,306 | 56.3% |
227
- | **1** | Subword | 1.4909 | 2.811 | 10.56 | 410 | 0.0% |
228
- | **2** | Word | 0.0764 | 1.054 | 1.12 | 46,305 | 92.4% |
229
- | **2** | Subword | 1.1481 | 2.216 | 5.61 | 4,325 | 0.0% |
230
- | **3** | Word | 0.0240 | 1.017 | 1.03 | 51,243 | 97.6% |
231
- | **3** | Subword | 0.7541 | 1.687 | 2.97 | 24,260 | 24.6% |
232
- | **4** | Word | 0.0128 🏆 | 1.009 | 1.02 | 52,387 | 98.7% |
233
- | **4** | Subword | 0.4304 | 1.348 | 1.86 | 72,077 | 57.0% |
234
 
235
  ### Generated Text Samples (Word-based)
236
 
@@ -238,27 +270,27 @@ Below are text samples generated from each word-based Markov chain model:
238
 
239
  **Context Size 1:**
240
 
241
- 1. `и 99 86 561 3 щагъэпсыгъ ахэр къуаджэ адыгэ алфавитэу зэхигъэуцуа гъэм ди лъэхъэнэм хухуабжэ зэкӏэпс...`
242
- 2. `адыгэ автоном хэку британиешхо пачъыхьыгъо темыр скотландием уэлсым пэгъунэгъу къэлэ нейпьидо мьянмэ...`
243
- 3. `м къехъу щэпсэу къэралыгъом икъэлэ хъугъэ шӏагъэхэм ащ ипшъэкӏэрэ лъэныкъо щыӏ сауд арабие бахрейн к...`
244
 
245
  **Context Size 2:**
246
 
247
- 1. `нэбгырэ млн 13 м паплъэхэзэ пчэдыжьым сыхьатыр 11 м машинэм дахэхэр къежьэхи исполкомым а ябыракъэу ...`
248
- 2. `къехъу щэпсэу хэгэгум 718 км китаибзэ англыбзэ малаибзэ тамилыбзэ дло м ез м и кандидат хэгъэгу фили...`
249
- 3. `картым тетэу экуадор къыблэ америкэм ит къэралыгъу къэлэ братиславэ нэбгырэ млн 8 8 фэдиз щэпсэу хэг...`
250
 
251
  **Context Size 3:**
252
 
253
- 1. `м къехъу щэпсэу хэгэгум 23 200 км арапыбзэ францыбзэ къэрал мохаммед ульд абдель азиз гуадзэр яхья у...`
254
- 2. `къехъу щэпсэу хэгэгум чӏырэу иӏэр 17 820 км бзэшъхьаӏэр арапыбз дло м хахьэ хэгъэгу гувернатор петер...`
255
- 3. `адыгэ республикэм и теуцуожь къедзыгъор къыгот краснодар краим и къэлэ нэбгырэ млн 1 3 фэдиз дэс къэ...`
256
 
257
  **Context Size 4:**
258
 
259
- 1. `м къехъу щэпсэу хэгэгум 82 880 км арапыбз дло м хахьэ хэгъэгу лӏышъхьэр сабах я 6 аль ахьмэд аль`
260
- 2. `дло м хахьэ хэгъэгу хамед бен исса аль халифа хэгъэгу тхьаматэр халифа бен салман аль халифа географ...`
261
- 3. `еуропэм хэт къэралыгъу къэлэ софие нэбгырэ млн 7 м къехъу щэпсэу хэгэгум 198 500 км кыргызыбзэрэ уры...`
262
 
263
 
264
  ### Generated Text Samples (Subword-based)
@@ -267,34 +299,34 @@ Below are text samples generated from each subword-based Markov chain model:
267
 
268
  **Context Size 1:**
269
 
270
- 1. `_хъ._зэтм_ьэлыбе`
271
- 2. `эмыны_ар,_е,_щӏэ`
272
- 3. `ыхьэ_хъайстоджау`
273
 
274
  **Context Size 2:**
275
 
276
- 1. `гъ._грайономафедо`
277
- 2. `ъэгугъэзетэм_зэра`
278
- 3. `э_сырикэмрэ_хылъэ`
279
 
280
  **Context Size 3:**
281
 
282
- 1. `гъэлэ_зинскэ_хы_фэ`
283
- 2. `_къикӏым_научнэр_б`
284
- 3. `эм_гу_рэ_исэугъэзе`
285
 
286
  **Context Size 4:**
287
 
288
- 1. `ыгъэ_поясхэм_сурэтх`
289
- 2. `хэр_зыдэщыӏу_ыкӏи_и`
290
- 3. `агъэцэкӏэ_иӏэм_итіо`
291
 
292
 
293
  ### Key Findings
294
 
295
  - **Best Predictability:** Context-4 (word) with 98.7% predictability
296
  - **Branching Factor:** Decreases with context size (more deterministic)
297
- - **Memory Trade-off:** Larger contexts require more storage (72,077 contexts)
298
  - **Recommendation:** Context-3 or Context-4 for text generation
299
 
300
  ---
@@ -310,64 +342,64 @@ Below are text samples generated from each subword-based Markov chain model:
310
 
311
  | Metric | Value |
312
  |--------|-------|
313
- | Vocabulary Size | 7,032 |
314
- | Total Tokens | 44,503 |
315
- | Mean Frequency | 6.33 |
316
  | Median Frequency | 3 |
317
- | Frequency Std Dev | 22.13 |
318
 
319
  ### Most Common Words
320
 
321
  | Rank | Word | Frequency |
322
  |------|------|-----------|
323
- | 1 | и | 1,013 |
324
- | 2 | адыгэ | 666 |
325
- | 3 | м | 489 |
326
- | 4 | илъэсым | 398 |
327
  | 5 | ащ | 391 |
328
- | 6 | я | 309 |
329
- | 7 | ары | 271 |
330
- | 8 | нэбгырэ | 247 |
331
- | 9 | а | 243 |
332
- | 10 | ыкӏи | 211 |
333
 
334
  ### Least Common Words (from vocabulary)
335
 
336
  | Rank | Word | Frequency |
337
  |------|------|-----------|
338
- | 1 | рсфср | 2 |
339
- | 2 | серийнэ | 2 |
340
- | 3 | ныбжьыкӏэхэри | 2 |
341
- | 4 | зэратебэнагъэр | 2 |
342
- | 5 | хираганэ | 2 |
343
- | 6 | катаканэ | 2 |
344
- | 7 | сербыбзэм | 2 |
345
- | 8 | къыздикӏыгъэр | 2 |
346
- | 9 | тыванбзэ | 2 |
347
- | 10 | къызыл | 2 |
348
 
349
  ### Zipf's Law Analysis
350
 
351
  | Metric | Value |
352
  |--------|-------|
353
- | Zipf Coefficient | 0.7821 |
354
- | R² (Goodness of Fit) | 0.977951 |
355
  | Adherence Quality | **excellent** |
356
 
357
  ### Coverage Analysis
358
 
359
  | Top N Words | Coverage |
360
  |-------------|----------|
361
- | Top 100 | 29.3% |
362
- | Top 1,000 | 60.6% |
363
- | Top 5,000 | 90.9% |
364
  | Top 10,000 | 0.0% |
365
 
366
  ### Key Findings
367
 
368
- - **Zipf Compliance:** R²=0.9780 indicates excellent adherence to Zipf's law
369
- - **High Frequency Dominance:** Top 100 words cover 29.3% of corpus
370
- - **Long Tail:** -2,968 words needed for remaining 100.0% coverage
371
 
372
  ---
373
  ## 5. Word Embeddings Evaluation
@@ -383,37 +415,40 @@ Below are text samples generated from each subword-based Markov chain model:
383
 
384
  ### 5.1 Cross-Lingual Alignment
385
 
386
- > *Note: Multilingual alignment visualization not available for this language.*
 
 
387
 
388
 
389
  ### 5.2 Model Comparison
390
 
391
  | Model | Dimension | Isotropy | Semantic Density | Alignment R@1 | Alignment R@10 |
392
  |-------|-----------|----------|------------------|---------------|----------------|
393
- | **mono_32d** | 32 | 0.4850 🏆 | 0.4355 | N/A | N/A |
394
- | **mono_64d** | 64 | 0.2076 | 0.3984 | N/A | N/A |
395
- | **mono_128d** | 128 | 0.0353 | 0.4111 | N/A | N/A |
 
 
 
396
 
397
  ### Key Findings
398
 
399
- - **Best Isotropy:** mono_32d with 0.4850 (more uniform distribution)
400
- - **Semantic Density:** Average pairwise similarity of 0.4150. Lower values indicate better semantic separation.
401
- - **Alignment Quality:** No aligned models evaluated in this run.
402
  - **Recommendation:** 128d aligned for best cross-lingual performance
403
 
404
  ---
405
  ## 6. Morphological Analysis (Experimental)
406
 
407
- > ⚠️ **Warning:** This language shows low morphological productivity. The statistical signals used for this analysis may be noisy or less reliable than for morphologically rich languages.
408
-
409
  This section presents an automated morphological analysis derived from the statistical divergence between word-level and subword-level models. By analyzing where subword predictability spikes and where word-level coverage fails, we can infer linguistic structures without supervised data.
410
 
411
  ### 6.1 Productivity & Complexity
412
 
413
  | Metric | Value | Interpretation | Recommendation |
414
  |--------|-------|----------------|----------------|
415
- | Productivity Index | **0.000** | Low morphological productivity | ⚠️ Likely unreliable |
416
- | Idiomaticity Gap | **-1.000** | Low formulaic content | - |
417
 
418
  ### 6.2 Affix Inventory (Productive Units)
419
 
@@ -422,21 +457,20 @@ These are the most productive prefixes and suffixes identified by sampling the v
422
  #### Productive Prefixes
423
  | Prefix | Examples |
424
  |--------|----------|
425
- | `-къ` | къатыгъ, къуаншэмрэ, къунетрэр |
426
- | `-зэ` | зэрэзэхагъэуцорэр, зэреджэхэрэр, зэкӏэм |
427
- | `-къы` | къыхэуутыным, къычӏэкӏы, къырафыгъэ |
428
 
429
  #### Productive Suffixes
430
  | Suffix | Examples |
431
  |--------|----------|
432
- | `-э` | зыхъожьыгъэ, инароднэ, шапсыгъэбзэкӏэ |
433
- | `-р` | зэрэзэхагъэуцорэр, зэреджэхэрэр, къунетрэр |
434
- | `-м` | къутамэхэм, къушъхьэм, ӏушъом |
435
- | `-эр` | зэрэзэхагъэуцорэр, зэреджэхэрэр, къунетрэр |
436
- | `-эм` | къутамэхэм, къушъхьэм, зэкӏэм |
437
- | `-эу` | псынкіэу, тамыгъэу, доцентэу |
438
- | `-хэр` | урымхэр, кӏэмгуехэр, зэлъаштагъэхэр |
439
- | `-рэ` | къуаншэмрэ, швециемрэ, тыркурэ |
440
 
441
  ### 6.3 Bound Stems (Lexical Roots)
442
 
@@ -444,18 +478,18 @@ Bound stems are high-frequency subword units that are semantically cohesive but
444
 
445
  | Stem | Cohesion | Substitutability | Examples |
446
  |------|----------|------------------|----------|
447
- | `тыгъ` | 1.78x | 28 contexts | тыгъу, итыгъ, тыгъэ |
448
- | `эпкъ` | 1.81x | 25 contexts | нэпкъ, нэпкъы, инэпкъ |
449
- | `ъагъ` | 2.15x | 14 contexts | пчъагъ, лъагъо, пчъагъэ |
450
- | `агъэ` | 1.54x | 41 contexts | тхагъэ, благъэ, багъэх |
451
- | `къуа` | 2.14x | 10 contexts | къуае, къуадж, къуажэ |
452
- | `дыгэ` | 1.88x | 14 contexts | адыгэ, адыгэш, адыгэм |
453
- | `ъхьэ` | 1.77x | 16 contexts | шъхьэ, пшъхьэ, шъхьэм |
454
- | `эхэр` | 1.59x | 21 contexts | бэхэр, дзэхэр, бзэхэр |
455
- | `шъхь` | 1.49x | 24 contexts | шъхьэ, пшъхьэ, ышъхьа |
456
- | `ыгъо` | 1.59x | 19 contexts | мыгъо, цыгъо, пщыгъо |
457
- | `псэу` | 1.57x | 19 contexts | нэпсэу, сыпсэу, щэпсэу |
458
- | `эхэм` | 1.55x | 17 contexts | бзэхэм, блэхэм, цӏэхэм |
459
 
460
  ### 6.4 Affix Compatibility (Co-occurrence)
461
 
@@ -463,16 +497,16 @@ This table shows which prefixes and suffixes most frequently co-occur on the sam
463
 
464
  | Prefix | Suffix | Frequency | Examples |
465
  |--------|--------|-----------|----------|
466
- | `-къ` | `-э` | 96 words | къэттыгъэ, къэлэшхомэ |
467
- | `-къ` | `-р` | 64 words | къэплъаныр, къэплъанхэр |
468
- | `-къ` | `-м` | 56 words | къызэхуэсахэм, къытыралъхьагъэм |
469
- | `-къ` | `-эр` | 52 words | къэплъанхэр, къэралыгъохэр |
470
- | `-зэ` | `-р` | 42 words | зэраукӏырэр, зэпэуцужьыныгъэхэр |
471
- | `-зэ` | `-м` | 41 words | зэрэратырэм, зэратебанэщтыгъэхэм |
472
- | `-къ` | `-эм` | 36 words | къызэхуэсахэм, къытыралъхьагъэм |
473
- | `-къ` | `-эу` | 34 words | къехыжьэу, къыдамылъытагъэу |
474
- | `-зэ` | `-эр` | 34 words | зэраукӏырэр, зэпэуцужьыныгъэхэр |
475
- | `-зэ` | `-э` | 31 words | зэхэзгъэуцуагъэ, зэхищэгъэгъэ |
476
 
477
  ### 6.5 Recursive Morpheme Segmentation
478
 
@@ -480,26 +514,28 @@ Using **Recursive Hierarchical Substitutability**, we decompose complex words in
480
 
481
  | Word | Suggested Split | Confidence | Stem |
482
  |------|-----------------|------------|------|
483
- | бзылъфыгъэмрэ | **`бзылъфыгъ-эм-рэ`** | 6.0 | `бзылъфыгъ` |
484
- | сарматхэмрэ | **`сармат-хэм-рэ`** | 6.0 | `сармат` |
485
- | меотхэмрэ | **`меот-хэм-рэ`** | 6.0 | `меот` |
486
- | адыгабзэмрэ | **`адыгабз-эм-рэ`** | 6.0 | `адыгабз` |
487
- | макъэхэмрэ | **`макъэ-хэм-рэ`** | 6.0 | `макъэ` |
488
- | республикэмрэ | **`республик-эм-рэ`** | 6.0 | `республик` |
489
- | зэхэлъхэм | **`зэ-хэлъ-хэм`** | 6.0 | `хэлъ` |
490
- | арапыбзэрэ | **`арапыбз��-рэ`** | 4.5 | `арапыбзэ` |
491
- | ягъунэгъухэр | **`ягъунэгъу-хэр`** | 4.5 | `ягъунэгъу` |
492
- | унагъохэр | **`унагъо-хэр`** | 4.5 | `унагъо` |
493
- | жьыбгъэхэр | **`жьыбгъэ-хэр`** | 4.5 | `жьыбгъэ` |
494
- | елъытыгъэу | **`елъытыгъ-эу`** | 4.5 | `елъытыгъ` |
495
- | чӏыпӏэхэм | **`чӏыпӏэ-хэм`** | 4.5 | `чӏыпӏэ` |
496
- | зыщыпсэухэрэр | **`зыщыпс-эу-хэр-эр`** | 4.5 | `зыщыпс` |
497
- | журналхэм | **`журнал-хэм`** | 4.5 | `журнал` |
498
 
499
  ### 6.6 Linguistic Interpretation
500
 
501
  > **Automated Insight:**
502
- The language ADY appears to be more isolating or has a highly fixed vocabulary. Word-level models perform nearly as well as subword models, indicating fewer productive morphological processes.
 
 
503
 
504
  ---
505
  ## 7. Summary & Recommendations
@@ -510,8 +546,8 @@ The language ADY appears to be more isolating or has a highly fixed vocabulary.
510
 
511
  | Component | Recommended | Rationale |
512
  |-----------|-------------|-----------|
513
- | Tokenizer | **32k BPE** | Best compression (4.23x) |
514
- | N-gram | **2-gram** | Lowest perplexity (399) |
515
  | Markov | **Context-4** | Highest predictability (98.7%) |
516
  | Embeddings | **100d** | Balanced semantic capture and isotropy |
517
 
@@ -726,4 +762,4 @@ MIT License - Free for academic and commercial use.
726
  ---
727
  *Generated by Wikilangs Models Pipeline*
728
 
729
- *Report Date: 2026-01-03 12:36:55*
 
1
  ---
2
  language: ady
3
+ language_name: Adyghe
4
  language_family: caucasian_northwest
5
  tags:
6
  - wikilangs
 
10
  - n-gram
11
  - markov
12
  - wikipedia
13
+ - feature-extraction
14
+ - sentence-similarity
15
+ - tokenization
16
+ - n-grams
17
+ - markov-chain
18
+ - text-mining
19
+ - fasttext
20
+ - babelvec
21
+ - vocabulous
22
+ - vocabulary
23
  - monolingual
24
  - family-caucasian_northwest
25
  license: mit
26
  library_name: wikilangs
27
+ pipeline_tag: text-generation
28
  datasets:
29
  - omarkamali/wikipedia-monthly
30
  dataset_info:
 
33
  metrics:
34
  - name: best_compression_ratio
35
  type: compression
36
+ value: 4.197
37
  - name: best_isotropy
38
  type: isotropy
39
+ value: 0.4929
40
  - name: vocabulary_size
41
  type: vocab
42
  value: 0
43
  generated: 2026-01-03
44
  ---
45
 
46
+ # Adyghe - Wikilangs Models
47
  ## Comprehensive Research Report & Full Ablation Study
48
 
49
+ This repository contains NLP models trained and evaluated by Wikilangs, specifically on **Adyghe** Wikipedia data.
50
  We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and word embeddings.
51
 
52
  ## 📋 Repository Contents
 
70
  - [3. Markov Chain Evaluation](#3-markov-chain-evaluation)
71
  - [4. Vocabulary Analysis](#4-vocabulary-analysis)
72
  - [5. Word Embeddings Evaluation](#5-word-embeddings-evaluation)
73
+ - [6. Morphological Analysis (Experimental)](#6--morphological-analysis-experimental)
74
  - [7. Summary & Recommendations](#7-summary--recommendations)
75
  - [Metrics Glossary](#appendix-metrics-glossary--interpretation-guide)
76
  - [Visualizations Index](#visualizations-index)
 
90
 
91
  | Vocab Size | Compression | Avg Token Len | UNK Rate | Total Tokens |
92
  |------------|-------------|---------------|----------|--------------|
93
+ | **8k** | 3.406x | 3.41 | 0.1685% | 137,125 |
94
+ | **16k** | 3.759x | 3.76 | 0.1859% | 124,248 |
95
+ | **32k** | 4.197x 🏆 | 4.20 | 0.2076% | 111,273 |
96
 
97
  ### Tokenization Examples
98
 
99
  Below are sample sentences tokenized with each vocabulary size:
100
 
101
+ **Sample 1:** `(Пынарбашы), Къайсэр къалэм и район. Адыгэхэ нахь бэрэу мы лъэныком щыӏпсэу.`
102
 
103
  | Vocab | Tokens | Count |
104
  |-------|--------|-------|
105
+ | 8k | `▁( пы н арбашы ), ▁къайсэр ▁къалэм ▁и ▁район . ... (+10 more)` | 20 |
106
+ | 16k | `▁( пы н арбашы ), ▁къайсэр ▁къалэм ▁и ▁район . ... (+10 more)` | 20 |
107
+ | 32k | `▁( пынарбашы ), ▁къайсэр ▁къалэм ▁и ▁район . ▁адыгэхэ ▁нахь ... (+5 more)` | 15 |
108
 
109
+ **Sample 2:** `Орэдус орэдхэр зыусырэр. пае классикэ орэд е мэкъамэ ягугъу композитор нахьы...`
110
 
111
  | Vocab | Tokens | Count |
112
  |-------|--------|-------|
113
+ | 8k | `▁орэдус ▁— ▁орэдхэр ▁зы ус ырэр . ▁пае ▁класс икэ ... (+18 more)` | 28 |
114
+ | 16k | `▁орэдус ▁— ▁орэдхэр ▁зы ус ырэр . ▁пае ▁класс икэ ... (+15 more)` | 25 |
115
+ | 32k | `▁орэдус ▁— ▁орэдхэр ▁зыусырэр . ▁пае ▁классикэ ▁орэд ▁е ▁мэкъамэ ... (+10 more)` | 20 |
116
 
117
+ **Sample 3:** `Эбрар Каракурт 17 Щылэмаз Балыкесирым къэхъугъ, Тыркуе Волэйболым и джэгуакӀу,Ты...`
118
 
119
  | Vocab | Tokens | Count |
120
  |-------|--------|-------|
121
+ | 8k | `▁э б рар ▁кара к урт 1 7 ▁щы ... (+21 more)` | 31 |
122
+ | 16k | `▁эбрар ▁карак урт 1 7 ▁щы л эм аз ... (+15 more)` | 25 |
123
+ | 32k | `▁эбрар ▁каракурт 1 7 ▁щылэмаз ▁балыкесирым ▁къэхъугъ , ▁тыркуе ... (+10 more)` | 20 |
124
 
125
 
126
  ### Key Findings
127
 
128
+ - **Best Compression:** 32k achieves 4.197x compression
129
+ - **Lowest UNK Rate:** 8k with 0.1685% unknown tokens
130
  - **Trade-off:** Larger vocabularies improve compression but increase model size
131
  - **Recommendation:** 32k vocabulary provides optimal balance for production use
132
 
 
143
 
144
  | N-gram | Variant | Perplexity | Entropy | Unique N-grams | Top-100 Coverage | Top-1000 Coverage |
145
  |--------|---------|------------|---------|----------------|------------------|-------------------|
146
+ | **2-gram** | Word | 453 | 8.82 | 625 | 42.6% | 100.0% |
147
+ | **2-gram** | Subword | 407 🏆 | 8.67 | 2,126 | 56.6% | 97.3% |
148
+ | **3-gram** | Word | 759 | 9.57 | 977 | 31.6% | 100.0% |
149
+ | **3-gram** | Subword | 2,854 | 11.48 | 11,856 | 24.3% | 64.6% |
150
+ | **4-gram** | Word | 2,909 | 11.51 | 3,378 | 13.2% | 45.0% |
151
+ | **4-gram** | Subword | 10,911 | 13.41 | 36,062 | 12.4% | 39.2% |
152
+ | **5-gram** | Word | 2,658 | 11.38 | 2,950 | 12.2% | 45.6% |
153
+ | **5-gram** | Subword | 21,199 | 14.37 | 52,393 | 8.2% | 28.3% |
154
 
155
  ### Top 5 N-grams by Size
156
 
 
158
 
159
  | Rank | N-gram | Count |
160
  |------|--------|-------|
161
+ | 1 | `нэбгырэ млн` | 168 |
162
  | 2 | `къехъу щэпсэу` | 104 |
163
+ | 3 | къехъу` | 89 |
164
+ | 4 | `дло м` | 87 |
165
+ | 5 | `адыгэ республикэм` | 80 |
166
 
167
  **3-grams (Word):**
168
 
 
170
  |------|--------|-------|
171
  | 1 | `м къехъу щэпсэу` | 76 |
172
  | 2 | `къехъу щэпсэу хэгэгум` | 70 |
173
+ | 3 | `адыгэ республикэм и` | 46 |
174
  | 4 | `дло м хахьэ` | 44 |
175
  | 5 | `м хахьэ хэгъэгу` | 39 |
176
 
 
184
  | 4 | `америкэм ит къэралыгъу къэлэ` | 19 |
185
  | 5 | `азием ит къэралыгъу къэлэ` | 18 |
186
 
187
+ **5-grams (Word):**
188
+
189
+ | Rank | N-gram | Count |
190
+ |------|--------|-------|
191
+ | 1 | `км гъогу щыӏ къуаджэм ис` | 17 |
192
+ | 2 | `гъогу щыӏ къуаджэм ис цӏыфхэр` | 17 |
193
+ | 3 | `щыӏ къуаджэм ис цӏыфхэр илъэсхэм` | 17 |
194
+ | 4 | `къуаджэм ис цӏыфхэр илъэсхэм тетэу` | 17 |
195
+ | 5 | `ис цӏыфхэр илъэсхэм тетэу къуаджэм` | 17 |
196
+
197
  **2-grams (Subword):**
198
 
199
  | Rank | N-gram | Count |
200
  |------|--------|-------|
201
+ | 1 | `г ъ` | 9,326 |
202
+ | 2 | `ъ э` | 9,249 |
203
+ | 3 | `э _` | 8,792 |
204
+ | 4 | `м _` | 7,740 |
205
+ | 5 | `э р` | 6,822 |
206
 
207
  **3-grams (Subword):**
208
 
209
  | Rank | N-gram | Count |
210
  |------|--------|-------|
211
+ | 1 | `г ъ э` | 4,961 |
212
+ | 2 | `_ к ъ` | 4,140 |
213
+ | 3 | `э м _` | 3,581 |
214
+ | 4 | `ы г ъ` | 3,362 |
215
+ | 5 | `э р _` | 3,020 |
216
 
217
  **4-grams (Subword):**
218
 
219
  | Rank | N-gram | Count |
220
  |------|--------|-------|
221
+ | 1 | `ы г ъ э` | 1,902 |
222
+ | 2 | `х э р _` | 1,448 |
223
+ | 3 | `а г �� э` | 1,342 |
224
+ | 4 | `х э м _` | 1,303 |
225
  | 5 | `_ к ъ э` | 1,289 |
226
 
227
+ **5-grams (Subword):**
228
+
229
+ | Rank | N-gram | Count |
230
+ |------|--------|-------|
231
+ | 1 | `_ а д ы г` | 1,062 |
232
+ | 2 | `а д ы г э` | 978 |
233
+ | 3 | `_ и л ъ э` | 670 |
234
+ | 4 | `д ы г э _` | 651 |
235
+ | 5 | `и л ъ э с` | 627 |
236
+
237
 
238
  ### Key Findings
239
 
240
+ - **Best Perplexity:** 2-gram (subword) with 407
241
  - **Entropy Trend:** Decreases with larger n-grams (more predictable)
242
+ - **Coverage:** Top-1000 patterns cover ~28% of corpus
243
  - **Recommendation:** 4-gram or 5-gram for best predictive performance
244
 
245
  ---
 
255
 
256
  | Context | Variant | Avg Entropy | Perplexity | Branching Factor | Unique Contexts | Predictability |
257
  |---------|---------|-------------|------------|------------------|-----------------|----------------|
258
+ | **1** | Word | 0.4341 | 1.351 | 2.09 | 22,655 | 56.6% |
259
+ | **1** | Subword | 1.4193 | 2.674 | 10.02 | 450 | 0.0% |
260
+ | **2** | Word | 0.0766 | 1.055 | 1.12 | 46,851 | 92.3% |
261
+ | **2** | Subword | 1.1376 | 2.200 | 5.57 | 4,503 | 0.0% |
262
+ | **3** | Word | 0.0248 | 1.017 | 1.04 | 51,794 | 97.5% |
263
+ | **3** | Subword | 0.7466 | 1.678 | 2.95 | 25,044 | 25.3% |
264
+ | **4** | Word | 0.0130 🏆 | 1.009 | 1.02 | 53,002 | 98.7% |
265
+ | **4** | Subword | 0.4264 | 1.344 | 1.85 | 73,859 | 57.4% |
266
 
267
  ### Generated Text Samples (Word-based)
268
 
 
270
 
271
  **Context Size 1:**
272
 
273
+ 1. `и нэхъышъхьэ лэжьыгъэм статистикэм теухуауэ интервью къэрал хассан аль джадид зэхащагъ илъэсым тэуфи...`
274
+ 2. `адыгэ литературэм ихьаси лъэшэу фэӏэзагъэх синдикэр къэралыгъоу тунисым и 20 м нэс тхыдэр нэхь мэхъу...`
275
+ 3. `м хахьэ ыужрэр алтай бзэунагъом хахьэ хэгъэгу тхьаматэр инь юн`
276
 
277
  **Context Size 2:**
278
 
279
+ 1. `нэбгырэ млн 7 къехъу щэпсэу хэгэгум 51 100 км арапыбзэ дло м еуро зэкъотыныгъэм ахахьэ хэгъэгу колин...`
280
+ 2. `къехъу щэпсэу хэгэгум 718 км китаибзэ англыбзэ малаибзэ тамилыбзэ дло м хахьэ хэгъэгу пачъыхьэу абду...`
281
+ 3. къехъу щэпсэу хэгэгум 267 667 км францыбзэ къэрал фор эссозимна гнассингбе хэгъэгу тхьаматэр даниэ...`
282
 
283
  **Context Size 3:**
284
 
285
+ 1. `м къехъу щэпсэу хэгэгум 765 км арапыбз арап къэралмэ анахь баймэ а��ыщ нефтыр лъэшдэдэу дло м хахьэ х...`
286
+ 2. `къехъу щэпсэу хэгэгум 147 570 км бенгалыбзэ дло м хахьэ хэгъэгу алмазбек атамбаев къэрал тхьэматэр т...`
287
+ 3. `адыгэ республикэм и шэуджэн къедзыгъом и къоджэ км 42 мыекъуапэ пэчыжь хэкум къинэжьыгъэ абдзэхэ къо...`
288
 
289
  **Context Size 4:**
290
 
291
+ 1. `м къехъу щэпсэу хэгэгум чӏырэу иӏэр 17 820 км бзэшъхьаӏэр арапыбз дло м хахьэ хэгъэгу хассанал болки...`
292
+ 2. `дло м хахьэ хэгъэгу тейн сейн географие азием и гъунэгъухэр урысые казахстан кыргызстан монголие ишъ...`
293
+ 3. `еуропэм хэт къэралыгъу къэлэ загреб нэбгырэ млн 4 м къехъу щэпсэу я 116 хэгэгум 49 035 км я 129`
294
 
295
 
296
  ### Generated Text Samples (Subword-based)
 
299
 
300
  **Context Size 1:**
301
 
302
+ 1. `_пчӏэр_гокъэме_д`
303
+ 2. `эр_цӏыӏэзэзынэ_я`
304
+ 3. `ыем_щщэра,_фадж.`
305
 
306
  **Context Size 2:**
307
 
308
+ 1. `гъэмьяхэр_арт_пре`
309
+ 2. `ъэу_дэхъ_зышӏэным`
310
+ 3. `э_гъэ_ратымэ_лъхь`
311
 
312
  **Context Size 3:**
313
 
314
+ 1. `гъэзекӏожьыдзэнэжы`
315
+ 2. `_къагъэхьыбэмэ,_гу`
316
+ 3. `эм_къурэтхъум__ищ`
317
 
318
  **Context Size 4:**
319
 
320
+ 1. `ыгъэ_хасэмрэ_млн_89`
321
+ 2. `хэр_къолэжъхэр_тхыг`
322
+ 3. `агъэкӏотэщтыр_ары._`
323
 
324
 
325
  ### Key Findings
326
 
327
  - **Best Predictability:** Context-4 (word) with 98.7% predictability
328
  - **Branching Factor:** Decreases with context size (more deterministic)
329
+ - **Memory Trade-off:** Larger contexts require more storage (73,859 contexts)
330
  - **Recommendation:** Context-3 or Context-4 for text generation
331
 
332
  ---
 
342
 
343
  | Metric | Value |
344
  |--------|-------|
345
+ | Vocabulary Size | 7,120 |
346
+ | Total Tokens | 45,308 |
347
+ | Mean Frequency | 6.36 |
348
  | Median Frequency | 3 |
349
+ | Frequency Std Dev | 22.08 |
350
 
351
  ### Most Common Words
352
 
353
  | Rank | Word | Frequency |
354
  |------|------|-----------|
355
+ | 1 | и | 999 |
356
+ | 2 | адыгэ | 660 |
357
+ | 3 | м | 508 |
358
+ | 4 | илъэсым | 406 |
359
  | 5 | ащ | 391 |
360
+ | 6 | я | 320 |
361
+ | 7 | ары | 274 |
362
+ | 8 | а | 257 |
363
+ | 9 | нэбгырэ | 250 |
364
+ | 10 | е | 223 |
365
 
366
  ### Least Common Words (from vocabulary)
367
 
368
  | Rank | Word | Frequency |
369
  |------|------|-----------|
370
+ | 1 | muzea | 2 |
371
+ | 2 | britishpedia | 2 |
372
+ | 3 | encyklopedia | 2 |
373
+ | 4 | osobistości | 2 |
374
+ | 5 | rzeczypospolitej | 2 |
375
+ | 6 | polskiej | 2 |
376
+ | 7 | bph | 2 |
377
+ | 8 | british | 2 |
378
+ | 9 | publishing | 2 |
379
+ | 10 | ltd | 2 |
380
 
381
  ### Zipf's Law Analysis
382
 
383
  | Metric | Value |
384
  |--------|-------|
385
+ | Zipf Coefficient | 0.7863 |
386
+ | R² (Goodness of Fit) | 0.977814 |
387
  | Adherence Quality | **excellent** |
388
 
389
  ### Coverage Analysis
390
 
391
  | Top N Words | Coverage |
392
  |-------------|----------|
393
+ | Top 100 | 28.8% |
394
+ | Top 1,000 | 60.5% |
395
+ | Top 5,000 | 90.6% |
396
  | Top 10,000 | 0.0% |
397
 
398
  ### Key Findings
399
 
400
+ - **Zipf Compliance:** R²=0.9778 indicates excellent adherence to Zipf's law
401
+ - **High Frequency Dominance:** Top 100 words cover 28.8% of corpus
402
+ - **Long Tail:** -2,880 words needed for remaining 100.0% coverage
403
 
404
  ---
405
  ## 5. Word Embeddings Evaluation
 
415
 
416
  ### 5.1 Cross-Lingual Alignment
417
 
418
+ ![Alignment Quality](visualizations/embedding_alignment_quality.png)
419
+
420
+ ![Multilingual t-SNE](visualizations/embedding_tsne_multilingual.png)
421
 
422
 
423
  ### 5.2 Model Comparison
424
 
425
  | Model | Dimension | Isotropy | Semantic Density | Alignment R@1 | Alignment R@10 |
426
  |-------|-----------|----------|------------------|---------------|----------------|
427
+ | **mono_32d** | 32 | 0.4929 🏆 | 0.4238 | N/A | N/A |
428
+ | **mono_64d** | 64 | 0.2008 | 0.4008 | N/A | N/A |
429
+ | **mono_128d** | 128 | 0.0373 | 0.3931 | N/A | N/A |
430
+ | **aligned_32d** | 32 | 0.4929 | 0.4303 | 0.0632 | 0.4080 |
431
+ | **aligned_64d** | 64 | 0.2008 | 0.3933 | 0.2011 | 0.7586 |
432
+ | **aligned_128d** | 128 | 0.0373 | 0.3923 | 0.2701 | 0.8046 |
433
 
434
  ### Key Findings
435
 
436
+ - **Best Isotropy:** mono_32d with 0.4929 (more uniform distribution)
437
+ - **Semantic Density:** Average pairwise similarity of 0.4056. Lower values indicate better semantic separation.
438
+ - **Alignment Quality:** Aligned models achieve up to 27.0% R@1 in cross-lingual retrieval.
439
  - **Recommendation:** 128d aligned for best cross-lingual performance
440
 
441
  ---
442
  ## 6. Morphological Analysis (Experimental)
443
 
 
 
444
  This section presents an automated morphological analysis derived from the statistical divergence between word-level and subword-level models. By analyzing where subword predictability spikes and where word-level coverage fails, we can infer linguistic structures without supervised data.
445
 
446
  ### 6.1 Productivity & Complexity
447
 
448
  | Metric | Value | Interpretation | Recommendation |
449
  |--------|-------|----------------|----------------|
450
+ | Productivity Index | **5.000** | High morphological productivity | Reliable analysis |
451
+ | Idiomaticity Gap | **0.610** | High formulaic/idiomatic content | - |
452
 
453
  ### 6.2 Affix Inventory (Productive Units)
454
 
 
457
  #### Productive Prefixes
458
  | Prefix | Examples |
459
  |--------|----------|
460
+ | `-къ` | къахэщых, къэштэжь, къыхагъэщэу |
461
+ | `-зэ` | зэмыпэсырэм, зэрагъэзэкӏуагъэу, зэфэшъхьафыбэмэ |
 
462
 
463
  #### Productive Suffixes
464
  | Suffix | Examples |
465
  |--------|----------|
466
+ | `-э` | инджылыбзэ, шъхьэгуащэ, ыкурэ |
467
+ | `-р` | хъулъфыгъэхэр, егъэблэгъэныр, усэхэр |
468
+ | `-м` | зэмыпэсырэм, бысымым, м |
469
+ | `-эр` | хъулъфыгъэхэр, усэхэр, благъэр |
470
+ | `-эм` | зэмыпэсырэм, къутамэм, пхъэм |
471
+ | `-эу` | бэрэу, зэрагъэзэкӏуагъэу, къыхагъэщэу |
472
+ | `-хэр` | хъулъфыгъэхэр, усэхэр, ӏутыхэр |
473
+ | `-рэ` | ыкурэ, цӏэмрэ, чэщрэ |
474
 
475
  ### 6.3 Bound Stems (Lexical Roots)
476
 
 
478
 
479
  | Stem | Cohesion | Substitutability | Examples |
480
  |------|----------|------------------|----------|
481
+ | `тыгъ` | 1.78x | 28 contexts | тыгъу, тыгъэ, итыгъ |
482
+ | `ъагъ` | 2.17x | 14 contexts | пчъагъ, лъагъо, тхъагъо |
483
+ | `эпкъ` | 1.76x | 25 contexts | нэпкъ, нэпкъы, инэпкъ |
484
+ | `агъэ` | 1.55x | 39 contexts | тхагъэ, багъэх, благъэ |
485
+ | `къуа` | 2.17x | 10 contexts | къуае, къуадж, къуажэ |
486
+ | `дыгэ` | 1.90x | 14 contexts | адыгэ, адыгэу, адыгэм |
487
+ | `псэу` | 1.64x | 20 contexts | упсэу, нэпсэу, щэпсэу |
488
+ | `эхэр` | 1.61x | 20 contexts | бэхэр, усэхэр, унэхэр |
489
+ | `ъхьэ` | 1.72x | 16 contexts | шъхьэ, ишъхьэ, шъхьэм |
490
+ | `ыгъо` | 1.62x | 19 contexts | цыгъо, мыгъо, мыгъом |
491
+ | `шъхь` | 1.51x | 23 contexts | шъхьэ, шъхьаф, ишъхьэ |
492
+ | `гъэх` | 1.67x | 14 contexts | багъэх, тхыгъэх, ежагъэх |
493
 
494
  ### 6.4 Affix Compatibility (Co-occurrence)
495
 
 
497
 
498
  | Prefix | Suffix | Frequency | Examples |
499
  |--------|--------|-----------|----------|
500
+ | `-къ` | `-э` | 94 words | къызэриӏорэмкӏэ, къыгъэпсыщтыгъэ |
501
+ | `-къ` | `-р` | 64 words | къалъхуахэр, къызэдыхэфэныр |
502
+ | `-къ` | `-м` | 56 words | къэралхэм, къожъхэм |
503
+ | `-къ` | `-эр` | 52 words | къалъхуахэр, къалэр |
504
+ | `-зэ` | `-р` | 43 words | зэрэзэтекӏыхэрэр, зэрыхъур |
505
+ | `-зэ` | `-м` | 41 words | зэрэхъурэм, зэкъотыныгъэм |
506
+ | `-къ` | `-эм` | 36 words | къэралхэм, къожъхэм |
507
+ | `-зэ` | `-эр` | 34 words | зэрэзэтекӏыхэрэр, зэриукъорэр |
508
+ | `-къ` | `-эу` | 33 words | къыхахыгъэу, къыдыхэлъытагъэу |
509
+ | `-зэ` | `-э` | 31 words | зэралэжьырэ, зэгъусэмэ |
510
 
511
  ### 6.5 Recursive Morpheme Segmentation
512
 
 
514
 
515
  | Word | Suggested Split | Confidence | Stem |
516
  |------|-----------------|------------|------|
517
+ | щыпсэухэрэр | **`щыпс-эу-хэр-эр`** | 7.5 | `щыпс` |
518
+ | литературэмрэ | **`литератур-эм-рэ`** | 6.0 | `литератур` |
519
+ | мыхъунхэр | **`мыхъун-хэр`** | 4.5 | `мыхъун` |
520
+ | джуртыбзэрэ | **`джуртыбзэ-рэ`** | 4.5 | `джуртыбзэ` |
521
+ | тхьаматэр | **`тхьамат-эр`** | 4.5 | `тхьамат` |
522
+ | фэхъугъэм | **`фэхъугъ-эм`** | 4.5 | `фэхъугъ` |
523
+ | игъунэгъухэр | **`игъунэгъу-хэр`** | 4.5 | `игъунэгъу` |
524
+ | нэмыкӏхэр | **`нэмыкӏ-хэр`** | 4.5 | `нэмыкӏ` |
525
+ | зэкъотыныгъэм | **`зэ-къ-отыныгъ-эм`** | 4.5 | `отыныгъ` |
526
+ | ипрезидентэу | **`ипрезидент-эу`** | 4.5 | `ипрезидент` |
527
+ | литературэр | **`литератур-эр`** | 4.5 | `литератур` |
528
+ | хъыбархэм | **`хъыбар-хэм`** | 4.5 | `хъыбар` |
529
+ | культурэм | **`культур-эм`** | 4.5 | `культур` |
530
+ | къыхафыгъэхэр | **`къ-ыхафыг-ъэ-хэр`** | 4.5 | `ыхафыг` |
531
+ | зэрэгущаӏэхэрэр | **`зэ-рэгущаӏэ-хэр-эр`** | 4.5 | `рэгущаӏэ` |
532
 
533
  ### 6.6 Linguistic Interpretation
534
 
535
  > **Automated Insight:**
536
+ The language Adyghe shows high morphological productivity. The subword models are significantly more efficient than word models, suggesting a rich system of affixation or compounding.
537
+
538
+ > **Note on Idiomaticity:** The high Idiomaticity Gap suggests a large number of frequent multi-word expressions or formulaic sequences that are statistically distinct from their component parts.
539
 
540
  ---
541
  ## 7. Summary & Recommendations
 
546
 
547
  | Component | Recommended | Rationale |
548
  |-----------|-------------|-----------|
549
+ | Tokenizer | **32k BPE** | Best compression (4.20x) |
550
+ | N-gram | **2-gram** | Lowest perplexity (407) |
551
  | Markov | **Context-4** | Highest predictability (98.7%) |
552
  | Embeddings | **100d** | Balanced semantic capture and isotropy |
553
 
 
762
  ---
763
  *Generated by Wikilangs Models Pipeline*
764
 
765
+ *Report Date: 2026-01-03 14:02:39*
models/embeddings/aligned/ady_128d.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f176dc3785048428a3b30c841bd84ed79b92ff591378272a599482b167641cd
3
+ size 1025644289
models/embeddings/aligned/ady_128d.meta.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lang": "ady", "dim": 128, "max_seq_len": 512, "is_aligned": true}
models/embeddings/aligned/ady_128d.projection.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d83549b0e20236b2c0988f4f13b7266f6f78751ae5dde671daea405bba933aae
3
+ size 65664
models/embeddings/aligned/ady_128d_metadata.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "language": "ady",
3
+ "dimension": 128,
4
+ "version": "aligned",
5
+ "hub_language": "en",
6
+ "seed_vocab_size": 174,
7
+ "vocab_size": 1573
8
+ }
models/embeddings/aligned/ady_32d.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:663d1bbe256cb5339ab746ba37d0f5e5ab034bafd747c390ebbdd1eabbe1b5fc
3
+ size 256436225
models/embeddings/aligned/ady_32d.meta.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lang": "ady", "dim": 32, "max_seq_len": 512, "is_aligned": true}
models/embeddings/aligned/ady_32d.projection.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7057b6fbc0c824ffad21041b164ea17f161de3d678d0a454ec12b77d682c863
3
+ size 4224
models/embeddings/aligned/ady_32d_metadata.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "language": "ady",
3
+ "dimension": 32,
4
+ "version": "aligned",
5
+ "hub_language": "en",
6
+ "seed_vocab_size": 174,
7
+ "vocab_size": 1573
8
+ }
models/embeddings/aligned/ady_64d.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a10090792b8196b70efae444f73787c1ec5103fd55dc37ceb14d647443ad1a4
3
+ size 512838913
models/embeddings/aligned/ady_64d.meta.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lang": "ady", "dim": 64, "max_seq_len": 512, "is_aligned": true}
models/embeddings/aligned/ady_64d.projection.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84ba46ef4bf4f82df308790cb1ba9f63722408b474837aeab9f8fc47f81c7e11
3
+ size 16512
models/embeddings/aligned/ady_64d_metadata.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "language": "ady",
3
+ "dimension": 64,
4
+ "version": "aligned",
5
+ "hub_language": "en",
6
+ "seed_vocab_size": 174,
7
+ "vocab_size": 1573
8
+ }
models/embeddings/monolingual/ady_128d.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:15314cc76788ddb26c68cad7ccc4d1a2fa1c42bf0d8a23163020f6c1be5e011c
3
- size 1025621365
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f176dc3785048428a3b30c841bd84ed79b92ff591378272a599482b167641cd
3
+ size 1025644289
models/embeddings/monolingual/ady_128d_metadata.json CHANGED
@@ -11,5 +11,5 @@
11
  "encoding_method": "rope",
12
  "dim": 128
13
  },
14
- "vocab_size": 1551
15
  }
 
11
  "encoding_method": "rope",
12
  "dim": 128
13
  },
14
+ "vocab_size": 1573
15
  }
models/embeddings/monolingual/ady_32d.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0a7d65d778045d5838991706962f1f2fda476d83f0eaf0d25e006a2c54b0addb
3
- size 256430197
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:663d1bbe256cb5339ab746ba37d0f5e5ab034bafd747c390ebbdd1eabbe1b5fc
3
+ size 256436225
models/embeddings/monolingual/ady_32d_metadata.json CHANGED
@@ -11,5 +11,5 @@
11
  "encoding_method": "rope",
12
  "dim": 32
13
  },
14
- "vocab_size": 1551
15
  }
 
11
  "encoding_method": "rope",
12
  "dim": 32
13
  },
14
+ "vocab_size": 1573
15
  }
models/embeddings/monolingual/ady_64d.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:462cda6a4c5bb17912b29e9f210bd916044788b457f8b2cd61463a18150385c6
3
- size 512827253
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a10090792b8196b70efae444f73787c1ec5103fd55dc37ceb14d647443ad1a4
3
+ size 512838913
models/embeddings/monolingual/ady_64d_metadata.json CHANGED
@@ -11,5 +11,5 @@
11
  "encoding_method": "rope",
12
  "dim": 64
13
  },
14
- "vocab_size": 1551
15
  }
 
11
  "encoding_method": "rope",
12
  "dim": 64
13
  },
14
+ "vocab_size": 1573
15
  }
models/subword_markov/ady_markov_ctx1_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:326d1899bc21c88f5d2ecbb60df90091a3f81877b05f9f3880179e64fc82bc39
3
- size 35559
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea1f21df4b685dfea6b5db6e44e57a8bb0ac92311c91c308143cf35fcbb61d68
3
+ size 38398
models/subword_markov/ady_markov_ctx1_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 1,
3
  "variant": "subword",
4
  "language": "ady",
5
- "unique_contexts": 410,
6
- "total_transitions": 461186
7
  }
 
2
  "context_size": 1,
3
  "variant": "subword",
4
  "language": "ady",
5
+ "unique_contexts": 450,
6
+ "total_transitions": 466024
7
  }
models/subword_markov/ady_markov_ctx2_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:439580e1d85b149b7a4e12434dcacbd8c63ce4f422949f7787d12f51a3b06768
3
- size 169546
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc4a78482132260c4af1de9c10f9cc91758d344526e771dd9ccb4fc1f194464c
3
+ size 177308
models/subword_markov/ady_markov_ctx2_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 2,
3
  "variant": "subword",
4
  "language": "ady",
5
- "unique_contexts": 4325,
6
- "total_transitions": 460515
7
  }
 
2
  "context_size": 2,
3
  "variant": "subword",
4
  "language": "ady",
5
+ "unique_contexts": 4503,
6
+ "total_transitions": 465338
7
  }
models/subword_markov/ady_markov_ctx3_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ba32cd485f139de6037badf290db15170159de5ebb63242fabe6473e50963984
3
- size 554556
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b821b2be71b4b958f03177625985421729fee5ea5f35f0bcbbba7f891cf8e38
3
+ size 567095
models/subword_markov/ady_markov_ctx3_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 3,
3
  "variant": "subword",
4
  "language": "ady",
5
- "unique_contexts": 24260,
6
- "total_transitions": 459844
7
  }
 
2
  "context_size": 3,
3
  "variant": "subword",
4
  "language": "ady",
5
+ "unique_contexts": 25044,
6
+ "total_transitions": 464652
7
  }
models/subword_markov/ady_markov_ctx4_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a23a4022efd4fb6c5360771eb3945cadabce7d4e08d9a04895e4535a30c12dad
3
- size 1240010
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5acbad74a99249e96dd951ee18917f894f3669704783fd95161beee040f3fce
3
+ size 1265262
models/subword_markov/ady_markov_ctx4_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 4,
3
  "variant": "subword",
4
  "language": "ady",
5
- "unique_contexts": 72077,
6
- "total_transitions": 459173
7
  }
 
2
  "context_size": 4,
3
  "variant": "subword",
4
  "language": "ady",
5
+ "unique_contexts": 73859,
6
+ "total_transitions": 463966
7
  }
models/subword_ngram/ady_2gram_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f2744392ac6f30902dc24ef71be21ed3d0027816ff58c51cc117dbe79b60b6eb
3
- size 26615
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9de0521f3604c270d9d5334436fe81fb540e8ea5906b0be25a092aa6179cf45
3
+ size 27108
models/subword_ngram/ady_2gram_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "n": 2,
3
  "variant": "subword",
4
  "language": "ady",
5
- "unique_ngrams": 2072,
6
- "total_ngrams": 461186
7
  }
 
2
  "n": 2,
3
  "variant": "subword",
4
  "language": "ady",
5
+ "unique_ngrams": 2126,
6
+ "total_ngrams": 466024
7
  }
models/subword_ngram/ady_3gram_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3e27920b4a896f7b33778a4da60561fc4bf2c40fe17414d4721a023a815cc6e6
3
- size 144946
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed0ae00b308ead9c908cdaabf26a84bec9172251d443c0d3d3aa02081e24292d
3
+ size 147714
models/subword_ngram/ady_3gram_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "n": 3,
3
  "variant": "subword",
4
  "language": "ady",
5
- "unique_ngrams": 11614,
6
- "total_ngrams": 460515
7
  }
 
2
  "n": 3,
3
  "variant": "subword",
4
  "language": "ady",
5
+ "unique_ngrams": 11856,
6
+ "total_ngrams": 465338
7
  }
models/subword_ngram/ady_4gram_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a1189db5bc7c400c2d162eabf55e0ca7ce6b8403c020f3330fe200b437eab55c
3
- size 467994
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8d87c1c9407f9d5599e4e07229e5b7ccf78b280fc06922a9134d9e6f48368fa
3
+ size 470826
models/subword_ngram/ady_4gram_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "n": 4,
3
  "variant": "subword",
4
  "language": "ady",
5
- "unique_ngrams": 35316,
6
- "total_ngrams": 459844
7
  }
 
2
  "n": 4,
3
  "variant": "subword",
4
  "language": "ady",
5
+ "unique_ngrams": 36062,
6
+ "total_ngrams": 464652
7
  }
models/subword_ngram/ady_5gram_subword.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5b99c1f7304e466988fd44a8987f2393709285de0fa76cc71513adbc5c16ebb
3
+ size 704587
models/subword_ngram/ady_5gram_subword_metadata.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "n": 5,
3
+ "variant": "subword",
4
+ "language": "ady",
5
+ "unique_ngrams": 52393,
6
+ "total_ngrams": 463966
7
+ }
models/tokenizer/ady_tokenizer_16k.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:39a0cc5483f8b60e51f162e1451531fbd79f4a7ef64ed4d9770c72634ad2f63f
3
- size 582264
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:907c1ae616234d526f5961f0c4f163fb497ac9c3805d3e14a7fe3fabf96c944f
3
+ size 579551
models/tokenizer/ady_tokenizer_16k.vocab CHANGED
The diff for this file is too large to render. See raw diff
 
models/tokenizer/ady_tokenizer_32k.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:60fe921e9bf5abac9a26d2e450368741ec923239ef1bfe3380b5fecc9f1ebc5e
3
- size 924924
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62873f8357831fb35e5a2f441ad474a6c2eb8cc0ae9734eb5c04190135969486
3
+ size 926359
models/tokenizer/ady_tokenizer_32k.vocab CHANGED
The diff for this file is too large to render. See raw diff
 
models/tokenizer/ady_tokenizer_8k.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eed6e8aa7a197f3cf93f47869aff6cc5bec5173974995af92b59a00c2be84d9a
3
- size 396510
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c0595fa5cbdefef5ffe8230419254db68de7f2fb652181e980c62b94f6eea3a
3
+ size 395183
models/tokenizer/ady_tokenizer_8k.vocab CHANGED
The diff for this file is too large to render. See raw diff
 
models/vocabulary/ady_vocabulary.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0394f3c6c2f9871c8756de87f7c0ab00cba32facf111d0c2027f0844c932241b
3
- size 124070
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e4972182eb8bd7e39e0e37b85fab21a678c05c8dec6fd4c262eca1c1d0e7439b
3
+ size 125482
models/vocabulary/ady_vocabulary_metadata.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "language": "ady",
3
- "vocabulary_size": 7032,
4
  "variant": "full",
5
  "statistics": {
6
- "type_token_ratio": 0.37416908841901325,
7
  "coverage": {
8
- "top_100": 0.21755686942579416,
9
- "top_1000": 0.4501954103617597,
10
- "top_5000": 0.6754016768547283,
11
- "top_10000": 0.7928483147944015
12
  },
13
- "hapax_count": 15371,
14
- "hapax_ratio": 0.6861134669463911,
15
- "total_documents": 671
16
  }
17
  }
 
1
  {
2
  "language": "ady",
3
+ "vocabulary_size": 7120,
4
  "variant": "full",
5
  "statistics": {
6
+ "type_token_ratio": 0.3733302701106042,
7
  "coverage": {
8
+ "top_100": 0.21408644852144804,
9
+ "top_1000": 0.45016246020545475,
10
+ "top_5000": 0.6739308805671338,
11
+ "top_10000": 0.7907709475204306
12
  },
13
+ "hapax_count": 15630,
14
+ "hapax_ratio": 0.687032967032967,
15
+ "total_documents": 686
16
  }
17
  }
models/word_markov/ady_markov_ctx1_word.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e529d92df652217192b1a53b43d340b1b2d134ff78b718c2331e8296f5e8a183
3
- size 772451
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c10b7085544a5b8115070723022c0bdfcead811182d57a2c877f7604d1ba7a53
3
+ size 782266
models/word_markov/ady_markov_ctx1_word_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 1,
3
  "variant": "word",
4
  "language": "ady",
5
- "unique_contexts": 22306,
6
- "total_transitions": 59203
7
  }
 
2
  "context_size": 1,
3
  "variant": "word",
4
  "language": "ady",
5
+ "unique_contexts": 22655,
6
+ "total_transitions": 60252
7
  }
models/word_markov/ady_markov_ctx2_word.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:949fea0fcff222936353244ef6edc60951c50ac50290dc3b857a31ce73658051
3
- size 1310186
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0bd26c61a14f1e19b592dca73900dc19aa0d3e8839ea60b95c3643dba42ae82
3
+ size 1323569
models/word_markov/ady_markov_ctx2_word_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 2,
3
  "variant": "word",
4
  "language": "ady",
5
- "unique_contexts": 46305,
6
- "total_transitions": 58532
7
  }
 
2
  "context_size": 2,
3
  "variant": "word",
4
  "language": "ady",
5
+ "unique_contexts": 46851,
6
+ "total_transitions": 59566
7
  }
models/word_markov/ady_markov_ctx3_word.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0edd29aaf52ac72ad56819e317223cc422f1de2fd93feb950a6fd0c9d7772f20
3
- size 1585798
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ede7cceaaeebf470cbbc85a329ad43223fcfd6126b983a75c435ce8f14238ff
3
+ size 1598892
models/word_markov/ady_markov_ctx3_word_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 3,
3
  "variant": "word",
4
  "language": "ady",
5
- "unique_contexts": 51243,
6
- "total_transitions": 57861
7
  }
 
2
  "context_size": 3,
3
  "variant": "word",
4
  "language": "ady",
5
+ "unique_contexts": 51794,
6
+ "total_transitions": 58880
7
  }