omarkamali commited on
Commit
a6e972b
·
verified ·
1 Parent(s): 16bbcc3

Upload all models and assets for ab (latest)

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +1 -0
  2. README.md +201 -165
  3. models/embeddings/aligned/ab_128d.bin +3 -0
  4. models/embeddings/aligned/ab_128d.meta.json +1 -0
  5. models/embeddings/aligned/ab_128d.projection.npy +3 -0
  6. models/embeddings/aligned/ab_128d_metadata.json +8 -0
  7. models/embeddings/aligned/ab_32d.bin +3 -0
  8. models/embeddings/aligned/ab_32d.meta.json +1 -0
  9. models/embeddings/aligned/ab_32d.projection.npy +3 -0
  10. models/embeddings/aligned/ab_32d_metadata.json +8 -0
  11. models/embeddings/aligned/ab_64d.bin +3 -0
  12. models/embeddings/aligned/ab_64d.meta.json +1 -0
  13. models/embeddings/aligned/ab_64d.projection.npy +3 -0
  14. models/embeddings/aligned/ab_64d_metadata.json +8 -0
  15. models/embeddings/monolingual/ab_128d.bin +2 -2
  16. models/embeddings/monolingual/ab_128d_metadata.json +1 -1
  17. models/embeddings/monolingual/ab_32d.bin +2 -2
  18. models/embeddings/monolingual/ab_32d_metadata.json +1 -1
  19. models/embeddings/monolingual/ab_64d.bin +2 -2
  20. models/embeddings/monolingual/ab_64d_metadata.json +1 -1
  21. models/subword_markov/ab_markov_ctx1_subword.parquet +2 -2
  22. models/subword_markov/ab_markov_ctx1_subword_metadata.json +2 -2
  23. models/subword_markov/ab_markov_ctx2_subword.parquet +2 -2
  24. models/subword_markov/ab_markov_ctx2_subword_metadata.json +2 -2
  25. models/subword_markov/ab_markov_ctx3_subword.parquet +2 -2
  26. models/subword_markov/ab_markov_ctx3_subword_metadata.json +2 -2
  27. models/subword_markov/ab_markov_ctx4_subword.parquet +2 -2
  28. models/subword_markov/ab_markov_ctx4_subword_metadata.json +2 -2
  29. models/subword_ngram/ab_2gram_subword.parquet +2 -2
  30. models/subword_ngram/ab_2gram_subword_metadata.json +2 -2
  31. models/subword_ngram/ab_3gram_subword.parquet +2 -2
  32. models/subword_ngram/ab_3gram_subword_metadata.json +2 -2
  33. models/subword_ngram/ab_4gram_subword.parquet +2 -2
  34. models/subword_ngram/ab_4gram_subword_metadata.json +2 -2
  35. models/subword_ngram/ab_5gram_subword.parquet +3 -0
  36. models/subword_ngram/ab_5gram_subword_metadata.json +7 -0
  37. models/tokenizer/ab_tokenizer_16k.model +2 -2
  38. models/tokenizer/ab_tokenizer_16k.vocab +0 -0
  39. models/tokenizer/ab_tokenizer_32k.model +2 -2
  40. models/tokenizer/ab_tokenizer_32k.vocab +0 -0
  41. models/tokenizer/ab_tokenizer_64k.model +2 -2
  42. models/tokenizer/ab_tokenizer_64k.vocab +0 -0
  43. models/tokenizer/ab_tokenizer_8k.model +2 -2
  44. models/tokenizer/ab_tokenizer_8k.vocab +0 -0
  45. models/vocabulary/ab_vocabulary.parquet +2 -2
  46. models/vocabulary/ab_vocabulary_metadata.json +9 -9
  47. models/word_markov/ab_markov_ctx1_word.parquet +2 -2
  48. models/word_markov/ab_markov_ctx1_word_metadata.json +2 -2
  49. models/word_markov/ab_markov_ctx2_word.parquet +2 -2
  50. models/word_markov/ab_markov_ctx2_word_metadata.json +2 -2
.gitattributes CHANGED
@@ -39,3 +39,4 @@ visualizations/position_encoding_comparison.png filter=lfs diff=lfs merge=lfs -t
39
  visualizations/tsne_sentences.png filter=lfs diff=lfs merge=lfs -text
40
  visualizations/tsne_words.png filter=lfs diff=lfs merge=lfs -text
41
  visualizations/zipf_law.png filter=lfs diff=lfs merge=lfs -text
 
 
39
  visualizations/tsne_sentences.png filter=lfs diff=lfs merge=lfs -text
40
  visualizations/tsne_words.png filter=lfs diff=lfs merge=lfs -text
41
  visualizations/zipf_law.png filter=lfs diff=lfs merge=lfs -text
42
+ visualizations/embedding_tsne_multilingual.png filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  language: ab
3
- language_name: AB
4
  language_family: caucasian_northwest
5
  tags:
6
  - wikilangs
@@ -10,11 +10,21 @@ tags:
10
  - n-gram
11
  - markov
12
  - wikipedia
 
 
 
 
 
 
 
 
 
 
13
  - monolingual
14
  - family-caucasian_northwest
15
  license: mit
16
  library_name: wikilangs
17
- pipeline_tag: feature-extraction
18
  datasets:
19
  - omarkamali/wikipedia-monthly
20
  dataset_info:
@@ -26,17 +36,17 @@ metrics:
26
  value: 4.193
27
  - name: best_isotropy
28
  type: isotropy
29
- value: 0.8185
30
  - name: vocabulary_size
31
  type: vocab
32
  value: 0
33
  generated: 2026-01-03
34
  ---
35
 
36
- # AB - Wikilangs Models
37
  ## Comprehensive Research Report & Full Ablation Study
38
 
39
- This repository contains NLP models trained and evaluated by Wikilangs, specifically on **AB** Wikipedia data.
40
  We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and word embeddings.
41
 
42
  ## 📋 Repository Contents
@@ -60,7 +70,7 @@ We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and
60
  - [3. Markov Chain Evaluation](#3-markov-chain-evaluation)
61
  - [4. Vocabulary Analysis](#4-vocabulary-analysis)
62
  - [5. Word Embeddings Evaluation](#5-word-embeddings-evaluation)
63
- - [6. Morphological Analysis (Experimental)](#6-morphological-analysis)
64
  - [7. Summary & Recommendations](#7-summary--recommendations)
65
  - [Metrics Glossary](#appendix-metrics-glossary--interpretation-guide)
66
  - [Visualizations Index](#visualizations-index)
@@ -80,47 +90,47 @@ We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and
80
 
81
  | Vocab Size | Compression | Avg Token Len | UNK Rate | Total Tokens |
82
  |------------|-------------|---------------|----------|--------------|
83
- | **8k** | 3.304x | 3.31 | 0.1494% | 223,505 |
84
- | **16k** | 3.652x | 3.66 | 0.1652% | 202,175 |
85
- | **32k** | 3.908x | 3.91 | 0.1768% | 188,952 |
86
- | **64k** | 4.193x 🏆 | 4.20 | 0.1897% | 176,103 |
87
 
88
  ### Tokenization Examples
89
 
90
  Below are sample sentences tokenized with each vocabulary size:
91
 
92
- **Sample 1:** `Ψ, ψбырзентәи аҩыратә нбан. Азхьарԥшқәа Graphemica (Ψ) Graphemica (ψ) аҩыратә...`
93
 
94
  | Vocab | Tokens | Count |
95
  |-------|--------|-------|
96
- | 8k | `▁ ψ , ▁ ψ ▁— ▁бырзентәи ▁аҩыратә ▁нбан . ... (+11 more)` | 21 |
97
- | 16k | `▁ψ , ▁ψ ▁— ▁бырзентәи ▁аҩыратә ▁нбан . ▁азхьарԥшқәа ▁graphemica ... (+9 more)` | 19 |
98
- | 32k | `▁ψ , ▁ψ ▁— ▁бырзентәи ▁аҩыратә ▁нбан . ▁азхьарԥшқәа ▁graphemica ... (+9 more)` | 19 |
99
- | 64k | `▁ψ , ▁ψ ▁— ▁бырзентәи ▁аҩыратә ▁нбан . ▁азхьарԥшқәа ▁graphemica ... (+9 more)` | 19 |
100
 
101
- **Sample 2:** `Амолдав бызшәа, амолдаван бызшәа (limba moldovenească, лимба молдовеняскэ) Азгәа...`
102
 
103
  | Vocab | Tokens | Count |
104
  |-------|--------|-------|
105
- | 8k | `▁ам ол да в ▁бызшәа , ▁ам ол да ван ... (+24 more)` | 34 |
106
- | 16k | `▁амол дав ▁бызшәа , ▁амол да ван ▁бызшәа( l ... (+20 more)` | 30 |
107
- | 32k | `▁амолдав ▁бызшәа , ▁амол да ван ▁бызшәа( l imba ... (+13 more)` | 23 |
108
- | 64k | `▁амолдав ▁бызшәа , ▁амолдаван ▁бызшәа( limbamoldoveneasc ă , ... (+5 more)` | 15 |
109
 
110
- **Sample 3:** `Ардешен () Ҭырқтәыла ақалақь. Иаланхо . Ахьарԥшқәа ақалақьқәа`
111
 
112
  | Vocab | Tokens | Count |
113
  |-------|--------|-------|
114
- | 8k | `▁ар деш ен ▁() ▁– ▁ҭырқтәыла ▁ақалақь . ▁иаланхо ▁. ... (+2 more)` | 12 |
115
- | 16k | `▁ар деш ен ▁() ▁– ▁ҭырқтәыла ▁ақалақь . ▁иаланхо ▁. ... (+2 more)` | 12 |
116
- | 32k | `▁ардешен ▁() ▁– ▁ҭырқтәыла ▁ақалақь . ▁иаланхо ▁. ▁ахьарԥшқәа ▁ақалақьқәа` | 10 |
117
- | 64k | `▁ардешен ▁() ▁– ▁ҭырқтәыла ▁ақалақь . ▁иаланхо ▁. ▁ахьарԥшқәа ▁ақалақьқәа` | 10 |
118
 
119
 
120
  ### Key Findings
121
 
122
  - **Best Compression:** 64k achieves 4.193x compression
123
- - **Lowest UNK Rate:** 8k with 0.1494% unknown tokens
124
  - **Trade-off:** Larger vocabularies improve compression but increase model size
125
  - **Recommendation:** 32k vocabulary provides optimal balance for production use
126
 
@@ -137,12 +147,14 @@ Below are sample sentences tokenized with each vocabulary size:
137
 
138
  | N-gram | Variant | Perplexity | Entropy | Unique N-grams | Top-100 Coverage | Top-1000 Coverage |
139
  |--------|---------|------------|---------|----------------|------------------|-------------------|
140
- | **2-gram** | Word | 724 | 9.50 | 5,814 | 51.4% | 72.0% |
141
- | **2-gram** | Subword | 363 | 8.50 | 4,104 | 60.3% | 96.8% |
142
- | **3-gram** | Word | 252 🏆 | 7.98 | 5,216 | 66.5% | 80.6% |
143
- | **3-gram** | Subword | 2,675 | 11.39 | 28,199 | 28.1% | 67.5% |
144
- | **4-gram** | Word | 342 | 8.42 | 9,800 | 64.0% | 74.0% |
145
- | **4-gram** | Subword | 11,090 | 13.44 | 112,541 | 16.8% | 44.7% |
 
 
146
 
147
  ### Top 5 N-grams by Size
148
 
@@ -154,7 +166,7 @@ Below are sample sentences tokenized with each vocabulary size:
154
  | 2 | `иит рыԥсҭазаара` | 3,938 |
155
  | 3 | `рашәарамза ԥхынгәымза` | 3,603 |
156
  | 4 | `жәабранмза хәажәкырамза` | 3,603 |
157
- | 5 | `ажьырныҳәамза жәабранмза` | 3,602 |
158
 
159
  **3-grams (Word):**
160
 
@@ -163,55 +175,75 @@ Below are sample sentences tokenized with each vocabulary size:
163
  | 1 | `иит рыԥсҭазаара иалҵит` | 3,938 |
164
  | 2 | `цәыббрамза жьҭаарамза абҵарамза` | 3,602 |
165
  | 3 | `нанҳәамза цәыббрамза жьҭаарамза` | 3,601 |
166
- | 4 | `рашәарамза ԥхынгәымза нанҳәамза` | 3,601 |
167
- | 5 | `мшаԥымза лаҵарамза рашәарамза` | 3,601 |
168
 
169
  **4-grams (Word):**
170
 
171
  | Rank | N-gram | Count |
172
  |------|--------|-------|
173
- | 1 | `мшаԥымза лаҵарамза рашәарамза ԥхынгәымза` | 3,601 |
174
- | 2 | `лаҵарамза рашәарамза ԥхынгәымза нанҳәамза` | 3,601 |
175
- | 3 | `рашәарамза ԥхынгәымза нанҳәамза цәыббрамза` | 3,601 |
176
- | 4 | `ԥхынгәымза нанҳәамза цәыббрамза жьҭаарамза` | 3,601 |
177
- | 5 | `нанҳәамза цәыббрамза жьҭаарамза абҵарамза` | 3,601 |
 
 
 
 
 
 
 
 
 
 
178
 
179
  **2-grams (Subword):**
180
 
181
  | Rank | N-gram | Count |
182
  |------|--------|-------|
183
- | 1 | `а _` | 154,707 |
184
- | 2 | `_ а` | 149,768 |
185
- | 3 | `р а` | 100,539 |
186
- | 4 | `а р` | 84,594 |
187
- | 5 | `ә а` | 75,990 |
188
 
189
  **3-grams (Subword):**
190
 
191
  | Rank | N-gram | Count |
192
  |------|--------|-------|
193
- | 1 | `а р а` | 50,267 |
194
- | 2 | `м з а` | 45,869 |
195
- | 3 | `з а _` | 44,868 |
196
- | 4 | `а _ а` | 35,436 |
197
- | 5 | `а м з` | 31,355 |
198
 
199
  **4-grams (Subword):**
200
 
201
  | Rank | N-gram | Count |
202
  |------|--------|-------|
203
- | 1 | `м з а _` | 44,436 |
204
- | 2 | `а м з а` | 30,784 |
205
- | 3 | `р а м з` | 22,744 |
206
- | 4 | `а р а _` | 19,480 |
207
- | 5 | `қ ә а _` | 17,506 |
 
 
 
 
 
 
 
 
 
 
208
 
209
 
210
  ### Key Findings
211
 
212
- - **Best Perplexity:** 3-gram (word) with 252
213
  - **Entropy Trend:** Decreases with larger n-grams (more predictable)
214
- - **Coverage:** Top-1000 patterns cover ~45% of corpus
215
  - **Recommendation:** 4-gram or 5-gram for best predictive performance
216
 
217
  ---
@@ -227,14 +259,14 @@ Below are sample sentences tokenized with each vocabulary size:
227
 
228
  | Context | Variant | Avg Entropy | Perplexity | Branching Factor | Unique Contexts | Predictability |
229
  |---------|---------|-------------|------------|------------------|-----------------|----------------|
230
- | **1** | Word | 0.6658 | 1.586 | 3.61 | 90,583 | 33.4% |
231
- | **1** | Subword | 1.3373 | 2.527 | 10.82 | 872 | 0.0% |
232
- | **2** | Word | 0.1207 | 1.087 | 1.22 | 326,705 | 87.9% |
233
- | **2** | Subword | 1.0105 | 2.015 | 5.95 | 9,435 | 0.0% |
234
- | **3** | Word | 0.0295 | 1.021 | 1.04 | 396,742 | 97.1% |
235
- | **3** | Subword | 0.7768 | 1.713 | 3.69 | 56,063 | 22.3% |
236
- | **4** | Word | 0.0100 🏆 | 1.007 | 1.01 | 412,289 | 99.0% |
237
- | **4** | Subword | 0.5289 | 1.443 | 2.33 | 206,852 | 47.1% |
238
 
239
  ### Generated Text Samples (Word-based)
240
 
@@ -242,27 +274,27 @@ Below are text samples generated from each word-based Markov chain model:
242
 
243
  **Context Size 1:**
244
 
245
- 1. `уи иалнаршеит абанкбжьаратә ҳасабкрақәа рҭыԥ инаркны шықәсанӡа наринџьованиархангельский а лазарев м...`
246
- 2. `рыԥсҭазаара иалҵит шықәса нанҳәа 5 ԥхынҷкәы́н 22 27 азы иҭыҵит раԥхьа ақырҭуа ҭыԥҳа викторина ж ж`
247
- 3. `иит рыԥсҭазаара иалҵит дидим халкентер абырзен бызшәала адраматургиа иақәшәаны ииашаҵәҟьаны еилкааӡа...`
248
 
249
  **Context Size 2:**
250
 
251
- 1. `иит рыԥсҭазаара иалҵит клеопатра селена ii марк антонии клеопатреи antony and cleopatra кориолан cor...`
252
- 2. `рашәарамза ԥхынгәымза нанҳәамза цәыббрамза жьҭаарамза абҵарамза ԥхынҷкәынмза иит рыԥсҭазаара иалҵит ...`
253
- 3. `жәабранмза хәажәкырамза мшаԥымза лаҵарамза рашәарамза ԥхынгәымза нанҳәамза цәыббрамза жьҭаарамза абҵ...`
254
 
255
  **Context Size 3:**
256
 
257
- 1. `цәыббрамза жьҭаарамза абҵарамза ԥхынҷкәынмза иит калигула римтәи аимператор дыԥсит 69 ш рыԥсҭазаара ...`
258
  2. `жьҭаарамза абҵарамза ԥхынҷкәынмза иит рыԥсҭазаара иалҵит аныҳәақәа араԥтә ар амш аҳәаахьчаҩцәа рамш ...`
259
  3. `ажьырныҳәамза жәабранмза хәажәкырамза мшаԥымза лаҵарамза рашәарамза ԥхынгәымза нанҳәамза цәыббрамза ...`
260
 
261
  **Context Size 4:**
262
 
263
- 1. `нанҳәамза цәыббрамза жьҭаарамза абҵарамза ԥхынҷкәынмза иит рыԥсҭазаара иалҵит арминии германиатә хер...`
264
- 2. `цәыббрамза жьҭаарамза абҵарамза ԥхынҷкәынмза иит рыԥсҭазаара иалҵит арминии германиатә херуски аимшь...`
265
- 3. `рашәарамза ԥхынгәымза нанҳәамза цәыббрамза жьҭаарамза абҵарамза ԥхынҷкәынмза иит рыԥсҭазаара иалҵит ...`
266
 
267
 
268
  ### Generated Text Samples (Subword-based)
@@ -271,34 +303,34 @@ Below are text samples generated from each subword-based Markov chain model:
271
 
272
  **Context Size 1:**
273
 
274
- 1. `аил_даит.._аарԥа`
275
- 2. `_дигьҭақәанбрббс`
276
- 3. `иунмақә_ser_иала`
277
 
278
  **Context Size 2:**
279
 
280
- 1. `а_иазара,_иа_мамҭ`
281
- 2. `_ашықәаԥсҭақәа_из`
282
- 3. `рамза_иционва,_30`
283
 
284
  **Context Size 3:**
285
 
286
- 1. `ара_амшын._уи_аҩны`
287
- 2. `мза_ԥхынгәымза_абе`
288
- 3. `за_мшаԥысуа_кәу._а`
289
 
290
  **Context Size 4:**
291
 
292
- 1. `мза_нанҳәамҭа._аҩаӡ`
293
- 2. `амза_рашәара_мап_рц`
294
- 3. `рамза_ԥхынгәы_9,_ш.`
295
 
296
 
297
  ### Key Findings
298
 
299
  - **Best Predictability:** Context-4 (word) with 99.0% predictability
300
  - **Branching Factor:** Decreases with context size (more deterministic)
301
- - **Memory Trade-off:** Larger contexts require more storage (206,852 contexts)
302
  - **Recommendation:** Context-3 or Context-4 for text generation
303
 
304
  ---
@@ -314,23 +346,23 @@ Below are text samples generated from each subword-based Markov chain model:
314
 
315
  | Metric | Value |
316
  |--------|-------|
317
- | Vocabulary Size | 32,686 |
318
- | Total Tokens | 440,475 |
319
- | Mean Frequency | 13.48 |
320
  | Median Frequency | 3 |
321
- | Frequency Std Dev | 100.81 |
322
 
323
  ### Most Common Words
324
 
325
  | Rank | Word | Frequency |
326
  |------|------|-----------|
327
- | 1 | уи | 4,159 |
328
  | 2 | рыԥсҭазаара | 4,025 |
329
  | 3 | иит | 3,987 |
330
  | 4 | иалҵит | 3,980 |
331
  | 5 | лаҵарамза | 3,752 |
332
  | 6 | жәабранмза | 3,722 |
333
- | 7 | хәажәкырамза | 3,701 |
334
  | 8 | абҵарамза | 3,701 |
335
  | 9 | нанҳәамза | 3,696 |
336
  | 10 | ԥхынҷкәынмза | 3,696 |
@@ -339,23 +371,23 @@ Below are text samples generated from each subword-based Markov chain model:
339
 
340
  | Rank | Word | Frequency |
341
  |------|------|-----------|
342
- | 1 | slang | 2 |
343
- | 2 | пуэрто | 2 |
344
- | 3 | испантәи | 2 |
345
- | 4 | reggaetón | 2 |
346
- | 5 | маратон | 2 |
347
- | 6 | урымтәыла | 2 |
348
- | 7 | византии | 2 |
349
- | 8 | белобров | 2 |
350
- | 9 | длины | 2 |
351
- | 10 | акармара | 2 |
352
 
353
  ### Zipf's Law Analysis
354
 
355
  | Metric | Value |
356
  |--------|-------|
357
- | Zipf Coefficient | 0.9638 |
358
- | R² (Goodness of Fit) | 0.995375 |
359
  | Adherence Quality | **excellent** |
360
 
361
  ### Coverage Analysis
@@ -363,7 +395,7 @@ Below are text samples generated from each subword-based Markov chain model:
363
  | Top N Words | Coverage |
364
  |-------------|----------|
365
  | Top 100 | 30.3% |
366
- | Top 1,000 | 55.8% |
367
  | Top 5,000 | 76.9% |
368
  | Top 10,000 | 85.7% |
369
 
@@ -371,7 +403,7 @@ Below are text samples generated from each subword-based Markov chain model:
371
 
372
  - **Zipf Compliance:** R²=0.9954 indicates excellent adherence to Zipf's law
373
  - **High Frequency Dominance:** Top 100 words cover 30.3% of corpus
374
- - **Long Tail:** 22,686 words needed for remaining 14.3% coverage
375
 
376
  ---
377
  ## 5. Word Embeddings Evaluation
@@ -387,37 +419,40 @@ Below are text samples generated from each subword-based Markov chain model:
387
 
388
  ### 5.1 Cross-Lingual Alignment
389
 
390
- > *Note: Multilingual alignment visualization not available for this language.*
 
 
391
 
392
 
393
  ### 5.2 Model Comparison
394
 
395
  | Model | Dimension | Isotropy | Semantic Density | Alignment R@1 | Alignment R@10 |
396
  |-------|-----------|----------|------------------|---------------|----------------|
397
- | **mono_32d** | 32 | 0.8185 🏆 | 0.3505 | N/A | N/A |
398
- | **mono_64d** | 64 | 0.6063 | 0.2895 | N/A | N/A |
399
- | **mono_128d** | 128 | 0.1746 | 0.2841 | N/A | N/A |
 
 
 
400
 
401
  ### Key Findings
402
 
403
- - **Best Isotropy:** mono_32d with 0.8185 (more uniform distribution)
404
  - **Semantic Density:** Average pairwise similarity of 0.3080. Lower values indicate better semantic separation.
405
- - **Alignment Quality:** No aligned models evaluated in this run.
406
  - **Recommendation:** 128d aligned for best cross-lingual performance
407
 
408
  ---
409
  ## 6. Morphological Analysis (Experimental)
410
 
411
- > ⚠️ **Warning:** This language shows low morphological productivity. The statistical signals used for this analysis may be noisy or less reliable than for morphologically rich languages.
412
-
413
  This section presents an automated morphological analysis derived from the statistical divergence between word-level and subword-level models. By analyzing where subword predictability spikes and where word-level coverage fails, we can infer linguistic structures without supervised data.
414
 
415
  ### 6.1 Productivity & Complexity
416
 
417
  | Metric | Value | Interpretation | Recommendation |
418
  |--------|-------|----------------|----------------|
419
- | Productivity Index | **0.000** | Low morphological productivity | ⚠️ Likely unreliable |
420
- | Idiomaticity Gap | **-1.000** | Low formulaic content | - |
421
 
422
  ### 6.2 Affix Inventory (Productive Units)
423
 
@@ -426,19 +461,19 @@ These are the most productive prefixes and suffixes identified by sampling the v
426
  #### Productive Prefixes
427
  | Prefix | Examples |
428
  |--------|----------|
429
- | `-иа` | иаанхон, иаҳәшьа, иагьлыхьӡиҵоит |
430
 
431
  #### Productive Suffixes
432
  | Suffix | Examples |
433
  |--------|----------|
434
- | `-а` | азербайджана, рыҟамлара, аҭӡысахьақәа |
435
- | `-әа` | аҭӡысахьақәа, льготақәа, ааимҭақәа |
436
- | `-қәа` | аҭӡысахьақәа, льготақәа, ааимҭақәа |
437
- | `-ит` | еиҭарҳәоит, иагьлыхьӡиҵоит, рыдиулоит |
438
- | `-ра` | рыҟамлара, архынҳәра, ахынҳәра |
439
- | `-еи` | аԥсреи, русқәеи, аглобализациеи |
440
- | `-тә` | антилопатә, аҳазылхратә, апрозатә |
441
- | `-ақәа` | аҭӡысахьақәа, льготақәа, ааимҭақәа |
442
 
443
  ### 6.3 Bound Stems (Lexical Roots)
444
 
@@ -446,18 +481,18 @@ Bound stems are high-frequency subword units that are semantically cohesive but
446
 
447
  | Stem | Cohesion | Substitutability | Examples |
448
  |------|----------|------------------|----------|
449
- | `гыла` | 1.69x | 82 contexts | гылан, ргылан, игылаз |
450
- | `аҵар` | 1.72x | 38 contexts | аҵара, лаҵара, хаҵара |
451
- | `әыла` | 1.72x | 34 contexts | тәыла, атәыла, тәылан |
452
- | `ықәс` | 1.85x | 26 contexts | шықәс, шықәсы, шықәса |
453
- | `әара` | 1.47x | 58 contexts | шәара, акәара, аҟәара |
454
- | `арам` | 2.03x | 17 contexts | нарам, харам, арамка |
455
- | `қәса` | 2.02x | 16 contexts | шықәса, щықәса, жәшықәса |
456
- | `шәар` | 1.69x | 26 contexts | шәара, ршәарц, шәарах |
457
- | `ҭаза` | 2.47x | 8 contexts | иԥсҭазара, ԥсҭазаара, ипсҭазаара |
458
  | `азаа` | 1.69x | 23 contexts | лазаа, амазаап, иазааит |
459
- | `ыҳәа` | 1.62x | 22 contexts | ныҳәа, иныҳәа, ныҳәан |
460
- | `заар` | 2.08x | 10 contexts | акзаара, акзаарак, ракзаара |
 
 
 
461
 
462
  ### 6.4 Affix Compatibility (Co-occurrence)
463
 
@@ -465,16 +500,15 @@ This table shows which prefixes and suffixes most frequently co-occur on the sam
465
 
466
  | Prefix | Suffix | Frequency | Examples |
467
  |--------|--------|-----------|----------|
468
- | `-иа` | `-ит` | 97 words | иахьӡырҵеит, иаҭаауеит |
469
- | `-иа` | `-еит` | 54 words | иахьӡырҵеит, иаҭаауеит |
470
- | `-иа` | `-а` | 42 words | иалихуа, иадыруа |
471
- | `-иа` | `-әа` | 14 words | иаламҵакәа, иажәа |
472
- | `-иа` | `-ра` | 5 words | иабшьҭра, иааӡара |
473
- | `-иа` | `-тә` | 5 words | иаҳратә, иашахаҵаратә |
474
- | `-иа` | `-қәа` | 5 words | иашақәа, ианҵамҭақәа |
475
- | `-иа` | `-ақәа` | 4 words | иашақәа, ианҵамҭақәа |
476
- | `-иа` | `-еи` | 3 words | иашьцәеи, иамхреи |
477
- | `-иа` | `-әи` | 1 words | иапониатәи |
478
 
479
  ### 6.5 Recursive Morpheme Segmentation
480
 
@@ -482,26 +516,28 @@ Using **Recursive Hierarchical Substitutability**, we decompose complex words in
482
 
483
  | Word | Suggested Split | Confidence | Stem |
484
  |------|-----------------|------------|------|
 
 
 
 
485
  | аҳәынҭқарқәа | **`аҳәынҭқар-қәа`** | 4.5 | `аҳәынҭқар` |
486
- | мраҭашәаратә | **`мраҭаш-әа-ра-тә`** | 4.5 | `мраҭаш` |
487
- | ашьауӷатә | **`ашьауӷа-тә`** | 4.5 | `ашьауӷа` |
488
- | алахәылара | **`алахәыла-ра`** | 4.5 | `алахәыла` |
489
- | амитингқәа | **`амитинг-қәа`** | 4.5 | `амитинг` |
490
- | аметаллургиатә | **`аметаллургиа-тә`** | 4.5 | `аметаллургиа` |
491
- | рекологиатә | **`рекологиа-тә`** | 4.5 | `рекологиа` |
492
- | асиужетқәа | **`асиужет-қәа`** | 4.5 | `асиужет` |
493
- | аконцерттә | **`аконцерт-тә`** | 4.5 | `аконцерт` |
494
- | ипартиатә | **`ипартиа-тә`** | 4.5 | `ипартиа` |
495
- | афасадқәа | **`афасад-қәа`** | 4.5 | `афасад` |
496
- | астадионқәа | **`астадион-қәа`** | 4.5 | `астадион` |
497
- | аномерқәа | **`аномер-қәа`** | 4.5 | `аномер` |
498
- | аныҳәаратә | **`аныҳ-әа-ра-тә`** | 4.5 | `аныҳ` |
499
- | атерминқәа | **`атермин-қәа`** | 4.5 | `атермин` |
500
 
501
  ### 6.6 Linguistic Interpretation
502
 
503
  > **Automated Insight:**
504
- The language AB appears to be more isolating or has a highly fixed vocabulary. Word-level models perform nearly as well as subword models, indicating fewer productive morphological processes.
 
 
505
 
506
  ---
507
  ## 7. Summary & Recommendations
@@ -513,7 +549,7 @@ The language AB appears to be more isolating or has a highly fixed vocabulary. W
513
  | Component | Recommended | Rationale |
514
  |-----------|-------------|-----------|
515
  | Tokenizer | **64k BPE** | Best compression (4.19x) |
516
- | N-gram | **3-gram** | Lowest perplexity (252) |
517
  | Markov | **Context-4** | Highest predictability (99.0%) |
518
  | Embeddings | **100d** | Balanced semantic capture and isotropy |
519
 
@@ -728,4 +764,4 @@ MIT License - Free for academic and commercial use.
728
  ---
729
  *Generated by Wikilangs Models Pipeline*
730
 
731
- *Report Date: 2026-01-03 05:05:55*
 
1
  ---
2
  language: ab
3
+ language_name: Abkhazian
4
  language_family: caucasian_northwest
5
  tags:
6
  - wikilangs
 
10
  - n-gram
11
  - markov
12
  - wikipedia
13
+ - feature-extraction
14
+ - sentence-similarity
15
+ - tokenization
16
+ - n-grams
17
+ - markov-chain
18
+ - text-mining
19
+ - fasttext
20
+ - babelvec
21
+ - vocabulous
22
+ - vocabulary
23
  - monolingual
24
  - family-caucasian_northwest
25
  license: mit
26
  library_name: wikilangs
27
+ pipeline_tag: text-generation
28
  datasets:
29
  - omarkamali/wikipedia-monthly
30
  dataset_info:
 
36
  value: 4.193
37
  - name: best_isotropy
38
  type: isotropy
39
+ value: 0.8394
40
  - name: vocabulary_size
41
  type: vocab
42
  value: 0
43
  generated: 2026-01-03
44
  ---
45
 
46
+ # Abkhazian - Wikilangs Models
47
  ## Comprehensive Research Report & Full Ablation Study
48
 
49
+ This repository contains NLP models trained and evaluated by Wikilangs, specifically on **Abkhazian** Wikipedia data.
50
  We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and word embeddings.
51
 
52
  ## 📋 Repository Contents
 
70
  - [3. Markov Chain Evaluation](#3-markov-chain-evaluation)
71
  - [4. Vocabulary Analysis](#4-vocabulary-analysis)
72
  - [5. Word Embeddings Evaluation](#5-word-embeddings-evaluation)
73
+ - [6. Morphological Analysis (Experimental)](#6--morphological-analysis-experimental)
74
  - [7. Summary & Recommendations](#7-summary--recommendations)
75
  - [Metrics Glossary](#appendix-metrics-glossary--interpretation-guide)
76
  - [Visualizations Index](#visualizations-index)
 
90
 
91
  | Vocab Size | Compression | Avg Token Len | UNK Rate | Total Tokens |
92
  |------------|-------------|---------------|----------|--------------|
93
+ | **8k** | 3.306x | 3.31 | 0.1493% | 223,032 |
94
+ | **16k** | 3.654x | 3.66 | 0.1650% | 201,823 |
95
+ | **32k** | 3.910x | 3.92 | 0.1766% | 188,563 |
96
+ | **64k** | 4.193x 🏆 | 4.20 | 0.1893% | 175,871 |
97
 
98
  ### Tokenization Examples
99
 
100
  Below are sample sentences tokenized with each vocabulary size:
101
 
102
+ **Sample 1:** `Ѳ, ѳкириллтәи аҩыратә архаикатә иажәхьоу нбан. Азхьарԥшқәа Graphemica (Ѳ) Gra...`
103
 
104
  | Vocab | Tokens | Count |
105
  |-------|--------|-------|
106
+ | 8k | `▁ ѳ , ▁ ѳ ▁— ▁кириллтәи ▁аҩыратә ▁архаикатә ▁иажәхьоу ... (+11 more)` | 21 |
107
+ | 16k | `▁ѳ , ▁ѳ ▁— ▁кириллтәи ▁аҩыратә ▁архаикатә ▁иажәхьоу ▁нбан . ... (+9 more)` | 19 |
108
+ | 32k | `▁ѳ , ▁ѳ ▁— ▁кириллтәи ▁аҩыратә ▁архаикатә ▁иажәхьоу ▁нбан . ... (+9 more)` | 19 |
109
+ | 64k | `▁ѳ , ▁ѳ ▁— ▁кириллтәи ▁аҩыратә ▁архаикатә ▁иажәхьоу ▁нбан . ... (+9 more)` | 19 |
110
 
111
+ **Sample 2:** `Скуо-Уелли Winter Olympics, Jeux olympiques d'hiver de - аӡынтәи Олимпиадатә хәм...`
112
 
113
  | Vocab | Tokens | Count |
114
  |-------|--------|-------|
115
+ | 8k | `▁с ку о - у елли ▁winter ▁olympics , ▁jeux ... (+12 more)` | 22 |
116
+ | 16k | `▁с ку о - у елли ▁winter ▁olympics , jeux ... (+12 more)` | 22 |
117
+ | 32k | `▁с ку о - уелли ▁winter ▁olympics , jeux ▁olympiques ... (+11 more)` | 21 |
118
+ | 64k | `▁скуо - уелли ▁winter ▁olympics , jeux ▁olympiquesd ' ... (+9 more)` | 19 |
119
 
120
+ **Sample 3:** `Ж, ж кириллтәи аҩыратә нбан. Азхьарԥшқәа Graphemica (Ж) Graphemica (ж)`
121
 
122
  | Vocab | Tokens | Count |
123
  |-------|--------|-------|
124
+ | 8k | `▁ж , ▁ж ▁— ▁кириллтәи ▁аҩыратә ▁нбан . ▁азхьарԥшқәа ▁graphemica ... (+7 more)` | 17 |
125
+ | 16k | `▁ж , ▁ж ▁— ▁кириллтәи ▁аҩыратә ▁нбан . ▁азхьарԥшқәа ▁graphemica ... (+7 more)` | 17 |
126
+ | 32k | `▁ж , ▁ж ▁— ▁кириллтәи ▁аҩыратә ▁нбан . ▁азхьарԥшқәа ▁graphemica ... (+7 more)` | 17 |
127
+ | 64k | `▁ж , ▁ж ▁— ▁кириллтәи ▁аҩыратә ▁нбан . ▁азхьарԥшқәа ▁graphemica ... (+7 more)` | 17 |
128
 
129
 
130
  ### Key Findings
131
 
132
  - **Best Compression:** 64k achieves 4.193x compression
133
+ - **Lowest UNK Rate:** 8k with 0.1493% unknown tokens
134
  - **Trade-off:** Larger vocabularies improve compression but increase model size
135
  - **Recommendation:** 32k vocabulary provides optimal balance for production use
136
 
 
147
 
148
  | N-gram | Variant | Perplexity | Entropy | Unique N-grams | Top-100 Coverage | Top-1000 Coverage |
149
  |--------|---------|------------|---------|----------------|------------------|-------------------|
150
+ | **2-gram** | Word | 723 | 9.50 | 5,814 | 51.5% | 72.0% |
151
+ | **2-gram** | Subword | 363 | 8.51 | 4,117 | 60.3% | 96.8% |
152
+ | **3-gram** | Word | 252 | 7.98 | 5,218 | 66.6% | 80.6% |
153
+ | **3-gram** | Subword | 2,678 | 11.39 | 28,284 | 28.1% | 67.5% |
154
+ | **4-gram** | Word | 341 | 8.41 | 9,794 | 64.0% | 74.0% |
155
+ | **4-gram** | Subword | 11,104 | 13.44 | 112,814 | 16.8% | 44.7% |
156
+ | **5-gram** | Word | 198 🏆 | 7.63 | 7,301 | 69.5% | 78.6% |
157
+ | **5-gram** | Subword | 26,131 | 14.67 | 211,528 | 13.8% | 34.5% |
158
 
159
  ### Top 5 N-grams by Size
160
 
 
166
  | 2 | `иит рыԥсҭазаара` | 3,938 |
167
  | 3 | `рашәарамза ԥхынгәымза` | 3,603 |
168
  | 4 | `жәабранмза хәажәкырамза` | 3,603 |
169
+ | 5 | `цәыббрамза жьҭаарамза` | 3,602 |
170
 
171
  **3-grams (Word):**
172
 
 
175
  | 1 | `иит рыԥсҭазаара иалҵит` | 3,938 |
176
  | 2 | `цәыббрамза жьҭаарамза абҵарамза` | 3,602 |
177
  | 3 | `нанҳәамза цәыббрамза жьҭаарамза` | 3,601 |
178
+ | 4 | `жьҭаарамза абҵарамза ԥхынҷкәынмза` | 3,601 |
179
+ | 5 | `лаҵарамза рашәарамза ԥхынгәымза` | 3,601 |
180
 
181
  **4-grams (Word):**
182
 
183
  | Rank | N-gram | Count |
184
  |------|--------|-------|
185
+ | 1 | `цәыббрамза жьҭаарамза абҵарамза ԥхынҷкәынмза` | 3,601 |
186
+ | 2 | `нанҳәамза цәыббрамза жьҭаарамза абҵарамза` | 3,601 |
187
+ | 3 | `ԥхынгәымза нанҳәамза цәыббрамза жьҭаарамза` | 3,601 |
188
+ | 4 | `мшаԥымза лаҵарамза рашәарамза ԥхынгәымза` | 3,601 |
189
+ | 5 | `лаҵарамза рашәарамза ԥхынгәымза нанҳәамза` | 3,601 |
190
+
191
+ **5-grams (Word):**
192
+
193
+ | Rank | N-gram | Count |
194
+ |------|--------|-------|
195
+ | 1 | `мшаԥымза лаҵарамза рашәарамза ԥхынгәымза нанҳәамза` | 3,601 |
196
+ | 2 | `рашәарамза ԥхынгәымза нанҳәамза цәыббрамза жьҭаарамза` | 3,601 |
197
+ | 3 | `ԥхынгәымза нанҳәамза цәыббрамза жьҭаарамза абҵарамза` | 3,601 |
198
+ | 4 | `нанҳәамза цәыббрамза жьҭаарамза абҵарамза ԥхынҷкәынмза` | 3,601 |
199
+ | 5 | `лаҵарамза рашәарамза ԥхынгәымза нанҳәамза цәыббрамза` | 3,601 |
200
 
201
  **2-grams (Subword):**
202
 
203
  | Rank | N-gram | Count |
204
  |------|--------|-------|
205
+ | 1 | `а _` | 154,936 |
206
+ | 2 | `_ а` | 150,057 |
207
+ | 3 | `р а` | 100,657 |
208
+ | 4 | `а р` | 84,729 |
209
+ | 5 | `ә а` | 76,114 |
210
 
211
  **3-grams (Subword):**
212
 
213
  | Rank | N-gram | Count |
214
  |------|--------|-------|
215
+ | 1 | `а р а` | 50,339 |
216
+ | 2 | `м з а` | 45,875 |
217
+ | 3 | `з а _` | 44,872 |
218
+ | 4 | `а _ а` | 35,534 |
219
+ | 5 | `а м з` | 31,361 |
220
 
221
  **4-grams (Subword):**
222
 
223
  | Rank | N-gram | Count |
224
  |------|--------|-------|
225
+ | 1 | `м з а _` | 44,438 |
226
+ | 2 | `а м з а` | 30,790 |
227
+ | 3 | `р а м з` | 22,745 |
228
+ | 4 | `а р а _` | 19,530 |
229
+ | 5 | `қ ә а _` | 17,562 |
230
+
231
+ **5-grams (Subword):**
232
+
233
+ | Rank | N-gram | Count |
234
+ |------|--------|-------|
235
+ | 1 | `а м з а _` | 29,604 |
236
+ | 2 | `р а м з а` | 22,366 |
237
+ | 3 | `а р а м з` | 15,138 |
238
+ | 4 | `т ә и _ а` | 11,926 |
239
+ | 5 | `а қ ә а _` | 9,350 |
240
 
241
 
242
  ### Key Findings
243
 
244
+ - **Best Perplexity:** 5-gram (word) with 198
245
  - **Entropy Trend:** Decreases with larger n-grams (more predictable)
246
+ - **Coverage:** Top-1000 patterns cover ~34% of corpus
247
  - **Recommendation:** 4-gram or 5-gram for best predictive performance
248
 
249
  ---
 
259
 
260
  | Context | Variant | Avg Entropy | Perplexity | Branching Factor | Unique Contexts | Predictability |
261
  |---------|---------|-------------|------------|------------------|-----------------|----------------|
262
+ | **1** | Word | 0.6658 | 1.586 | 3.61 | 90,782 | 33.4% |
263
+ | **1** | Subword | 1.3353 | 2.523 | 10.79 | 879 | 0.0% |
264
+ | **2** | Word | 0.1206 | 1.087 | 1.22 | 327,437 | 87.9% |
265
+ | **2** | Subword | 1.0094 | 2.013 | 5.94 | 9,477 | 0.0% |
266
+ | **3** | Word | 0.0294 | 1.021 | 1.04 | 397,532 | 97.1% |
267
+ | **3** | Subword | 0.7766 | 1.713 | 3.69 | 56,288 | 22.3% |
268
+ | **4** | Word | 0.0100 🏆 | 1.007 | 1.01 | 413,065 | 99.0% |
269
+ | **4** | Subword | 0.5281 | 1.442 | 2.33 | 207,598 | 47.2% |
270
 
271
  ### Generated Text Samples (Word-based)
272
 
 
274
 
275
  **Context Size 1:**
276
 
277
+ 1. `уи зыхҟьаз зеиҧш дыҟамыз аҧҳәызба ссир иргылеит еидҵоу қырҭтәыла адемократиатә хдырра асоциалтә хьча...`
278
+ 2. `рыԥсҭазаара иалҵит пиотр актәи амаӡаныҟәгаҩыс ш вуковар vukovar jedna prica ш азхьарԥшқәа heritagesi...`
279
+ 3. `иит рыԥсҭазаара иалҵит кринагор абырзен бызшәа афранцыз италиа иалаигалоит флоренцианӡагьы инеиуеит ...`
280
 
281
  **Context Size 2:**
282
 
283
+ 1. `иит рыԥсҭазаара иалҵит октавиан август аԥеиԥа диит ҳ ҟ 326 мцхеҭа ҳ ҟ 14 ш абанктә система`
284
+ 2. `жәабранмза хәажәкырамза мшаԥымза лаҵарамза рашәарамза ԥхынгәымза нанҳәамза цәыббрамза жьҭаарамза абҵ...`
285
+ 3. `рашәарамза ԥхынгәымза нанҳәамза цәыббрамза жьҭаарамза абҵарамза ԥхынҷкәынмза иит рыԥсҭазаара иалҵит ...`
286
 
287
  **Context Size 3:**
288
 
289
+ 1. `цәыббрамза жьҭаарамза абҵарамза ԥхынҷкәынмза иит рыԥсҭазаара иалҵит аныҳәақәа араԥтә ар амш аҳәаахьч...`
290
  2. `жьҭаарамза абҵарамза ԥхынҷкәынмза иит рыԥсҭазаара иалҵит аныҳәақәа араԥтә ар амш аҳәаахьчаҩцәа рамш ...`
291
  3. `ажьырныҳәамза жәабранмза хәажәкырамза мшаԥымза лаҵарамза рашәарамза ԥхынгәымза нанҳәамза цәыббрамза ...`
292
 
293
  **Context Size 4:**
294
 
295
+ 1. `цәыббрамза жьҭаарамза абҵарамза ԥхынҷкәынмза иит рыԥсҭазаара иалҵит клавдиа пульхра римтәи аамсҭаԥхә...`
296
+ 2. `лаҵарамза рашәарамза ԥхынгәымза нанҳәамза цәыббрамза жьҭаарамза абҵарамза ԥхынҷкәынмза иит рыԥсҭазаа...`
297
+ 3. `ԥхынгәымза нанҳәамза цәыббрамза жьҭаарамза абҵарамза ԥхынҷкәынмза иит рыԥсҭазаара иалҵит клавдиа пул...`
298
 
299
 
300
  ### Generated Text Samples (Subword-based)
 
303
 
304
  **Context Size 1:**
305
 
306
+ 1. `аякарамаҟәаҿы_«п`
307
+ 2. `_жьы_ажәынқәсп_и`
308
+ 3. `иха_аббарран._ло`
309
 
310
  **Context Size 2:**
311
 
312
+ 1. `а_уи_ахьы_иркую_с`
313
+ 2. `_ареит._ара_ихьам`
314
+ 3. `рала_ԥхын,_хьшара`
315
 
316
  **Context Size 3:**
317
 
318
+ 1. `араҟнытә_бызшәалеи`
319
+ 2. `мза_жьҭаарамза_жәа`
320
+ 3. `за_ԥхынгәырый_фано`
321
 
322
  **Context Size 4:**
323
 
324
+ 1. `мза_ракәзар,_зныз_х`
325
+ 2. `амза_рашәара,_шықәс`
326
+ 3. `рамза_ԥхынгәымза_ла`
327
 
328
 
329
  ### Key Findings
330
 
331
  - **Best Predictability:** Context-4 (word) with 99.0% predictability
332
  - **Branching Factor:** Decreases with context size (more deterministic)
333
+ - **Memory Trade-off:** Larger contexts require more storage (207,598 contexts)
334
  - **Recommendation:** Context-3 or Context-4 for text generation
335
 
336
  ---
 
346
 
347
  | Metric | Value |
348
  |--------|-------|
349
+ | Vocabulary Size | 32,744 |
350
+ | Total Tokens | 441,086 |
351
+ | Mean Frequency | 13.47 |
352
  | Median Frequency | 3 |
353
+ | Frequency Std Dev | 100.78 |
354
 
355
  ### Most Common Words
356
 
357
  | Rank | Word | Frequency |
358
  |------|------|-----------|
359
+ | 1 | уи | 4,161 |
360
  | 2 | рыԥсҭазаара | 4,025 |
361
  | 3 | иит | 3,987 |
362
  | 4 | иалҵит | 3,980 |
363
  | 5 | лаҵарамза | 3,752 |
364
  | 6 | жәабранмза | 3,722 |
365
+ | 7 | хәажәкырамза | 3,702 |
366
  | 8 | абҵарамза | 3,701 |
367
  | 9 | нанҳәамза | 3,696 |
368
  | 10 | ԥхынҷкәынмза | 3,696 |
 
371
 
372
  | Rank | Word | Frequency |
373
  |------|------|-----------|
374
+ | 1 | sons | 2 |
375
+ | 2 | extended | 2 |
376
+ | 3 | stream | 2 |
377
+ | 4 | block | 2 |
378
+ | 5 | stru | 2 |
379
+ | 6 | compressed | 2 |
380
+ | 7 | deflate | 2 |
381
+ | 8 | january | 2 |
382
+ | 9 | видеохәмарроуп | 2 |
383
+ | 10 | роблокс | 2 |
384
 
385
  ### Zipf's Law Analysis
386
 
387
  | Metric | Value |
388
  |--------|-------|
389
+ | Zipf Coefficient | 0.9626 |
390
+ | R² (Goodness of Fit) | 0.995444 |
391
  | Adherence Quality | **excellent** |
392
 
393
  ### Coverage Analysis
 
395
  | Top N Words | Coverage |
396
  |-------------|----------|
397
  | Top 100 | 30.3% |
398
+ | Top 1,000 | 55.7% |
399
  | Top 5,000 | 76.9% |
400
  | Top 10,000 | 85.7% |
401
 
 
403
 
404
  - **Zipf Compliance:** R²=0.9954 indicates excellent adherence to Zipf's law
405
  - **High Frequency Dominance:** Top 100 words cover 30.3% of corpus
406
+ - **Long Tail:** 22,744 words needed for remaining 14.3% coverage
407
 
408
  ---
409
  ## 5. Word Embeddings Evaluation
 
419
 
420
  ### 5.1 Cross-Lingual Alignment
421
 
422
+ ![Alignment Quality](visualizations/embedding_alignment_quality.png)
423
+
424
+ ![Multilingual t-SNE](visualizations/embedding_tsne_multilingual.png)
425
 
426
 
427
  ### 5.2 Model Comparison
428
 
429
  | Model | Dimension | Isotropy | Semantic Density | Alignment R@1 | Alignment R@10 |
430
  |-------|-----------|----------|------------------|---------------|----------------|
431
+ | **mono_32d** | 32 | 0.8394 | 0.3485 | N/A | N/A |
432
+ | **mono_64d** | 64 | 0.5679 | 0.2942 | N/A | N/A |
433
+ | **mono_128d** | 128 | 0.1636 | 0.2836 | N/A | N/A |
434
+ | **aligned_32d** | 32 | 0.8394 🏆 | 0.3421 | 0.0220 | 0.1360 |
435
+ | **aligned_64d** | 64 | 0.5679 | 0.2946 | 0.0360 | 0.1960 |
436
+ | **aligned_128d** | 128 | 0.1636 | 0.2850 | 0.0420 | 0.2180 |
437
 
438
  ### Key Findings
439
 
440
+ - **Best Isotropy:** aligned_32d with 0.8394 (more uniform distribution)
441
  - **Semantic Density:** Average pairwise similarity of 0.3080. Lower values indicate better semantic separation.
442
+ - **Alignment Quality:** Aligned models achieve up to 4.2% R@1 in cross-lingual retrieval.
443
  - **Recommendation:** 128d aligned for best cross-lingual performance
444
 
445
  ---
446
  ## 6. Morphological Analysis (Experimental)
447
 
 
 
448
  This section presents an automated morphological analysis derived from the statistical divergence between word-level and subword-level models. By analyzing where subword predictability spikes and where word-level coverage fails, we can infer linguistic structures without supervised data.
449
 
450
  ### 6.1 Productivity & Complexity
451
 
452
  | Metric | Value | Interpretation | Recommendation |
453
  |--------|-------|----------------|----------------|
454
+ | Productivity Index | **2.615** | High morphological productivity | Reliable analysis |
455
+ | Idiomaticity Gap | **1.280** | High formulaic/idiomatic content | - |
456
 
457
  ### 6.2 Affix Inventory (Productive Units)
458
 
 
461
  #### Productive Prefixes
462
  | Prefix | Examples |
463
  |--------|----------|
464
+ | `-иа` | иалаҵоу, иааргазар, иадлоит |
465
 
466
  #### Productive Suffixes
467
  | Suffix | Examples |
468
  |--------|----------|
469
+ | `-а` | акандидатцәа, азура, ашәара |
470
+ | `-әа` | акандидатцәа, акрақәа, аконсультациақәа |
471
+ | `-ит` | иҟамлеит, дагәыланахалоит, дашьҭалоит |
472
+ | `-қәа` | акрақәа, аконсультациақәа, дунеихәаԥшрақәа |
473
+ | `-ра` | азура, ашәара, рықәцара |
474
+ | `-тә` | алашаратә, аҵакырадгьылтә, аетнографиатә |
475
+ | `-еи` | ргәыԥқәеи, астатуиақәеи, аизгақәеи |
476
+ | `-еит` | иҟамлеит, ԥхасҭахеит, игәарҭеит |
477
 
478
  ### 6.3 Bound Stems (Lexical Roots)
479
 
 
481
 
482
  | Stem | Cohesion | Substitutability | Examples |
483
  |------|----------|------------------|----------|
484
+ | `гыла` | 1.73x | 82 contexts | гылан, ргылан, дгылан |
485
+ | `ықәс` | 1.84x | 26 contexts | шықәс, щықәса, ашықәс |
486
+ | `әыла` | 1.68x | 34 contexts | тәыла, тәылак, ртәыла |
487
+ | `аҵар` | 1.63x | 38 contexts | аҵара, лаҵара, аҵареи |
488
+ | `қәса` | 1.96x | 16 contexts | щықәса, шықәса, шиқәсазы |
489
+ | `арам` | 1.86x | 17 contexts | харам, нарам, гуарам |
 
 
 
490
  | `азаа` | 1.69x | 23 contexts | лазаа, амазаап, иазааит |
491
+ | `әара` | 1.30x | 58 contexts | шәара, акәара, ҿҳәара |
492
+ | `ҭаза` | 2.37x | 8 contexts | иԥсҭазара, ԥсҭазаара, иԥсҭазаара |
493
+ | `шәар` | 1.56x | 26 contexts | шәара, шәарах, ашәара |
494
+ | `заар` | 2.09x | 10 contexts | акзаара, аҟазаара, акзаареи |
495
+ | `ыҳәа` | 1.57x | 22 contexts | ныҳәа, рныҳәа, иныҳәа |
496
 
497
  ### 6.4 Affix Compatibility (Co-occurrence)
498
 
 
500
 
501
  | Prefix | Suffix | Frequency | Examples |
502
  |--------|--------|-----------|----------|
503
+ | `-иа` | `-ит` | 83 words | иаабоит, иацхраауеит |
504
+ | `-иа` | `-еит` | 50 words | иацхраауеит, иартәеит |
505
+ | `-иа` | `-а` | 43 words | ианырба, ианрылага |
506
+ | `-иа` | `-әа` | 11 words | иацәыхарамкәа, иаламлакәа |
507
+ | `-иа` | `-тә` | 5 words | иааникыларатә, иавтобиографиатә |
508
+ | `-иа` | `-ра` | 3 words | иавторра, иамхра |
509
+ | `-иа` | `-еи` | 2 words | ианԥсеи, иашьцәеи |
510
+ | `-иа` | `-қәа` | 2 words | иажәақәа, иажәамаанақәа |
511
+ | `-иа` | `-ақәа` | 1 words | иажәақәа, иажәамаанақәа |
 
512
 
513
  ### 6.5 Recursive Morpheme Segmentation
514
 
 
516
 
517
  | Word | Suggested Split | Confidence | Stem |
518
  |------|-----------------|------------|------|
519
+ | анхарҭатә | **`анхарҭа-тә`** | 4.5 | `анхарҭа` |
520
+ | рхыԥхьаӡараҟнытә | **`рхыԥхьаӡараҟны-тә`** | 4.5 | `рхыԥхьаӡараҟны` |
521
+ | аӡхықәқәа | **`аӡхықә-қәа`** | 4.5 | `аӡхықә` |
522
+ | астуденттә | **`астудент-тә`** | 4.5 | `астудент` |
523
  | аҳәынҭқарқәа | **`аҳәынҭқар-қәа`** | 4.5 | `аҳәынҭқар` |
524
+ | каталониатә | **`каталониа-тә`** | 4.5 | `каталониа` |
525
+ | абиблиографиатә | **`абиблиографиа-тә`** | 4.5 | `абиблиографиа` |
526
+ | аредакциатә | **`аредакциа-тә`** | 4.5 | `аредакциа` |
527
+ | амилициатә | **`амилициа-тә`** | 4.5 | `амилициа` |
528
+ | амилаҭқәа | **`амилаҭ-қәа`** | 4.5 | `амилаҭ` |
529
+ | аекологиатә | **`аекологиа-тә`** | 4.5 | `аекологиа` |
530
+ | адемографиатә | **`адемографиа-тә`** | 4.5 | `адемографиа` |
531
+ | аконсервациатә | **`аконсервациа-тә`** | 4.5 | `аконсервациа` |
532
+ | ауаҩытәыҩсатә | **`ауаҩытәыҩса-тә`** | 4.5 | `ауаҩытәыҩса` |
533
+ | аелементқәа | **`аелемент-қәа`** | 4.5 | `аелемент` |
 
 
 
 
534
 
535
  ### 6.6 Linguistic Interpretation
536
 
537
  > **Automated Insight:**
538
+ The language Abkhazian shows high morphological productivity. The subword models are significantly more efficient than word models, suggesting a rich system of affixation or compounding.
539
+
540
+ > **Note on Idiomaticity:** The high Idiomaticity Gap suggests a large number of frequent multi-word expressions or formulaic sequences that are statistically distinct from their component parts.
541
 
542
  ---
543
  ## 7. Summary & Recommendations
 
549
  | Component | Recommended | Rationale |
550
  |-----------|-------------|-----------|
551
  | Tokenizer | **64k BPE** | Best compression (4.19x) |
552
+ | N-gram | **5-gram** | Lowest perplexity (198) |
553
  | Markov | **Context-4** | Highest predictability (99.0%) |
554
  | Embeddings | **100d** | Balanced semantic capture and isotropy |
555
 
 
764
  ---
765
  *Generated by Wikilangs Models Pipeline*
766
 
767
+ *Report Date: 2026-01-03 16:16:58*
models/embeddings/aligned/ab_128d.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71c42722da6595336ced095cc89ab98dce46aa2764b97c56a059a072aa38c40f
3
+ size 1036432309
models/embeddings/aligned/ab_128d.meta.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lang": "ab", "dim": 128, "max_seq_len": 512, "is_aligned": true}
models/embeddings/aligned/ab_128d.projection.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6554a5b2123562574512b96e8803301b6bd19ee23ecae768b30b02038949093
3
+ size 65664
models/embeddings/aligned/ab_128d_metadata.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "language": "ab",
3
+ "dimension": 128,
4
+ "version": "aligned",
5
+ "hub_language": "en",
6
+ "seed_vocab_size": 927,
7
+ "vocab_size": 11860
8
+ }
models/embeddings/aligned/ab_32d.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26e75c47cb2d2d69347eb0db2b7206527bdfe9a74fd1254d6e81ebb46b8e86f4
3
+ size 259323829
models/embeddings/aligned/ab_32d.meta.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lang": "ab", "dim": 32, "max_seq_len": 512, "is_aligned": true}
models/embeddings/aligned/ab_32d.projection.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61563ea00eba259124ac0cd787d4e2f7845d51497b14dbe427b37e79e0366086
3
+ size 4224
models/embeddings/aligned/ab_32d_metadata.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "language": "ab",
3
+ "dimension": 32,
4
+ "version": "aligned",
5
+ "hub_language": "en",
6
+ "seed_vocab_size": 927,
7
+ "vocab_size": 11860
8
+ }
models/embeddings/aligned/ab_64d.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24b083b8f182ed5799370c8e2ae70954e4785710342ad0ebd087e7b12310f7f4
3
+ size 518359989
models/embeddings/aligned/ab_64d.meta.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lang": "ab", "dim": 64, "max_seq_len": 512, "is_aligned": true}
models/embeddings/aligned/ab_64d.projection.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf87c42f43dc72a3c48c49eb5a683b1a5c25149e7ab05d83bee2dba66444b19b
3
+ size 16512
models/embeddings/aligned/ab_64d_metadata.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "language": "ab",
3
+ "dimension": 64,
4
+ "version": "aligned",
5
+ "hub_language": "en",
6
+ "seed_vocab_size": 927,
7
+ "vocab_size": 11860
8
+ }
models/embeddings/monolingual/ab_128d.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:974accbf5e248437084213db34906d889d1387be4ba287f5bf3decc278094a76
3
- size 1036399852
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71c42722da6595336ced095cc89ab98dce46aa2764b97c56a059a072aa38c40f
3
+ size 1036432309
models/embeddings/monolingual/ab_128d_metadata.json CHANGED
@@ -11,5 +11,5 @@
11
  "encoding_method": "rope",
12
  "dim": 128
13
  },
14
- "vocab_size": 11829
15
  }
 
11
  "encoding_method": "rope",
12
  "dim": 128
13
  },
14
+ "vocab_size": 11860
15
  }
models/embeddings/monolingual/ab_32d.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:20cc29bc8688764f66dd79168f5b38b9bc2c37d4cff74832ca2676787afff53e
3
- size 259315180
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26e75c47cb2d2d69347eb0db2b7206527bdfe9a74fd1254d6e81ebb46b8e86f4
3
+ size 259323829
models/embeddings/monolingual/ab_32d_metadata.json CHANGED
@@ -11,5 +11,5 @@
11
  "encoding_method": "rope",
12
  "dim": 32
13
  },
14
- "vocab_size": 11829
15
  }
 
11
  "encoding_method": "rope",
12
  "dim": 32
13
  },
14
+ "vocab_size": 11860
15
  }
models/embeddings/monolingual/ab_64d.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f6e336f851f9509b219587c555a83bc96b5e5f8af5ceca5f7c676a4c11e0eec7
3
- size 518343404
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24b083b8f182ed5799370c8e2ae70954e4785710342ad0ebd087e7b12310f7f4
3
+ size 518359989
models/embeddings/monolingual/ab_64d_metadata.json CHANGED
@@ -11,5 +11,5 @@
11
  "encoding_method": "rope",
12
  "dim": 64
13
  },
14
- "vocab_size": 11829
15
  }
 
11
  "encoding_method": "rope",
12
  "dim": 64
13
  },
14
+ "vocab_size": 11860
15
  }
models/subword_markov/ab_markov_ctx1_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:08d81996caede4f16305dff37bfaf00cc4aa080b329a05b6e813e43b32d4b5b1
3
- size 74300
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19cd311913da52a816b4226234ba5047af2e1d508ac9c65583ea9a88645432ac
3
+ size 74098
models/subword_markov/ab_markov_ctx1_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 1,
3
  "variant": "subword",
4
  "language": "ab",
5
- "unique_contexts": 872,
6
- "total_transitions": 4086915
7
  }
 
2
  "context_size": 1,
3
  "variant": "subword",
4
  "language": "ab",
5
+ "unique_contexts": 879,
6
+ "total_transitions": 4093303
7
  }
models/subword_markov/ab_markov_ctx2_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:391adb8ed94471fe378fd39fa4bafc27bc22517c04a0688782555f658bfa5660
3
- size 430405
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:927b982ad94c14844b6d7630912a44f97b2ed837ecc6612f0334f1f20f72c9e3
3
+ size 431394
models/subword_markov/ab_markov_ctx2_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 2,
3
  "variant": "subword",
4
  "language": "ab",
5
- "unique_contexts": 9435,
6
- "total_transitions": 4080892
7
  }
 
2
  "context_size": 2,
3
  "variant": "subword",
4
  "language": "ab",
5
+ "unique_contexts": 9477,
6
+ "total_transitions": 4087276
7
  }
models/subword_markov/ab_markov_ctx3_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2554570f44dad84400a3a91e792ec596f90a065c3a673af002a820c70dca44e9
3
- size 1518240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f54ce1a72b48d81d104298015c3c2deb43f8272a78af7071406a6e3a7b786f0
3
+ size 1525586
models/subword_markov/ab_markov_ctx3_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 3,
3
  "variant": "subword",
4
  "language": "ab",
5
- "unique_contexts": 56063,
6
- "total_transitions": 4074869
7
  }
 
2
  "context_size": 3,
3
  "variant": "subword",
4
  "language": "ab",
5
+ "unique_contexts": 56288,
6
+ "total_transitions": 4081249
7
  }
models/subword_markov/ab_markov_ctx4_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a36f4483bb607e4182b1e8a367258cf1570a36ccf6debdfa0c5ed6d83eb6baa9
3
- size 4368269
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96368e0437f26828beb02a17e213d6db03b94d43cfc0d2d0a1f95b152a4198e8
3
+ size 4373446
models/subword_markov/ab_markov_ctx4_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 4,
3
  "variant": "subword",
4
  "language": "ab",
5
- "unique_contexts": 206852,
6
- "total_transitions": 4068846
7
  }
 
2
  "context_size": 4,
3
  "variant": "subword",
4
  "language": "ab",
5
+ "unique_contexts": 207598,
6
+ "total_transitions": 4075222
7
  }
models/subword_ngram/ab_2gram_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6546e93ab975088eb711841de0aac0eef95cedd64a0648eef8735695ed55ff88
3
- size 54917
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a1ed758e72646135180f530b31bbda049e7ce468f83f1559ec140a8ce2d91db
3
+ size 54947
models/subword_ngram/ab_2gram_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "n": 2,
3
  "variant": "subword",
4
  "language": "ab",
5
- "unique_ngrams": 4104,
6
- "total_ngrams": 4086915
7
  }
 
2
  "n": 2,
3
  "variant": "subword",
4
  "language": "ab",
5
+ "unique_ngrams": 4117,
6
+ "total_ngrams": 4093303
7
  }
models/subword_ngram/ab_3gram_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7b1b02523b1b148c8a70500352648cf3a5210acb9cff8bd3b8c613fbe55e72ef
3
- size 368926
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:25e278dc8f9634827e842fd4fa0a7ab1634032e48a71621686047e4cd1ca2c00
3
+ size 369086
models/subword_ngram/ab_3gram_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "n": 3,
3
  "variant": "subword",
4
  "language": "ab",
5
- "unique_ngrams": 28199,
6
- "total_ngrams": 4080892
7
  }
 
2
  "n": 3,
3
  "variant": "subword",
4
  "language": "ab",
5
+ "unique_ngrams": 28284,
6
+ "total_ngrams": 4087276
7
  }
models/subword_ngram/ab_4gram_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d7eebb21889eb39845863d08c2bb3db59acb306cc610d9cff31719d1fccb8a9f
3
- size 1420721
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:665986334ceeda8aa48fbc99ac922f66ba2fa5fe34df93b4d6ba07fc1f9d5d48
3
+ size 1427310
models/subword_ngram/ab_4gram_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "n": 4,
3
  "variant": "subword",
4
  "language": "ab",
5
- "unique_ngrams": 112541,
6
- "total_ngrams": 4074869
7
  }
 
2
  "n": 4,
3
  "variant": "subword",
4
  "language": "ab",
5
+ "unique_ngrams": 112814,
6
+ "total_ngrams": 4081249
7
  }
models/subword_ngram/ab_5gram_subword.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:839b1b7f710bb721be0571bcffc69c450506f2c7de4a3cede70a0d8f89d9a2c5
3
+ size 2860189
models/subword_ngram/ab_5gram_subword_metadata.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "n": 5,
3
+ "variant": "subword",
4
+ "language": "ab",
5
+ "unique_ngrams": 211528,
6
+ "total_ngrams": 4075222
7
+ }
models/tokenizer/ab_tokenizer_16k.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5a2085566fc5290d88287b8216db2b990d022c905f7619c834e7a919406ca669
3
- size 569009
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:faab02fa0c6d322c4b29526ae72183b3c6bdefcdd1dde3661ce29441818b789e
3
+ size 569291
models/tokenizer/ab_tokenizer_16k.vocab CHANGED
The diff for this file is too large to render. See raw diff
 
models/tokenizer/ab_tokenizer_32k.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1e4a3fa167aa62fd56d1a517f1285f051a7587310895cdc82a80895e1bd8badc
3
- size 936940
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6373f1b123f84fb1dc1394e345603a0a661907f056224dd29e84248278023e8e
3
+ size 936901
models/tokenizer/ab_tokenizer_32k.vocab CHANGED
The diff for this file is too large to render. See raw diff
 
models/tokenizer/ab_tokenizer_64k.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:09581380da9729bf63f3885bf0e54ee06e9b0659cd517ad67eab83cf01150542
3
- size 1690301
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33980fec6cedaeeab4a3b516f2447c745616b229192936ac6f9bcc71b8c92cd0
3
+ size 1690475
models/tokenizer/ab_tokenizer_64k.vocab CHANGED
The diff for this file is too large to render. See raw diff
 
models/tokenizer/ab_tokenizer_8k.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fb1da49fc5cbf61e37585d974f3bcb0262f7b8e403c201974781b65a1079fa28
3
- size 402830
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c621351e7846d8e21f2dc64c9034dd95cd0718a1712a0b7ed4b16694ad52bf38
3
+ size 403151
models/tokenizer/ab_tokenizer_8k.vocab CHANGED
The diff for this file is too large to render. See raw diff
 
models/vocabulary/ab_vocabulary.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1b9954423f516ffad7b6b395410e0a3ed33a13656770c69fcd1e37b9a2500496
3
- size 613460
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96921566daf39bc2fc2c78e494582b420e246beb86cef2d2868dd655be330fad
3
+ size 618434
models/vocabulary/ab_vocabulary_metadata.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "language": "ab",
3
- "vocabulary_size": 32686,
4
  "variant": "full",
5
  "statistics": {
6
- "type_token_ratio": 0.18190225895758938,
7
  "coverage": {
8
- "top_100": 0.26752598001845684,
9
- "top_1000": 0.4927917987401196,
10
- "top_5000": 0.6798940737471412,
11
- "top_10000": 0.7575291899049071
12
  },
13
- "hapax_count": 57985,
14
- "hapax_ratio": 0.639509876366203,
15
- "total_documents": 6023
16
  }
17
  }
 
1
  {
2
  "language": "ab",
3
+ "vocabulary_size": 32744,
4
  "variant": "full",
5
  "statistics": {
6
+ "type_token_ratio": 0.18202523582212732,
7
  "coverage": {
8
+ "top_100": 0.26734987810765387,
9
+ "top_1000": 0.4923108665474118,
10
+ "top_5000": 0.679398090186314,
11
+ "top_10000": 0.7571407681321125
12
  },
13
+ "hapax_count": 58125,
14
+ "hapax_ratio": 0.6396570887761502,
15
+ "total_documents": 6027
16
  }
17
  }
models/word_markov/ab_markov_ctx1_word.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:477c26b2510447d28ecaa37e0e7ab40a517773cfffb1210fbde670a44914bc00
3
- size 4536676
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2bf7260c71b0507ac13f5ca76a0540969736af005153e3ea30a842a8fbc49138
3
+ size 4532845
models/word_markov/ab_markov_ctx1_word_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 1,
3
  "variant": "word",
4
  "language": "ab",
5
- "unique_contexts": 90583,
6
- "total_transitions": 492437
7
  }
 
2
  "context_size": 1,
3
  "variant": "word",
4
  "language": "ab",
5
+ "unique_contexts": 90782,
6
+ "total_transitions": 493184
7
  }
models/word_markov/ab_markov_ctx2_word.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b4bc4862d6258993f1f724d0d5118dc5b0d53b173d1761b40efd9fab5f483159
3
- size 10110358
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a16ee7a61ea5b043c3495cdd5ddaadb1eaab15dfa5e1487cdeaa37c6f45c0fd0
3
+ size 10123960
models/word_markov/ab_markov_ctx2_word_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 2,
3
  "variant": "word",
4
  "language": "ab",
5
- "unique_contexts": 326705,
6
- "total_transitions": 486414
7
  }
 
2
  "context_size": 2,
3
  "variant": "word",
4
  "language": "ab",
5
+ "unique_contexts": 327437,
6
+ "total_transitions": 487157
7
  }