omarkamali commited on
Commit
3d0ac29
·
verified ·
1 Parent(s): 8164949

Upload all models and assets for am (latest)

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +1 -0
  2. README.md +185 -148
  3. models/embeddings/aligned/am_128d.bin +3 -0
  4. models/embeddings/aligned/am_128d.meta.json +1 -0
  5. models/embeddings/aligned/am_128d.projection.npy +3 -0
  6. models/embeddings/aligned/am_128d_metadata.json +8 -0
  7. models/embeddings/aligned/am_32d.bin +3 -0
  8. models/embeddings/aligned/am_32d.meta.json +1 -0
  9. models/embeddings/aligned/am_32d.projection.npy +3 -0
  10. models/embeddings/aligned/am_32d_metadata.json +8 -0
  11. models/embeddings/aligned/am_64d.bin +3 -0
  12. models/embeddings/aligned/am_64d.meta.json +1 -0
  13. models/embeddings/aligned/am_64d.projection.npy +3 -0
  14. models/embeddings/aligned/am_64d_metadata.json +8 -0
  15. models/embeddings/monolingual/am_128d.bin +2 -2
  16. models/embeddings/monolingual/am_128d_metadata.json +1 -1
  17. models/embeddings/monolingual/am_32d.bin +2 -2
  18. models/embeddings/monolingual/am_32d_metadata.json +1 -1
  19. models/embeddings/monolingual/am_64d.bin +2 -2
  20. models/embeddings/monolingual/am_64d_metadata.json +1 -1
  21. models/subword_markov/am_markov_ctx1_subword.parquet +2 -2
  22. models/subword_markov/am_markov_ctx1_subword_metadata.json +2 -2
  23. models/subword_markov/am_markov_ctx2_subword.parquet +2 -2
  24. models/subword_markov/am_markov_ctx2_subword_metadata.json +2 -2
  25. models/subword_markov/am_markov_ctx3_subword.parquet +2 -2
  26. models/subword_markov/am_markov_ctx3_subword_metadata.json +2 -2
  27. models/subword_markov/am_markov_ctx4_subword.parquet +2 -2
  28. models/subword_markov/am_markov_ctx4_subword_metadata.json +2 -2
  29. models/subword_ngram/am_2gram_subword.parquet +2 -2
  30. models/subword_ngram/am_2gram_subword_metadata.json +2 -2
  31. models/subword_ngram/am_3gram_subword.parquet +2 -2
  32. models/subword_ngram/am_3gram_subword_metadata.json +2 -2
  33. models/subword_ngram/am_4gram_subword.parquet +2 -2
  34. models/subword_ngram/am_4gram_subword_metadata.json +2 -2
  35. models/subword_ngram/am_5gram_subword.parquet +3 -0
  36. models/subword_ngram/am_5gram_subword_metadata.json +7 -0
  37. models/tokenizer/am_tokenizer_16k.model +2 -2
  38. models/tokenizer/am_tokenizer_16k.vocab +0 -0
  39. models/tokenizer/am_tokenizer_32k.model +2 -2
  40. models/tokenizer/am_tokenizer_32k.vocab +0 -0
  41. models/tokenizer/am_tokenizer_64k.model +2 -2
  42. models/tokenizer/am_tokenizer_64k.vocab +0 -0
  43. models/tokenizer/am_tokenizer_8k.model +2 -2
  44. models/tokenizer/am_tokenizer_8k.vocab +0 -0
  45. models/vocabulary/am_vocabulary.parquet +2 -2
  46. models/vocabulary/am_vocabulary_metadata.json +8 -8
  47. models/word_markov/am_markov_ctx1_word.parquet +2 -2
  48. models/word_markov/am_markov_ctx1_word_metadata.json +2 -2
  49. models/word_markov/am_markov_ctx2_word.parquet +2 -2
  50. models/word_markov/am_markov_ctx2_word_metadata.json +2 -2
.gitattributes CHANGED
@@ -39,3 +39,4 @@ visualizations/position_encoding_comparison.png filter=lfs diff=lfs merge=lfs -t
39
  visualizations/tsne_sentences.png filter=lfs diff=lfs merge=lfs -text
40
  visualizations/tsne_words.png filter=lfs diff=lfs merge=lfs -text
41
  visualizations/zipf_law.png filter=lfs diff=lfs merge=lfs -text
 
 
39
  visualizations/tsne_sentences.png filter=lfs diff=lfs merge=lfs -text
40
  visualizations/tsne_words.png filter=lfs diff=lfs merge=lfs -text
41
  visualizations/zipf_law.png filter=lfs diff=lfs merge=lfs -text
42
+ visualizations/embedding_tsne_multilingual.png filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  language: am
3
- language_name: AM
4
  language_family: semitic_ethiopic
5
  tags:
6
  - wikilangs
@@ -10,11 +10,21 @@ tags:
10
  - n-gram
11
  - markov
12
  - wikipedia
 
 
 
 
 
 
 
 
 
 
13
  - monolingual
14
  - family-semitic_ethiopic
15
  license: mit
16
  library_name: wikilangs
17
- pipeline_tag: feature-extraction
18
  datasets:
19
  - omarkamali/wikipedia-monthly
20
  dataset_info:
@@ -23,20 +33,20 @@ dataset_info:
23
  metrics:
24
  - name: best_compression_ratio
25
  type: compression
26
- value: 3.287
27
  - name: best_isotropy
28
  type: isotropy
29
- value: 0.9163
30
  - name: vocabulary_size
31
  type: vocab
32
  value: 0
33
  generated: 2026-01-03
34
  ---
35
 
36
- # AM - Wikilangs Models
37
  ## Comprehensive Research Report & Full Ablation Study
38
 
39
- This repository contains NLP models trained and evaluated by Wikilangs, specifically on **AM** Wikipedia data.
40
  We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and word embeddings.
41
 
42
  ## 📋 Repository Contents
@@ -60,7 +70,7 @@ We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and
60
  - [3. Markov Chain Evaluation](#3-markov-chain-evaluation)
61
  - [4. Vocabulary Analysis](#4-vocabulary-analysis)
62
  - [5. Word Embeddings Evaluation](#5-word-embeddings-evaluation)
63
- - [6. Morphological Analysis (Experimental)](#6-morphological-analysis)
64
  - [7. Summary & Recommendations](#7-summary--recommendations)
65
  - [Metrics Glossary](#appendix-metrics-glossary--interpretation-guide)
66
  - [Visualizations Index](#visualizations-index)
@@ -80,47 +90,47 @@ We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and
80
 
81
  | Vocab Size | Compression | Avg Token Len | UNK Rate | Total Tokens |
82
  |------------|-------------|---------------|----------|--------------|
83
- | **8k** | 2.436x | 2.44 | 0.1557% | 683,952 |
84
- | **16k** | 2.745x | 2.75 | 0.1754% | 607,060 |
85
- | **32k** | 3.031x | 3.03 | 0.1937% | 549,802 |
86
- | **64k** | 3.287x 🏆 | 3.29 | 0.2101% | 506,938 |
87
 
88
  ### Tokenization Examples
89
 
90
  Below are sample sentences tokenized with each vocabulary size:
91
 
92
- **Sample 1:** `አዋሳ ከነማ ስታዲየም በአዋሳ፣ ኢትዮጵያ የሚገኝ ስታዲዮም ነው። ፳፭ ሺህ ሰዎችን መያዝ ሲችል የአዋሳ ከተማ የእግር ኳስ ክለብ...`
93
 
94
  | Vocab | Tokens | Count |
95
  |-------|--------|-------|
96
- | 8k | `▁አዋ ▁ከነ ▁ስታዲየም ▁በአ ሳ፣ ▁ኢትዮጵያ ▁የሚገኝ ... (+25 more)` | 35 |
97
- | 16k | `▁አዋሳ ▁ከነ ▁ስታዲየም ▁በአ ሳ፣ ▁ኢትዮጵያ ▁የሚገኝ ▁ስታ ... (+22 more)` | 32 |
98
- | 32k | `▁አዋሳ ▁ከነማ ▁ስታዲየም ▁በአ ሳ፣ ▁ኢትዮጵያ ▁የሚገኝ ▁ስታ ዲዮ ... (+20 more)` | 30 |
99
- | 64k | `▁አዋሳ ▁ከነማ ▁ስታዲየም ▁በአዋ ሳ፣ ▁ኢትዮጵያ ▁የሚገኝ ▁ስታ ዲዮ ... (+19 more)` | 29 |
100
 
101
- **Sample 2:** `የዝንጀሮ ስብሰባ በውሻ ጩኸት ይበተናል የአማርኛ ምሳሌ ነው። የዝንጀሮ ስብሰባ በውሻ ጩኸት ይበተናል የአማርኛ ምሳሌ ነው። ትር...`
102
 
103
  | Vocab | Tokens | Count |
104
  |-------|--------|-------|
105
- | 8k | `▁የዝ ንጀሮ ▁ስብሰባ ▁በው ▁ይበ ... (+29 more)` | 39 |
106
- | 16k | `▁የዝ ንጀሮ ▁ስብሰባ ▁በው ▁ጩ ኸት ▁ይበ ናል ... (+25 more)` | 35 |
107
- | 32k | `▁የዝንጀሮ ▁ስብሰባ ▁በው ▁ጩ ኸት ▁ይበ ተናል ▁የ���ማርኛ ▁ምሳሌ ... (+21 more)` | 31 |
108
- | 64k | `▁የዝንጀሮ ▁ስብሰባ ▁በውሻ ▁ጩኸት ▁ይበ ተናል ▁የአማርኛ ▁ምሳሌ ▁ነው። ▁የዝንጀሮ ... (+17 more)` | 27 |
109
 
110
- **Sample 3:** `የሐረሪ ብሔራዊ ሊግ የኢትዮጵያ ፖለቲካ ፓርቲ ነው። ዓላማ ሊቀመንበር ታሪክ መደብ: በምርጫ የተሳተፉ የኢትዮጵያ ፓርቲዎች መደብ...`
111
 
112
  | Vocab | Tokens | Count |
113
  |-------|--------|-------|
114
- | 8k | `▁የሐ ▁ብሔራዊ ▁ሊግ ▁የኢትዮጵያ ▁ፖለቲካ ▁ፓርቲ ▁ነው። ▁ዓላማ ... (+13 more)` | 23 |
115
- | 16k | `▁የሐ ረሪ ▁ብሔራዊ ▁ሊግ ▁የኢትዮጵያ ▁ፖለቲካ ▁ፓርቲ ▁ነው። ▁ዓላማ ▁ሊቀመንበር ... (+12 more)` | 22 |
116
- | 32k | `▁የሐረሪ ▁ብሔራዊ ▁ሊግ ▁የኢትዮጵያ ▁ፖለቲካ ▁ፓርቲ ▁ነው። ▁ዓላማ ▁ሊቀመንበር ▁ታሪክ ... (+11 more)` | 21 |
117
- | 64k | `▁የሐረሪ ▁ብሔራዊ ▁ሊግ ▁የኢትዮጵያ ▁ፖለቲካ ▁ፓርቲ ▁ነው። ▁ዓላማ ▁ሊቀመንበር ▁ታሪክ ... (+11 more)` | 21 |
118
 
119
 
120
  ### Key Findings
121
 
122
- - **Best Compression:** 64k achieves 3.287x compression
123
- - **Lowest UNK Rate:** 8k with 0.1557% unknown tokens
124
  - **Trade-off:** Larger vocabularies improve compression but increase model size
125
  - **Recommendation:** 32k vocabulary provides optimal balance for production use
126
 
@@ -137,12 +147,14 @@ Below are sample sentences tokenized with each vocabulary size:
137
 
138
  | N-gram | Variant | Perplexity | Entropy | Unique N-grams | Top-100 Coverage | Top-1000 Coverage |
139
  |--------|---------|------------|---------|----------------|------------------|-------------------|
140
- | **2-gram** | Word | 8,988 | 13.13 | 27,901 | 19.7% | 39.7% |
141
- | **2-gram** | Subword | 2,079 🏆 | 11.02 | 23,804 | 34.0% | 69.2% |
142
- | **3-gram** | Word | 9,944 | 13.28 | 35,714 | 22.1% | 40.5% |
143
- | **3-gram** | Subword | 19,139 | 14.22 | 153,027 | 11.8% | 35.5% |
144
- | **4-gram** | Word | 36,744 | 15.17 | 90,792 | 13.9% | 25.8% |
145
- | **4-gram** | Subword | 94,777 | 16.53 | 549,996 | 6.6% | 19.5% |
 
 
146
 
147
  ### Top 5 N-grams by Size
148
 
@@ -150,68 +162,88 @@ Below are sample sentences tokenized with each vocabulary size:
150
 
151
  | Rank | N-gram | Count |
152
  |------|--------|-------|
153
- | 1 | `ዓ ም` | 8,324 |
154
- | 2 | `ምሳሌ ነው` | 5,625 |
155
- | 3 | `የአማርኛ ምሳሌ` | 5,563 |
156
- | 4 | `እ ኤ` | 4,026 |
157
- | 5 | `ኤ አ` | 3,961 |
158
 
159
  **3-grams (Word):**
160
 
161
  | Rank | N-gram | Count |
162
  |------|--------|-------|
163
- | 1 | `የአማርኛ ምሳሌ ነው` | 5,563 |
164
- | 2 | `እ ኤ አ` | 3,908 |
165
  | 3 | `ምሳሌ ነው ትርጉሙ` | 3,454 |
166
  | 4 | `መደብ ተረትና ምሳሌ` | 3,051 |
167
- | 5 | `ነ�� ትርጉሙ መደብ` | 2,533 |
168
 
169
  **4-grams (Word):**
170
 
171
  | Rank | N-gram | Count |
172
  |------|--------|-------|
173
  | 1 | `የአማርኛ ምሳሌ ነው ትርጉሙ` | 3,452 |
174
- | 2 | `ምሳሌ ነው ትርጉሙ መደብ` | 2,533 |
175
- | 3 | `ትርጉሙ መደብ ያልተተረጎመ ምሳሌ` | 2,118 |
176
- | 4 | `ነው ትርጉሙ መደብ ያልተተረጎመ` | 2,114 |
177
  | 5 | `ምሳሌ መደብ ተረትና ምሳሌ` | 1,854 |
178
 
 
 
 
 
 
 
 
 
 
 
179
  **2-grams (Subword):**
180
 
181
  | Rank | N-gram | Count |
182
  |------|--------|-------|
183
- | 1 | `_ የ` | 170,716 |
184
- | 2 | `ት _` | 145,051 |
185
- | 3 | `_ በ` | 140,839 |
186
- | 4 | `ን _` | 132,909 |
187
- | 5 | `_ አ` | 113,769 |
188
 
189
  **3-grams (Subword):**
190
 
191
  | Rank | N-gram | Count |
192
  |------|--------|-------|
193
- | 1 | `_ እ ን` | 32,319 |
194
- | 2 | `_ ነ ው` | 26,511 |
195
- | 3 | `ው _` | 24,155 |
196
- | 4 | `_ ና` | 23,843 |
197
- | 5 | `እ ና _` | 22,397 |
198
 
199
  **4-grams (Subword):**
200
 
201
  | Rank | N-gram | Count |
202
  |------|--------|-------|
203
- | 1 | `_ እ ና _` | 22,267 |
204
- | 2 | `_ ነ ው ።` | 19,378 |
205
- | 3 | `ነ ው ። _` | 18,922 |
206
- | 4 | `_ እ ን ደ` | 13,836 |
207
- | 5 | `_ ላ ይ _` | 12,924 |
 
 
 
 
 
 
 
 
 
 
208
 
209
 
210
  ### Key Findings
211
 
212
- - **Best Perplexity:** 2-gram (subword) with 2,079
213
  - **Entropy Trend:** Decreases with larger n-grams (more predictable)
214
- - **Coverage:** Top-1000 patterns cover ~19% of corpus
215
  - **Recommendation:** 4-gram or 5-gram for best predictive performance
216
 
217
  ---
@@ -227,14 +259,14 @@ Below are sample sentences tokenized with each vocabulary size:
227
 
228
  | Context | Variant | Avg Entropy | Perplexity | Branching Factor | Unique Contexts | Predictability |
229
  |---------|---------|-------------|------------|------------------|-----------------|----------------|
230
- | **1** | Word | 0.7502 | 1.682 | 4.80 | 236,353 | 25.0% |
231
- | **1** | Subword | 1.2235 | 2.335 | 17.52 | 2,854 | 0.0% |
232
- | **2** | Word | 0.1468 | 1.107 | 1.28 | 1,130,961 | 85.3% |
233
- | **2** | Subword | 1.0397 | 2.056 | 6.98 | 49,981 | 0.0% |
234
- | **3** | Word | 0.0355 | 1.025 | 1.06 | 1,446,616 | 96.4% |
235
- | **3** | Subword | 0.6354 | 1.553 | 3.36 | 348,535 | 36.5% |
236
- | **4** | Word | 0.0159 🏆 | 1.011 | 1.02 | 1,520,994 | 98.4% |
237
- | **4** | Subword | 0.4515 | 1.367 | 2.14 | 1,171,344 | 54.9% |
238
 
239
  ### Generated Text Samples (Word-based)
240
 
@@ -242,27 +274,27 @@ Below are text samples generated from each word-based Markov chain model:
242
 
243
  **Context Size 1:**
244
 
245
- 1. `ነው በጥሩ አስተዳደር የህዝብ ልውውጥ ኮሚሽን ሕንፃ በሥነ ሕንጻ ተጠናቆ ክፍት አረግንጓዴ ከጂዮርጂያ በእ የሆነ`
246
- 2. `እና ሲሞት ቤተሰቦቹ ጋር ቢኮብለል ወይም ቀበሮ ዝርያ ያለበት ቦታ 4 5 14 847 ቅጥር የተለጠፈ`
247
- 3. `ላይ መስኮት እና ከደቡብ ህንድ ቀጥሎ ቀዳማዊ ኃይለ ሥላሴ ለአፍሪካ ገሞጂማ ሣርማ ቦታዎች ይቀመጡ`
248
 
249
  **Context Size 2:**
250
 
251
- 1. `ዓ ም አስቀድሞ ወይም የአርጡስ ወንድም 4 ካርል 663 669 ዓ ም ሁላቸው ሲስማሙ ወደ`
252
- 2. `ምሳሌ ነው ጀርባዬን እከክልኝ ለኔ ራቀኝ የአማርኛ ምሳሌ ነው ዝርክርክ ከወንፊት የባሰ ዝክዝክ የአማርኛ ምሳሌ ነው ትርጉሙ`
253
- 3. `የአማርኛ ምሳሌ ነው የምትጠላው ሰው ፈሱ እሆዱ ውስጥ ሳለ ኔሽን ኦፍ ኢስላም ጋር ያለው ዝምድና ግልጽ ነው`
254
 
255
  **Context Size 3:**
256
 
257
- 1. `የአማርኛ ምሳሌ ነው ትርጉሙ መደብ ያልተተረጎመ ምሳሌ መደብ ተረትና ምሳሌ ቁና ሰፋች`
258
- 2. `እ ኤ አ በሂትለር ተጽዕኖ ሙሶሎኒ በጣሊያን ፀረ ሴማዊ የዘር ህጎች እንዲፀድቁ ደገፈ በመጋቢት ጀርመን ቼኮዝሎቫኪያን ከቀላቀለች`
259
- 3. `ምሳሌ ነው ትርጉሙ መደብ ያልተተረጎመ ምሳሌ መደብ ተረትና ምሳሌ ምግባር ሳይኖር ስም እንደማለት ነዉ`
260
 
261
  **Context Size 4:**
262
 
263
- 1. `የአማርኛ ምሳሌ ነው ትርጉሙ ሁለቱም አያዋጡም መደብ ተረትና ምሳሌ`
264
- 2. `ምሳሌ ነው ትርጉሙ መደብ ያልተተረጎመ ምሳሌ መደብ ተረትና ምሳሌ መደብ ፈሊጣዊ አነጋገር መደብ ተረትና ምሳሌ ቁና ሰፋች`
265
- 3. `ነው ትርጉሙ መደብ ያልተተረጎመ ምሳሌ መደብ ተረትና ምሳሌ ሴት ሁሉን ቻይ ናት`
266
 
267
 
268
  ### Generated Text Samples (Subword-based)
@@ -271,34 +303,34 @@ Below are text samples generated from each subword-based Markov chain model:
271
 
272
  **Context Size 1:**
273
 
274
- 1. `_አንግብ፣_ለማር_(“ሀንን`
275
- 2. `ን_(dicole_ገደቡድ_ሞ`
276
- 3. `ት__ሓምበላስድ__ይለት`
277
 
278
  **Context Size 2:**
279
 
280
- 1. `_የኢትዮጵያ_አፖሎኛ_,00_`
281
- 2. `ት_ተመለሰብን_ኣሉ።_የባህር`
282
- 3. `_በዘመዴ_ሲፀድቅ_በተአምስተ`
283
 
284
  **Context Size 3:**
285
 
286
- 1. `_እንደ_ማርኮ_ከተማ_እና_ግሮ`
287
- 2. `_ነው።_ከተማው_ባልሞራል_እን`
288
- 3. `ው።_ኮምፕዩተራይዝ_ካሊፎርኒያ`
289
 
290
  **Context Size 4:**
291
 
292
- 1. `_እና_ማከማቸት_ጉዳት_ቁጭ_ብለ`
293
- 2. `_ነው።_ዋጋው_ወቅት_የመጀመሪያ`
294
- 3. `ነው።_ትርጉሙ_አንቴና_ይፈሳሉ፡`
295
 
296
 
297
  ### Key Findings
298
 
299
  - **Best Predictability:** Context-4 (word) with 98.4% predictability
300
  - **Branching Factor:** Decreases with context size (more deterministic)
301
- - **Memory Trade-off:** Larger contexts require more storage (1,171,344 contexts)
302
  - **Recommendation:** Context-3 or Context-4 for text generation
303
 
304
  ---
@@ -314,48 +346,48 @@ Below are text samples generated from each subword-based Markov chain model:
314
 
315
  | Metric | Value |
316
  |--------|-------|
317
- | Vocabulary Size | 99,716 |
318
- | Total Tokens | 1,636,892 |
319
- | Mean Frequency | 16.42 |
320
  | Median Frequency | 3 |
321
- | Frequency Std Dev | 174.41 |
322
 
323
  ### Most Common Words
324
 
325
  | Rank | Word | Frequency |
326
  |------|------|-----------|
327
- | 1 | ነው | 26,460 |
328
- | 2 | እና | 22,392 |
329
- | 3 | ላይ | 13,250 |
330
- | 4 | ምሳሌ | 11,607 |
331
- | 5 | ውስጥ | 9,622 |
332
- | 6 | ነበር | 9,005 |
333
- | 7 | ዓ | 8,679 |
334
- | 8 | | 8,584 |
335
- | 9 | ወደ | 8,446 |
336
- | 10 | እንደ | 6,776 |
337
 
338
  ### Least Common Words (from vocabulary)
339
 
340
  | Rank | Word | Frequency |
341
  |------|------|-----------|
342
- | 1 | ቫሊን | 2 |
343
- | 2 | ግሎቡላር | 2 |
344
- | 3 | ኢንዛይሞች | 2 |
345
- | 4 | የማከማቻ | 2 |
346
- | 5 | ለph | 2 |
347
- | 6 | ግብረመልሶችን | 2 |
348
- | 7 | behi | 2 |
349
- | 8 | ቤሂ | 2 |
350
- | 9 | goli | 2 |
351
- | 10 | ክሩድስ | 2 |
352
 
353
  ### Zipf's Law Analysis
354
 
355
  | Metric | Value |
356
  |--------|-------|
357
- | Zipf Coefficient | 0.9367 |
358
- | R² (Goodness of Fit) | 0.995214 |
359
  | Adherence Quality | **excellent** |
360
 
361
  ### Coverage Analysis
@@ -371,7 +403,7 @@ Below are text samples generated from each subword-based Markov chain model:
371
 
372
  - **Zipf Compliance:** R²=0.9952 indicates excellent adherence to Zipf's law
373
  - **High Frequency Dominance:** Top 100 words cover 22.7% of corpus
374
- - **Long Tail:** 89,716 words needed for remaining 25.1% coverage
375
 
376
  ---
377
  ## 5. Word Embeddings Evaluation
@@ -387,37 +419,40 @@ Below are text samples generated from each subword-based Markov chain model:
387
 
388
  ### 5.1 Cross-Lingual Alignment
389
 
390
- > *Note: Multilingual alignment visualization not available for this language.*
 
 
391
 
392
 
393
  ### 5.2 Model Comparison
394
 
395
  | Model | Dimension | Isotropy | Semantic Density | Alignment R@1 | Alignment R@10 |
396
  |-------|-----------|----------|------------------|---------------|----------------|
397
- | **mono_32d** | 32 | 0.9125 | 0.3250 | N/A | N/A |
398
- | **mono_64d** | 64 | 0.9163 🏆 | 0.2292 | N/A | N/A |
399
- | **mono_128d** | 128 | 0.8535 | 0.1745 | N/A | N/A |
 
 
 
400
 
401
  ### Key Findings
402
 
403
- - **Best Isotropy:** mono_64d with 0.9163 (more uniform distribution)
404
- - **Semantic Density:** Average pairwise similarity of 0.2429. Lower values indicate better semantic separation.
405
- - **Alignment Quality:** No aligned models evaluated in this run.
406
  - **Recommendation:** 128d aligned for best cross-lingual performance
407
 
408
  ---
409
  ## 6. Morphological Analysis (Experimental)
410
 
411
- > ⚠️ **Warning:** This language shows low morphological productivity. The statistical signals used for this analysis may be noisy or less reliable than for morphologically rich languages.
412
-
413
  This section presents an automated morphological analysis derived from the statistical divergence between word-level and subword-level models. By analyzing where subword predictability spikes and where word-level coverage fails, we can infer linguistic structures without supervised data.
414
 
415
  ### 6.1 Productivity & Complexity
416
 
417
  | Metric | Value | Interpretation | Recommendation |
418
  |--------|-------|----------------|----------------|
419
- | Productivity Index | **0.000** | Low morphological productivity | ⚠️ Likely unreliable |
420
- | Idiomaticity Gap | **-1.000** | Low formulaic content | - |
421
 
422
  ### 6.2 Affix Inventory (Productive Units)
423
 
@@ -432,18 +467,18 @@ Bound stems are high-frequency subword units that are semantically cohesive but
432
 
433
  | Stem | Cohesion | Substitutability | Examples |
434
  |------|----------|------------------|----------|
435
- | `እንደሚ` | 2.46x | 153 contexts | እንደሚሹ, እንደሚል, እንደሚሉ |
436
- | `ርስቲያ` | 2.48x | 60 contexts | ክርስቲያ, ክርስቲያኗ, ክርስቲያኖ |
437
- | `ትዮጵያ` | 2.23x | 57 contexts | እትዮጵያ, ኢትዮጵያ, ኢትዮጵያው |
438
- | `ግዚአብ` | 2.73x | 24 contexts | እግዚአብሔር, እግዚአብሐር, እግዚአብሄር |
439
- | `ኢትዮጵ` | 2.24x | 46 contexts | ኢትዮጵያ, ኢትዮጵያው, ኢትዮጵስት |
440
- | `መንግሥ` | 2.21x | 46 contexts | መንግሥተ, መንግሥት, መንግሥቱ |
441
- | `መንግስ` | 2.16x | 48 contexts | መንግስት, መንግስተ, መንግስቱ |
442
- | `ፈረንሳ` | 2.33x | 34 contexts | ፈረንሳዊ, ፈረንሳይ, በፈረንሳዩ |
443
- | `አስተዳ` | 2.33x | 33 contexts | አስተዳዳሪ, አስተዳደጓ, አስተዳደረ |
444
- | `እንግሊ` | 2.05x | 53 contexts | እንግሊዝ, እንግሊዙ, እንግሊኛ |
445
- | `tion` | 2.82x | 17 contexts | nation, action, section |
446
- | `ጀመሪያ` | 2.28x | 33 contexts | መጀመሪያ, ለመጀመሪያ, መጀመሪያው |
447
 
448
  ### 6.4 Affix Compatibility (Co-occurrence)
449
 
@@ -462,7 +497,9 @@ Using **Recursive Hierarchical Substitutability**, we decompose complex words in
462
  ### 6.6 Linguistic Interpretation
463
 
464
  > **Automated Insight:**
465
- The language AM appears to be more isolating or has a highly fixed vocabulary. Word-level models perform nearly as well as subword models, indicating fewer productive morphological processes.
 
 
466
 
467
  ---
468
  ## 7. Summary & Recommendations
@@ -474,7 +511,7 @@ The language AM appears to be more isolating or has a highly fixed vocabulary. W
474
  | Component | Recommended | Rationale |
475
  |-----------|-------------|-----------|
476
  | Tokenizer | **64k BPE** | Best compression (3.29x) |
477
- | N-gram | **2-gram** | Lowest perplexity (2,079) |
478
  | Markov | **Context-4** | Highest predictability (98.4%) |
479
  | Embeddings | **100d** | Balanced semantic capture and isotropy |
480
 
@@ -689,4 +726,4 @@ MIT License - Free for academic and commercial use.
689
  ---
690
  *Generated by Wikilangs Models Pipeline*
691
 
692
- *Report Date: 2026-01-03 05:13:17*
 
1
  ---
2
  language: am
3
+ language_name: Amharic
4
  language_family: semitic_ethiopic
5
  tags:
6
  - wikilangs
 
10
  - n-gram
11
  - markov
12
  - wikipedia
13
+ - feature-extraction
14
+ - sentence-similarity
15
+ - tokenization
16
+ - n-grams
17
+ - markov-chain
18
+ - text-mining
19
+ - fasttext
20
+ - babelvec
21
+ - vocabulous
22
+ - vocabulary
23
  - monolingual
24
  - family-semitic_ethiopic
25
  license: mit
26
  library_name: wikilangs
27
+ pipeline_tag: text-generation
28
  datasets:
29
  - omarkamali/wikipedia-monthly
30
  dataset_info:
 
33
  metrics:
34
  - name: best_compression_ratio
35
  type: compression
36
+ value: 3.293
37
  - name: best_isotropy
38
  type: isotropy
39
+ value: 0.9137
40
  - name: vocabulary_size
41
  type: vocab
42
  value: 0
43
  generated: 2026-01-03
44
  ---
45
 
46
+ # Amharic - Wikilangs Models
47
  ## Comprehensive Research Report & Full Ablation Study
48
 
49
+ This repository contains NLP models trained and evaluated by Wikilangs, specifically on **Amharic** Wikipedia data.
50
  We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and word embeddings.
51
 
52
  ## 📋 Repository Contents
 
70
  - [3. Markov Chain Evaluation](#3-markov-chain-evaluation)
71
  - [4. Vocabulary Analysis](#4-vocabulary-analysis)
72
  - [5. Word Embeddings Evaluation](#5-word-embeddings-evaluation)
73
+ - [6. Morphological Analysis (Experimental)](#6--morphological-analysis-experimental)
74
  - [7. Summary & Recommendations](#7-summary--recommendations)
75
  - [Metrics Glossary](#appendix-metrics-glossary--interpretation-guide)
76
  - [Visualizations Index](#visualizations-index)
 
90
 
91
  | Vocab Size | Compression | Avg Token Len | UNK Rate | Total Tokens |
92
  |------------|-------------|---------------|----------|--------------|
93
+ | **8k** | 2.438x | 2.44 | 0.1566% | 682,453 |
94
+ | **16k** | 2.748x | 2.75 | 0.1765% | 605,553 |
95
+ | **32k** | 3.035x | 3.04 | 0.1950% | 548,316 |
96
+ | **64k** | 3.293x 🏆 | 3.29 | 0.2116% | 505,279 |
97
 
98
  ### Tokenization Examples
99
 
100
  Below are sample sentences tokenized with each vocabulary size:
101
 
102
+ **Sample 1:** `እኔ እውነት እናገራለሁ ሌላውን አስኮንናለሁ የአማርኛ ምሳሌ ነው። ትርጉሙ መደብ: ያልተተረጎመ ምሳሌ`
103
 
104
  | Vocab | Tokens | Count |
105
  |-------|--------|-------|
106
+ | 8k | `▁እኔ ▁እውነት ▁እና ገራ ለሁ ▁ሌላ ውን ▁አስ ንና ... (+9 more)` | 19 |
107
+ | 16k | `▁እኔ ▁እውነት ▁እና ገራ ለሁ ▁ሌላውን ▁አስ ንና ለሁ ... (+8 more)` | 18 |
108
+ | 32k | `▁እኔ ▁እውነት ▁እናገራ ለሁ ▁ሌላውን ▁አስ ንና ለሁ ▁የአማርኛ ... (+7 more)` | 17 |
109
+ | 64k | `▁እኔ ▁እውነት ▁እናገራለሁ ▁ሌላውን ▁አስ ኮንና ለሁ ▁የአማርኛ ▁ምሳሌ ▁ነው። ... (+5 more)` | 15 |
110
 
111
+ **Sample 2:** `እንኳን ለገንፎ ለሙቅም አልደነግጥ የአማርኛ ምሳሌ ነው። ትርጉሙ መደብ: ያልተተረጎመ ምሳሌ`
112
 
113
  | Vocab | Tokens | Count |
114
  |-------|--------|-------|
115
+ | 8k | `▁እንኳን ▁ለ ገን ▁ለ ቅም ▁አል ነግ ... (+9 more)` | 19 |
116
+ | 16k | `▁እንኳን ▁ለ ገን ▁ለሙ ቅም ▁አል ደነግ ▁የአማርኛ ... (+7 more)` | 17 |
117
+ | 32k | `▁እንኳን ▁ለ ገንፎ ▁ለሙ ቅም ▁አል ደነግጥ ▁የአማርኛ ▁ምሳሌ ▁ነው። ... (+5 more)` | 15 |
118
+ | 64k | `▁እንኳን ▁ለገንፎ ▁ለሙ ቅም ▁አል ደነግጥ ▁የአማርኛ ▁ምሳሌ ▁ነው። ▁ትርጉሙ ... (+4 more)` | 14 |
119
 
120
+ **Sample 3:** `ሞፈር ረገጠ በአማርኛ ፈሊጣዊ አነጋገር የሆነ ዘይቤ ነው። ትርጉም እራሱን ቻለ። ከቤተሰብ ቁጥጥር ውጭ ሆነ። ምሳሌ ደበበ ዕድሜ...`
121
 
122
  | Vocab | Tokens | Count |
123
  |-------|--------|-------|
124
+ | 8k | `▁ሞ ፈር ▁ረ ▁በአማርኛ ▁ፈሊጣዊ ▁አነጋገር ▁የሆነ ▁ዘይቤ ... (+29 more)` | 39 |
125
+ | 16k | `▁ሞ ፈር ▁ረገ ▁በአማርኛ ▁ፈሊጣዊ ▁አነጋገር ▁የሆነ ▁ዘይቤ ▁ነው። ... (+24 more)` | 34 |
126
+ | 32k | `▁ሞፈር ▁ረገ ▁በአማርኛ ▁ፈሊጣዊ ▁አነጋገር ▁የሆነ ▁ዘይቤ ▁ነው። ▁ትርጉም ... (+21 more)` | 31 |
127
+ | 64k | `▁ሞፈር ▁ረገ ▁በአማርኛ ▁ፈሊጣዊ ▁አነጋገር ▁የሆነ ▁ዘይቤ ▁ነው። ▁ትርጉም ... (+21 more)` | 31 |
128
 
129
 
130
  ### Key Findings
131
 
132
+ - **Best Compression:** 64k achieves 3.293x compression
133
+ - **Lowest UNK Rate:** 8k with 0.1566% unknown tokens
134
  - **Trade-off:** Larger vocabularies improve compression but increase model size
135
  - **Recommendation:** 32k vocabulary provides optimal balance for production use
136
 
 
147
 
148
  | N-gram | Variant | Perplexity | Entropy | Unique N-grams | Top-100 Coverage | Top-1000 Coverage |
149
  |--------|---------|------------|---------|----------------|------------------|-------------------|
150
+ | **2-gram** | Word | 9,101 | 13.15 | 28,185 | 19.6% | 39.5% |
151
+ | **2-gram** | Subword | 2,069 🏆 | 11.01 | 23,787 | 34.1% | 69.3% |
152
+ | **3-gram** | Word | 9,934 | 13.28 | 35,745 | 22.2% | 40.6% |
153
+ | **3-gram** | Subword | 19,035 | 14.22 | 153,217 | 11.9% | 35.6% |
154
+ | **4-gram** | Word | 36,871 | 15.17 | 91,072 | 13.9% | 25.7% |
155
+ | **4-gram** | Subword | 94,475 | 16.53 | 551,504 | 6.6% | 19.5% |
156
+ | **5-gram** | Word | 32,696 | 15.00 | 78,497 | 14.6% | 26.2% |
157
+ | **5-gram** | Subword | 213,435 | 17.70 | 879,311 | 5.0% | 14.3% |
158
 
159
  ### Top 5 N-grams by Size
160
 
 
162
 
163
  | Rank | N-gram | Count |
164
  |------|--------|-------|
165
+ | 1 | `ዓ ም` | 8,266 |
166
+ | 2 | `ምሳሌ ነው` | 5,623 |
167
+ | 3 | `የአማርኛ ምሳሌ` | 5,562 |
168
+ | 4 | `እ ኤ` | 4,014 |
169
+ | 5 | `ኤ አ` | 3,948 |
170
 
171
  **3-grams (Word):**
172
 
173
  | Rank | N-gram | Count |
174
  |------|--------|-------|
175
+ | 1 | `የአማርኛ ምሳሌ ነው` | 5,562 |
176
+ | 2 | `እ ኤ አ` | 3,896 |
177
  | 3 | `ምሳሌ ነው ትርጉሙ` | 3,454 |
178
  | 4 | `መደብ ተረትና ምሳሌ` | 3,051 |
179
+ | 5 | `ነው ትርጉሙ መደብ` | 2,530 |
180
 
181
  **4-grams (Word):**
182
 
183
  | Rank | N-gram | Count |
184
  |------|--------|-------|
185
  | 1 | `የአማርኛ ምሳሌ ነው ትርጉሙ` | 3,452 |
186
+ | 2 | `ምሳሌ ነው ትርጉሙ መደብ` | 2,530 |
187
+ | 3 | `ትርጉሙ መደብ ያልተተረጎመ ምሳሌ` | 2,115 |
188
+ | 4 | `ነው ትርጉሙ መደብ ያልተተረጎመ` | 2,111 |
189
  | 5 | `ምሳሌ መደብ ተረትና ምሳሌ` | 1,854 |
190
 
191
+ **5-grams (Word):**
192
+
193
+ | Rank | N-gram | Count |
194
+ |------|--------|-------|
195
+ | 1 | `የአማርኛ ምሳሌ ነው ትርጉሙ መደብ` | 2,529 |
196
+ | 2 | `ምሳሌ ነው ትርጉሙ መደብ ያልተተረጎመ` | 2,111 |
197
+ | 3 | `ነው ትርጉሙ መደብ ያልተተረጎመ ምሳሌ` | 2,111 |
198
+ | 4 | `መደብ ያልተተረጎመ ምሳሌ መደብ ተረትና` | 1,812 |
199
+ | 5 | `ያልተተረጎመ ምሳሌ መደብ ተረትና ምሳሌ` | 1,811 |
200
+
201
  **2-grams (Subword):**
202
 
203
  | Rank | N-gram | Count |
204
  |------|--------|-------|
205
+ | 1 | `_ የ` | 172,656 |
206
+ | 2 | `ት _` | 146,889 |
207
+ | 3 | `_ በ` | 142,558 |
208
+ | 4 | `ን _` | 134,273 |
209
+ | 5 | `_ አ` | 115,168 |
210
 
211
  **3-grams (Subword):**
212
 
213
  | Rank | N-gram | Count |
214
  |------|--------|-------|
215
+ | 1 | `_ እ ን` | 32,943 |
216
+ | 2 | `_ ነ ው` | 26,886 |
217
+ | 3 | `_ ና` | 24,633 |
218
+ | 4 | `ው _` | 24,427 |
219
+ | 5 | `እ ና _` | 23,097 |
220
 
221
  **4-grams (Subword):**
222
 
223
  | Rank | N-gram | Count |
224
  |------|--------|-------|
225
+ | 1 | `_ እ ና _` | 22,966 |
226
+ | 2 | `_ ነ ው ።` | 19,603 |
227
+ | 3 | `ነ ው ። _` | 19,130 |
228
+ | 4 | `_ እ ን ደ` | 14,167 |
229
+ | 5 | `_ ላ ይ _` | 13,064 |
230
+
231
+ **5-grams (Subword):**
232
+
233
+ | Rank | N-gram | Count |
234
+ |------|--------|-------|
235
+ | 1 | `_ ነ ው ። _` | 19,000 |
236
+ | 2 | `_ ው ስ ጥ _` | 9,650 |
237
+ | 3 | `ኢ ት ዮ ጵ ያ` | 7,988 |
238
+ | 4 | `_ ም ሳ ሌ _` | 7,852 |
239
+ | 5 | `_ እ ን ደ _` | 6,562 |
240
 
241
 
242
  ### Key Findings
243
 
244
+ - **Best Perplexity:** 2-gram (subword) with 2,069
245
  - **Entropy Trend:** Decreases with larger n-grams (more predictable)
246
+ - **Coverage:** Top-1000 patterns cover ~14% of corpus
247
  - **Recommendation:** 4-gram or 5-gram for best predictive performance
248
 
249
  ---
 
259
 
260
  | Context | Variant | Avg Entropy | Perplexity | Branching Factor | Unique Contexts | Predictability |
261
  |---------|---------|-------------|------------|------------------|-----------------|----------------|
262
+ | **1** | Word | 0.7520 | 1.684 | 4.82 | 237,556 | 24.8% |
263
+ | **1** | Subword | 1.2212 | 2.331 | 17.49 | 2,857 | 0.0% |
264
+ | **2** | Word | 0.1473 | 1.108 | 1.28 | 1,142,374 | 85.3% |
265
+ | **2** | Subword | 1.0395 | 2.055 | 6.98 | 49,956 | 0.0% |
266
+ | **3** | Word | 0.0354 | 1.025 | 1.06 | 1,462,526 | 96.5% |
267
+ | **3** | Subword | 0.6359 | 1.554 | 3.37 | 348,652 | 36.4% |
268
+ | **4** | Word | 0.0157 🏆 | 1.011 | 1.02 | 1,537,232 | 98.4% |
269
+ | **4** | Subword | 0.4526 | 1.368 | 2.15 | 1,173,222 | 54.7% |
270
 
271
  ### Generated Text Samples (Word-based)
272
 
 
274
 
275
  **Context Size 1:**
276
 
277
+ 1. `ነው ወደዚህም የተሳበው ተፈጥሮን ፀባዮች የሚመረኮዝ ፆታዊ ውሳኔ ተላልፏል ውድድሩን ኤርት��ውያን የኩራት ምንጭ የበለጠ የተገደበ`
278
+ 2. `እና ሳይንሶችን እንዲሁም ከላስታና ከላሊበላ ከፍተኛ የመብት ጥሰቶች የተካሄደውን መፈንቅለ መንግሥት በጌሤም የአሦርም`
279
+ 3. `ላይ እንዲገኝ ስለሚያስገድድ ነው ኬንያ ወደሚገኘው ማይ ጎጋ የተባለ የህንድ ጥቃቶች የተጠበቀ እና በችግር ጊዜ የተረጋገጠ`
280
 
281
  **Context Size 2:**
282
 
283
+ 1. `ዓ ም የዊስቡር ልጅ ዶማር 300 307 ተከለከለ ታጂኪስታን ዓ ም በነሐሴ ወር 450 ዓ`
284
+ 2. `ምሳሌ ነው ጦጣ መጀመሪያ የመቀመጫዬን አለች አሉ ጦጣ ባለቤቱን ታስወጣ የአማርኛ ምሳሌ ነው ትርጉሙ መደብ ያልተተረጎመ ምሳሌ`
285
+ 3. `የአማርኛ ምሳሌ ነው ለላሙ መንጃ ለሸማው መቅደጃ የአማርኛ ምሳሌ ነው ትርጉሙ መደብ ተረትና ምሳሌ መደብ ተረትና ምሳሌ`
286
 
287
  **Context Size 3:**
288
 
289
+ 1. `የአማርኛ ምሳሌ ነው ትርጉሙ መደብ ያልተተረጎመ ምሳሌ መደብ ተረትና ምሳሌ wiz`
290
+ 2. `እ ኤ አ ቦራስ ስዊድን የግሪክ ዘፋኝ ነች አልበሞች protereotita my number one iparhi logos the game of`
291
+ 3. `ምሳሌ ነው ትርጉሙ መደብ ያልተተረጎመ ምሳሌ መደብ ተረትና ምሳሌ መደብ ተረትና ምሳሌ መደብ ተረትና ምሳሌ ምሳሌ`
292
 
293
  **Context Size 4:**
294
 
295
+ 1. `የአማርኛ ምሳሌ ነው ትርጉሙ መደብ ያልተተረጎመ ምሳሌ መደብ ተረትና ምሳሌ wiz`
296
+ 2. `ምሳሌ ነው ትርጉሙ መደብ ተረትና ምሳሌ በሬ ካራጁ ይዉላል`
297
+ 3. `ነው ትርጉሙ መደብ ያልተተረጎመ ምሳሌ መደብ ተረትና ምሳሌ መደብ ያልተተረጎመ ምሳሌ`
298
 
299
 
300
  ### Generated Text Samples (Subword-based)
 
303
 
304
  **Context Size 1:**
305
 
306
+ 1. `_493_የለቀድ_አት_በተገ`
307
+ 2. `ንዳዎች_20_የሳት_ወቀን_`
308
+ 3. `ት_ገኙ_ነበትላን_ጆች_በዚ`
309
 
310
  **Context Size 2:**
311
 
312
+ 1. `_የጠፈ_እጅጉ_ሙከራ_ተፈጥሮ`
313
+ 2. `ት_ከተማ_እቃ_ለማብራዶሮ_ሶ`
314
+ 3. `_በሁለተቸት_ስለ_ተመሳር_ከ`
315
 
316
  **Context Size 3:**
317
 
318
+ 1. `_እንዲሁም_ዘር።_ከነዚህ_ጊዜ`
319
+ 2. `_ነው_፡፡_አየሩ_በኋላም_ብዙ`
320
+ 3. `_እና_ጁላይ_ጥይቶቹ_ላይ_(2`
321
 
322
  **Context Size 4:**
323
 
324
+ 1. `_እና_የከተማ፡-_ጎንደርና_አገ`
325
+ 2. `_ነው።_ባብዛኛው_ህይወት_ውስጥ`
326
+ 3. `ነው።_ዓ.ም_ኪዮሺ_ሱጊዩራ_(1`
327
 
328
 
329
  ### Key Findings
330
 
331
  - **Best Predictability:** Context-4 (word) with 98.4% predictability
332
  - **Branching Factor:** Decreases with context size (more deterministic)
333
+ - **Memory Trade-off:** Larger contexts require more storage (1,173,222 contexts)
334
  - **Recommendation:** Context-3 or Context-4 for text generation
335
 
336
  ---
 
346
 
347
  | Metric | Value |
348
  |--------|-------|
349
+ | Vocabulary Size | 100,186 |
350
+ | Total Tokens | 1,652,256 |
351
+ | Mean Frequency | 16.49 |
352
  | Median Frequency | 3 |
353
+ | Frequency Std Dev | 176.36 |
354
 
355
  ### Most Common Words
356
 
357
  | Rank | Word | Frequency |
358
  |------|------|-----------|
359
+ | 1 | ነው | 26,831 |
360
+ | 2 | እና | 23,089 |
361
+ | 3 | ላይ | 13,382 |
362
+ | 4 | ምሳሌ | 11,608 |
363
+ | 5 | ውስጥ | 9,891 |
364
+ | 6 | ነበር | 9,130 |
365
+ | 7 | ዓ | 8,627 |
366
+ | 8 | ወደ | 8,565 |
367
+ | 9 | | 8,525 |
368
+ | 10 | እንደ | 6,906 |
369
 
370
  ### Least Common Words (from vocabulary)
371
 
372
  | Rank | Word | Frequency |
373
  |------|------|-----------|
374
+ | 1 | ጂኒካ | 2 |
375
+ | 2 | ዲኒካላ | 2 |
376
+ | 3 | ወስደሽ | 2 |
377
+ | 4 | አንኳኳ | 2 |
378
+ | 5 | መዳልወ | 2 |
379
+ | 6 | ረድእ | 2 |
380
+ | 7 | አንደኛይቱ | 2 |
381
+ | 8 | ወደሰልፍ | 2 |
382
+ | 9 | የኒኮፖሊስ | 2 |
383
+ | 10 | ጂምናዚየም | 2 |
384
 
385
  ### Zipf's Law Analysis
386
 
387
  | Metric | Value |
388
  |--------|-------|
389
+ | Zipf Coefficient | 0.9364 |
390
+ | R² (Goodness of Fit) | 0.995158 |
391
  | Adherence Quality | **excellent** |
392
 
393
  ### Coverage Analysis
 
403
 
404
  - **Zipf Compliance:** R²=0.9952 indicates excellent adherence to Zipf's law
405
  - **High Frequency Dominance:** Top 100 words cover 22.7% of corpus
406
+ - **Long Tail:** 90,186 words needed for remaining 25.1% coverage
407
 
408
  ---
409
  ## 5. Word Embeddings Evaluation
 
419
 
420
  ### 5.1 Cross-Lingual Alignment
421
 
422
+ ![Alignment Quality](visualizations/embedding_alignment_quality.png)
423
+
424
+ ![Multilingual t-SNE](visualizations/embedding_tsne_multilingual.png)
425
 
426
 
427
  ### 5.2 Model Comparison
428
 
429
  | Model | Dimension | Isotropy | Semantic Density | Alignment R@1 | Alignment R@10 |
430
  |-------|-----------|----------|------------------|---------------|----------------|
431
+ | **mono_32d** | 32 | 0.9080 | 0.3255 | N/A | N/A |
432
+ | **mono_64d** | 64 | 0.9137 | 0.2344 | N/A | N/A |
433
+ | **mono_128d** | 128 | 0.8453 | 0.1726 | N/A | N/A |
434
+ | **aligned_32d** | 32 | 0.9080 | 0.3232 | 0.0220 | 0.1700 |
435
+ | **aligned_64d** | 64 | 0.9137 🏆 | 0.2323 | 0.0420 | 0.1840 |
436
+ | **aligned_128d** | 128 | 0.8453 | 0.1725 | 0.0680 | 0.2480 |
437
 
438
  ### Key Findings
439
 
440
+ - **Best Isotropy:** aligned_64d with 0.9137 (more uniform distribution)
441
+ - **Semantic Density:** Average pairwise similarity of 0.2434. Lower values indicate better semantic separation.
442
+ - **Alignment Quality:** Aligned models achieve up to 6.8% R@1 in cross-lingual retrieval.
443
  - **Recommendation:** 128d aligned for best cross-lingual performance
444
 
445
  ---
446
  ## 6. Morphological Analysis (Experimental)
447
 
 
 
448
  This section presents an automated morphological analysis derived from the statistical divergence between word-level and subword-level models. By analyzing where subword predictability spikes and where word-level coverage fails, we can infer linguistic structures without supervised data.
449
 
450
  ### 6.1 Productivity & Complexity
451
 
452
  | Metric | Value | Interpretation | Recommendation |
453
  |--------|-------|----------------|----------------|
454
+ | Productivity Index | **5.000** | High morphological productivity | Reliable analysis |
455
+ | Idiomaticity Gap | **0.840** | High formulaic/idiomatic content | - |
456
 
457
  ### 6.2 Affix Inventory (Productive Units)
458
 
 
467
 
468
  | Stem | Cohesion | Substitutability | Examples |
469
  |------|----------|------------------|----------|
470
+ | `እንደሚ` | 2.39x | 158 contexts | እንደሚሉ, እንደሚሻ, እንደሚል |
471
+ | `ርስቲያ` | 2.46x | 61 contexts | ክርስቲያ, ክርስቲያኗ, ክርስቲያኑ |
472
+ | `ትዮጵያ` | 2.23x | 57 contexts | ኢትዮጵያ, እትዮጵያ, ኢትዮጵያና |
473
+ | `መንግስ` | 2.21x | 49 contexts | መንግስቱ, መንግስት, መንግስተ |
474
+ | `ግዚአብ` | 2.66x | 23 contexts | እግዚአብሔር, እግዚአብሐር, እግዚአብሄር |
475
+ | `ኢትዮጵ` | 2.18x | 46 contexts | ኢትዮጵያ, ኢትዮጵያና, የኢትዮጵያ |
476
+ | `እንግሊ` | 2.08x | 52 contexts | እንግሊኛ, እንግሊዙ, እንግሊዝ |
477
+ | `መንግሥ` | 2.12x | 46 contexts | መንግሥት, መንግሥተ, መንግሥቱ |
478
+ | `ጀመሪያ` | 2.29x | 33 contexts | መጀመሪያ, በመጀመሪያ, ለመጀመሪያ |
479
+ | `ፈረንሳ` | 2.27x | 34 contexts | ፈረንሳይ, ፈረንሳዊ, የፈረንሳዩ |
480
+ | `tion` | 2.77x | 17 contexts | nation, action, section |
481
+ | `መጀመሪ` | 2.29x | 31 contexts | መጀመሪአ, መጀመሪያ, የመጀመሪ |
482
 
483
  ### 6.4 Affix Compatibility (Co-occurrence)
484
 
 
497
  ### 6.6 Linguistic Interpretation
498
 
499
  > **Automated Insight:**
500
+ The language Amharic shows high morphological productivity. The subword models are significantly more efficient than word models, suggesting a rich system of affixation or compounding.
501
+
502
+ > **Note on Idiomaticity:** The high Idiomaticity Gap suggests a large number of frequent multi-word expressions or formulaic sequences that are statistically distinct from their component parts.
503
 
504
  ---
505
  ## 7. Summary & Recommendations
 
511
  | Component | Recommended | Rationale |
512
  |-----------|-------------|-----------|
513
  | Tokenizer | **64k BPE** | Best compression (3.29x) |
514
+ | N-gram | **2-gram** | Lowest perplexity (2,069) |
515
  | Markov | **Context-4** | Highest predictability (98.4%) |
516
  | Embeddings | **100d** | Balanced semantic capture and isotropy |
517
 
 
726
  ---
727
  *Generated by Wikilangs Models Pipeline*
728
 
729
+ *Report Date: 2026-01-03 14:11:24*
models/embeddings/aligned/am_128d.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3964ee0c4f9ca092d9f74907f56f9a3d93b19752347882a64e552d576a095e2b
3
+ size 1064306440
models/embeddings/aligned/am_128d.meta.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lang": "am", "dim": 128, "max_seq_len": 512, "is_aligned": true}
models/embeddings/aligned/am_128d.projection.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfbe39d8562cf38339d4ba377db708ff89f8f76337965c05da5c47a5511cd90d
3
+ size 65664
models/embeddings/aligned/am_128d_metadata.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "language": "am",
3
+ "dimension": 128,
4
+ "version": "aligned",
5
+ "hub_language": "en",
6
+ "seed_vocab_size": 2411,
7
+ "vocab_size": 38514
8
+ }
models/embeddings/aligned/am_32d.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43acd2454c73ec6a90e9b75922f4a4ecc9db024320951b2c88c136aaaf3e57dc
3
+ size 266727688
models/embeddings/aligned/am_32d.meta.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lang": "am", "dim": 32, "max_seq_len": 512, "is_aligned": true}
models/embeddings/aligned/am_32d.projection.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3d0d5ccf51dddf65743a5123bfc0ecd1944f4cfce415736d0e147acf3d55f4b
3
+ size 4224
models/embeddings/aligned/am_32d_metadata.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "language": "am",
3
+ "dimension": 32,
4
+ "version": "aligned",
5
+ "hub_language": "en",
6
+ "seed_vocab_size": 2411,
7
+ "vocab_size": 38514
8
+ }
models/embeddings/aligned/am_64d.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad86c60bc61db296fa76aedb6ab90d476fc19f98d61421e743d16270d0805cf5
3
+ size 532587272
models/embeddings/aligned/am_64d.meta.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lang": "am", "dim": 64, "max_seq_len": 512, "is_aligned": true}
models/embeddings/aligned/am_64d.projection.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:56b5628675cc1634a4f9405dd4c0ad1d8ef1da74827604d4d4f4a7c37742850a
3
+ size 16512
models/embeddings/aligned/am_64d_metadata.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "language": "am",
3
+ "dimension": 64,
4
+ "version": "aligned",
5
+ "hub_language": "en",
6
+ "seed_vocab_size": 2411,
7
+ "vocab_size": 38514
8
+ }
models/embeddings/monolingual/am_128d.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1d4be827f7a31270ecd980640a7032e11c5ed3885d037c1b8cd46df3d9007492
3
- size 1063990202
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3964ee0c4f9ca092d9f74907f56f9a3d93b19752347882a64e552d576a095e2b
3
+ size 1064306440
models/embeddings/monolingual/am_128d_metadata.json CHANGED
@@ -11,5 +11,5 @@
11
  "encoding_method": "rope",
12
  "dim": 128
13
  },
14
- "vocab_size": 38213
15
  }
 
11
  "encoding_method": "rope",
12
  "dim": 128
13
  },
14
+ "vocab_size": 38514
15
  }
models/embeddings/monolingual/am_32d.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1facd0cecd15328df01b8eb00c409adeaac9321a5d62748462e3055c6a4b976e
3
- size 266642618
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43acd2454c73ec6a90e9b75922f4a4ecc9db024320951b2c88c136aaaf3e57dc
3
+ size 266727688
models/embeddings/monolingual/am_32d_metadata.json CHANGED
@@ -11,5 +11,5 @@
11
  "encoding_method": "rope",
12
  "dim": 32
13
  },
14
- "vocab_size": 38213
15
  }
 
11
  "encoding_method": "rope",
12
  "dim": 32
13
  },
14
+ "vocab_size": 38514
15
  }
models/embeddings/monolingual/am_64d.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f42620197f1e1b51b39471cc1e5886ab38c161775048e9529ceb06fdf065e6e1
3
- size 532425146
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad86c60bc61db296fa76aedb6ab90d476fc19f98d61421e743d16270d0805cf5
3
+ size 532587272
models/embeddings/monolingual/am_64d_metadata.json CHANGED
@@ -11,5 +11,5 @@
11
  "encoding_method": "rope",
12
  "dim": 64
13
  },
14
- "vocab_size": 38213
15
  }
 
11
  "encoding_method": "rope",
12
  "dim": 64
13
  },
14
+ "vocab_size": 38514
15
  }
models/subword_markov/am_markov_ctx1_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0dea8b9be146dc2d3f0da26601ce42020e0c84af4656d9f7692b3f4d34d5d6ae
3
- size 356106
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8839cf6220dd43e6dea9746686fa8a81a9a14dd02c6132ff81405220cf661652
3
+ size 350051
models/subword_markov/am_markov_ctx1_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 1,
3
  "variant": "subword",
4
  "language": "am",
5
- "unique_contexts": 2854,
6
- "total_transitions": 8936316
7
  }
 
2
  "context_size": 1,
3
  "variant": "subword",
4
  "language": "am",
5
+ "unique_contexts": 2857,
6
+ "total_transitions": 9022824
7
  }
models/subword_markov/am_markov_ctx2_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0e9b01d5674bb78549bd0f7a41241639ddc7934db689c638aee8d0daacc70172
3
- size 2154891
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:27974fda5f8693d4915e9b657526e84bd0fb36fde9d966f19e785d000aedbe5c
3
+ size 2143373
models/subword_markov/am_markov_ctx2_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 2,
3
  "variant": "subword",
4
  "language": "am",
5
- "unique_contexts": 49981,
6
- "total_transitions": 8923902
7
  }
 
2
  "context_size": 2,
3
  "variant": "subword",
4
  "language": "am",
5
+ "unique_contexts": 49956,
6
+ "total_transitions": 9010410
7
  }
models/subword_markov/am_markov_ctx3_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:36e05d4c86f22f4d3f6c7dff80e4b868513b425317942829e61a871110f71ee0
3
- size 8408029
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:82c019b23a13a08dd018f89e36328fb55a539c91738572761bc8164c5041b3f4
3
+ size 8416653
models/subword_markov/am_markov_ctx3_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 3,
3
  "variant": "subword",
4
  "language": "am",
5
- "unique_contexts": 348535,
6
- "total_transitions": 8911488
7
  }
 
2
  "context_size": 3,
3
  "variant": "subword",
4
  "language": "am",
5
+ "unique_contexts": 348652,
6
+ "total_transitions": 8997996
7
  }
models/subword_markov/am_markov_ctx4_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ab3d345339fb5097d316a6acad43d3cbd7f781c24bd690b17950a9d30bdd7dfa
3
- size 23278804
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f07c64385b4c55e182b6ffee304e1810b2d71ef22b3c5a0b251758b633fca59c
3
+ size 23309934
models/subword_markov/am_markov_ctx4_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 4,
3
  "variant": "subword",
4
  "language": "am",
5
- "unique_contexts": 1171344,
6
- "total_transitions": 8899074
7
  }
 
2
  "context_size": 4,
3
  "variant": "subword",
4
  "language": "am",
5
+ "unique_contexts": 1173222,
6
+ "total_transitions": 8985582
7
  }
models/subword_ngram/am_2gram_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6bcf32e20adcf3b9aa77e784d9fd125484c7c7d01f306652cf74a76522fbc536
3
- size 300324
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5eabd2483a9b43b8f8fd572f71957cc349ed926f2f419138f2f3a188ed78c42f
3
+ size 300472
models/subword_ngram/am_2gram_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "n": 2,
3
  "variant": "subword",
4
  "language": "am",
5
- "unique_ngrams": 23804,
6
- "total_ngrams": 8936316
7
  }
 
2
  "n": 2,
3
  "variant": "subword",
4
  "language": "am",
5
+ "unique_ngrams": 23787,
6
+ "total_ngrams": 9022824
7
  }
models/subword_ngram/am_3gram_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9c8cece5e05ace848d6a685201f3e49e33ed838dc7744aec46795f09bbfdb9e0
3
- size 1881540
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c9b592f96c76cc3101fee1a715cf6c4149dfe51039e391d6098096f9714abe5
3
+ size 1885405
models/subword_ngram/am_3gram_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "n": 3,
3
  "variant": "subword",
4
  "language": "am",
5
- "unique_ngrams": 153027,
6
- "total_ngrams": 8923902
7
  }
 
2
  "n": 3,
3
  "variant": "subword",
4
  "language": "am",
5
+ "unique_ngrams": 153217,
6
+ "total_ngrams": 9010410
7
  }
models/subword_ngram/am_4gram_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f6cf11a9b9599d5b462fc5775302067d643de8e8fd248de5060f89351b18b9ab
3
- size 7082960
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:899527071ee5237fc41c2fdb6de104208b69f563db321792d2b61794349f7f99
3
+ size 7120084
models/subword_ngram/am_4gram_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "n": 4,
3
  "variant": "subword",
4
  "language": "am",
5
- "unique_ngrams": 549996,
6
- "total_ngrams": 8911488
7
  }
 
2
  "n": 4,
3
  "variant": "subword",
4
  "language": "am",
5
+ "unique_ngrams": 551504,
6
+ "total_ngrams": 8997996
7
  }
models/subword_ngram/am_5gram_subword.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6256142f289450d718eafc3693efd234fd5e20489751f38071b275a053df337
3
+ size 12113547
models/subword_ngram/am_5gram_subword_metadata.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "n": 5,
3
+ "variant": "subword",
4
+ "language": "am",
5
+ "unique_ngrams": 879311,
6
+ "total_ngrams": 8985582
7
+ }
models/tokenizer/am_tokenizer_16k.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f02cb4592939e59c4831f57ca855266e8d28172e5efebde508c8b57daefbafb6
3
- size 559482
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:802c7e23f92ccb7959b0feb6dc8f82635d55d846dbd9f4570913915ce17d5785
3
+ size 559625
models/tokenizer/am_tokenizer_16k.vocab CHANGED
The diff for this file is too large to render. See raw diff
 
models/tokenizer/am_tokenizer_32k.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:93bd53271f8c4e6a918072a97dd0581c020a8af645616093bc8213a30b88940e
3
- size 902409
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15d15bc01b2e176dbce09f1705536a89afa2737570c8c47d3353634f9f68a94a
3
+ size 902568
models/tokenizer/am_tokenizer_32k.vocab CHANGED
The diff for this file is too large to render. See raw diff
 
models/tokenizer/am_tokenizer_64k.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b8481661d7e8d938bace980d6a5c315614597c92fb98ccbef70df78797efdecc
3
- size 1589488
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5022a070bcee7d18664184b36241ade2425fae873f2d51e1b98af97932ce4f68
3
+ size 1589838
models/tokenizer/am_tokenizer_64k.vocab CHANGED
The diff for this file is too large to render. See raw diff
 
models/tokenizer/am_tokenizer_8k.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4ad931bdd41158b2542d8aac191257d87f612a4ef884c47473efc0d7a86dd011
3
- size 394754
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a8188f6c50ce22b57642fe6d5b7e098ba217e95117378d4c365656422f23b18
3
+ size 394741
models/tokenizer/am_tokenizer_8k.vocab CHANGED
The diff for this file is too large to render. See raw diff
 
models/vocabulary/am_vocabulary.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:294dfe907fb7e7f0defd53a349bd532546806ddc295a92c292188997c0d498bd
3
- size 1787217
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd540986de5c037c2d8004d9247b062277611308cc1b9bc858ffc3394fee94dd
3
+ size 1782777
models/vocabulary/am_vocabulary_metadata.json CHANGED
@@ -1,17 +1,17 @@
1
  {
2
  "language": "am",
3
- "vocabulary_size": 99716,
4
  "variant": "full",
5
  "statistics": {
6
- "type_token_ratio": 0.13335844069738334,
7
  "coverage": {
8
- "top_100": 0.20952339607919193,
9
- "top_1000": 0.42309028051841446,
10
- "top_5000": 0.6109410976729082,
11
- "top_10000": 0.6909696930060957
12
  },
13
- "hapax_count": 136824,
14
- "hapax_ratio": 0.5784391646233196,
15
  "total_documents": 12414
16
  }
17
  }
 
1
  {
2
  "language": "am",
3
+ "vocabulary_size": 100186,
4
  "variant": "full",
5
  "statistics": {
6
+ "type_token_ratio": 0.13283071071151606,
7
  "coverage": {
8
+ "top_100": 0.209810304098978,
9
+ "top_1000": 0.42304666635378463,
10
+ "top_5000": 0.6110407126558544,
11
+ "top_10000": 0.6911446565337589
12
  },
13
+ "hapax_count": 137556,
14
+ "hapax_ratio": 0.5785936014671367,
15
  "total_documents": 12414
16
  }
17
  }
models/word_markov/am_markov_ctx1_word.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:61264cf69517fe843352a0b6cc8383107c7cd20cb9613edb6a901779636cd1bd
3
- size 13693182
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95ccb7912954a141c9b8b7ca95e221a9fccec6dd51f1c444544019f2ac02b83d
3
+ size 13813306
models/word_markov/am_markov_ctx1_word_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 1,
3
  "variant": "word",
4
  "language": "am",
5
- "unique_contexts": 236353,
6
- "total_transitions": 1761302
7
  }
 
2
  "context_size": 1,
3
  "variant": "word",
4
  "language": "am",
5
+ "unique_contexts": 237556,
6
+ "total_transitions": 1777398
7
  }
models/word_markov/am_markov_ctx2_word.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c565bae97b8cf56a6c0d532f8f790eb6482096f8f6f9ee9069aec8edcf717f1d
3
- size 29639293
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1434179ed6804e7277902166d67afa9d340c443633dfac8c4f6574077bc3705
3
+ size 30016914
models/word_markov/am_markov_ctx2_word_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 2,
3
  "variant": "word",
4
  "language": "am",
5
- "unique_contexts": 1130961,
6
- "total_transitions": 1748889
7
  }
 
2
  "context_size": 2,
3
  "variant": "word",
4
  "language": "am",
5
+ "unique_contexts": 1142374,
6
+ "total_transitions": 1764985
7
  }