omarkamali commited on
Commit
6afc7d3
·
verified ·
1 Parent(s): e32d67b

Upload all models and assets for cr (latest)

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +1 -0
  2. README.md +131 -96
  3. models/embeddings/aligned/cr_128d.bin +3 -0
  4. models/embeddings/aligned/cr_128d.meta.json +1 -0
  5. models/embeddings/aligned/cr_128d.projection.npy +3 -0
  6. models/embeddings/aligned/cr_128d_metadata.json +8 -0
  7. models/embeddings/aligned/cr_32d.bin +3 -0
  8. models/embeddings/aligned/cr_32d.meta.json +1 -0
  9. models/embeddings/aligned/cr_32d.projection.npy +3 -0
  10. models/embeddings/aligned/cr_32d_metadata.json +8 -0
  11. models/embeddings/aligned/cr_64d.bin +3 -0
  12. models/embeddings/aligned/cr_64d.meta.json +1 -0
  13. models/embeddings/aligned/cr_64d.projection.npy +3 -0
  14. models/embeddings/aligned/cr_64d_metadata.json +8 -0
  15. models/embeddings/monolingual/cr_128d.bin +1 -1
  16. models/embeddings/monolingual/cr_32d.bin +1 -1
  17. models/embeddings/monolingual/cr_64d.bin +1 -1
  18. models/subword_markov/cr_markov_ctx1_subword.parquet +2 -2
  19. models/subword_markov/cr_markov_ctx1_subword_metadata.json +2 -2
  20. models/subword_markov/cr_markov_ctx2_subword.parquet +2 -2
  21. models/subword_markov/cr_markov_ctx2_subword_metadata.json +2 -2
  22. models/subword_markov/cr_markov_ctx3_subword.parquet +2 -2
  23. models/subword_markov/cr_markov_ctx3_subword_metadata.json +2 -2
  24. models/subword_markov/cr_markov_ctx4_subword.parquet +2 -2
  25. models/subword_markov/cr_markov_ctx4_subword_metadata.json +2 -2
  26. models/subword_ngram/cr_2gram_subword.parquet +2 -2
  27. models/subword_ngram/cr_2gram_subword_metadata.json +2 -2
  28. models/subword_ngram/cr_3gram_subword.parquet +2 -2
  29. models/subword_ngram/cr_3gram_subword_metadata.json +2 -2
  30. models/subword_ngram/cr_4gram_subword.parquet +2 -2
  31. models/subword_ngram/cr_4gram_subword_metadata.json +2 -2
  32. models/subword_ngram/cr_5gram_subword.parquet +3 -0
  33. models/subword_ngram/cr_5gram_subword_metadata.json +7 -0
  34. models/tokenizer/cr_tokenizer_8k.model +2 -2
  35. models/tokenizer/cr_tokenizer_8k.vocab +0 -0
  36. models/vocabulary/cr_vocabulary.parquet +2 -2
  37. models/vocabulary/cr_vocabulary_metadata.json +6 -6
  38. models/word_markov/cr_markov_ctx1_word.parquet +2 -2
  39. models/word_markov/cr_markov_ctx1_word_metadata.json +2 -2
  40. models/word_markov/cr_markov_ctx2_word.parquet +2 -2
  41. models/word_markov/cr_markov_ctx2_word_metadata.json +2 -2
  42. models/word_markov/cr_markov_ctx3_word.parquet +2 -2
  43. models/word_markov/cr_markov_ctx3_word_metadata.json +2 -2
  44. models/word_markov/cr_markov_ctx4_word.parquet +2 -2
  45. models/word_markov/cr_markov_ctx4_word_metadata.json +2 -2
  46. models/word_ngram/cr_2gram_word_metadata.json +1 -1
  47. models/word_ngram/cr_3gram_word_metadata.json +1 -1
  48. models/word_ngram/cr_4gram_word.parquet +2 -2
  49. models/word_ngram/cr_4gram_word_metadata.json +2 -2
  50. models/word_ngram/cr_5gram_word.parquet +3 -0
.gitattributes CHANGED
@@ -36,3 +36,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
36
  visualizations/embedding_similarity.png filter=lfs diff=lfs merge=lfs -text
37
  visualizations/performance_dashboard.png filter=lfs diff=lfs merge=lfs -text
38
  visualizations/position_encoding_comparison.png filter=lfs diff=lfs merge=lfs -text
 
 
36
  visualizations/embedding_similarity.png filter=lfs diff=lfs merge=lfs -text
37
  visualizations/performance_dashboard.png filter=lfs diff=lfs merge=lfs -text
38
  visualizations/position_encoding_comparison.png filter=lfs diff=lfs merge=lfs -text
39
+ visualizations/embedding_tsne_multilingual.png filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  language: cr
3
- language_name: CR
4
  language_family: american_algonquian
5
  tags:
6
  - wikilangs
@@ -10,11 +10,21 @@ tags:
10
  - n-gram
11
  - markov
12
  - wikipedia
 
 
 
 
 
 
 
 
 
 
13
  - monolingual
14
  - family-american_algonquian
15
  license: mit
16
  library_name: wikilangs
17
- pipeline_tag: feature-extraction
18
  datasets:
19
  - omarkamali/wikipedia-monthly
20
  dataset_info:
@@ -23,20 +33,20 @@ dataset_info:
23
  metrics:
24
  - name: best_compression_ratio
25
  type: compression
26
- value: 3.182
27
  - name: best_isotropy
28
  type: isotropy
29
- value: 0.0381
30
  - name: vocabulary_size
31
  type: vocab
32
  value: 0
33
  generated: 2026-01-03
34
  ---
35
 
36
- # CR - Wikilangs Models
37
  ## Comprehensive Research Report & Full Ablation Study
38
 
39
- This repository contains NLP models trained and evaluated by Wikilangs, specifically on **CR** Wikipedia data.
40
  We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and word embeddings.
41
 
42
  ## 📋 Repository Contents
@@ -60,7 +70,7 @@ We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and
60
  - [3. Markov Chain Evaluation](#3-markov-chain-evaluation)
61
  - [4. Vocabulary Analysis](#4-vocabulary-analysis)
62
  - [5. Word Embeddings Evaluation](#5-word-embeddings-evaluation)
63
- - [6. Morphological Analysis (Experimental)](#6-morphological-analysis)
64
  - [7. Summary & Recommendations](#7-summary--recommendations)
65
  - [Metrics Glossary](#appendix-metrics-glossary--interpretation-guide)
66
  - [Visualizations Index](#visualizations-index)
@@ -80,35 +90,35 @@ We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and
80
 
81
  | Vocab Size | Compression | Avg Token Len | UNK Rate | Total Tokens |
82
  |------------|-------------|---------------|----------|--------------|
83
- | **8k** | 3.182x 🏆 | 3.19 | 2.9567% | 6,629 |
84
 
85
  ### Tokenization Examples
86
 
87
  Below are sample sentences tokenized with each vocabulary size:
88
 
89
- **Sample 1:** `ᐊᓐ ᐊᒋᐦᑖᓱᓐ ᐯᔭᒄ ᐃᔑᓂᐦᑳᑌᒡ, ᐋᐸᑎᓐ ᐃᑣᓅᐦᒡ ᐯᔭᒄ ᒉᒀᓐ ᒫᒃ ᐊᐌᓐ᙮ ᐊᓐ ᒫᒃ ᐊᒋᐦᑖᓱᓐ ᐯᔭᒄ, ᐁᐅᑯᓐ ᓃ...`
90
 
91
  | Vocab | Tokens | Count |
92
  |-------|--------|-------|
93
- | 8k | `▁ᐊᓐ ▁ᐊᒋᐦᑖᓱᓐ ▁ᐯᔭᒄ ▁ᑲ ▁ᐃᔑᓂᐦᑳᑌᒡ , ▁ᐋᐸᑎᓐ ▁ᒉ ▁ᒌ ▁ᐃᑣᓅᐦᒡ ... (+19 more)` | 29 |
94
 
95
- **Sample 2:** `ᓀᐦᐃᔭᐁᐧᐃᐧᐣ ᑕᐣᓯᐃᓯᐲᑭᐢᑫᐧᕁ ᓵᓴᕀ ᐳᓂ ᐱᑭᐢᑫᐧᐃᐧᐣ ᐱᐦᒑᔨᕁ ᑳᓇᑕ. ᓵᓴᕀ ᐳᓂ ᐱᑭᐢᑫᐧᐃᐧᐣ ᓇᐊᐧᐨ ᐳᑯ ᒌᑳᐦᑕ...`
96
 
97
  | Vocab | Tokens | Count |
98
  |-------|--------|-------|
99
- | 8k | `▁ᓀᐦᐃᔭᐁᐧᐃᐧᐣ ▁ᑕᐣᓯ ▁ᑲ ▁ᐃᓯᐲᑭᐢᑫᐧᕁ ▁ᓵᓴᕀ ▁ᐳᓂ ▁ᐱᑭᐢᑫᐧᐃᐧᐣ ▁ᐱᐦᒑᔨᕁ ▁ᑳᓇᑕ . ... (+11 more)` | 21 |
100
 
101
- **Sample 3:** `ᒨᔅ, Muus, Mush ( ; ) n.a. ᐊᐧᐁᓰᔅ ᐆ᙮ ᒨᔅ ᒥᐦᒑᐱᔅᒋᓲ᙮ ᓂᒥᑕᐦᐊᒻ ᑲᔦᐦ᙮ ᐸᐹᒦᒋᓲ᙮ ᒨᒥᓀᐤ᙮ ᒨᔅ ᒦᒎ ᓂᐦ...`
102
 
103
  | Vocab | Tokens | Count |
104
  |-------|--------|-------|
105
- | 8k | `▁ᒨᔅ , muus , ▁mush( ▁; ▁) ▁n . ... (+17 more)` | 27 |
106
 
107
 
108
  ### Key Findings
109
 
110
- - **Best Compression:** 8k achieves 3.182x compression
111
- - **Lowest UNK Rate:** 8k with 2.9567% unknown tokens
112
  - **Trade-off:** Larger vocabularies improve compression but increase model size
113
  - **Recommendation:** 32k vocabulary provides optimal balance for production use
114
 
@@ -126,11 +136,13 @@ Below are sample sentences tokenized with each vocabulary size:
126
  | N-gram | Variant | Perplexity | Entropy | Unique N-grams | Top-100 Coverage | Top-1000 Coverage |
127
  |--------|---------|------------|---------|----------------|------------------|-------------------|
128
  | **2-gram** | Word | 16 | 4.04 | 17 | 100.0% | 100.0% |
129
- | **2-gram** | Subword | 492 | 8.94 | 848 | 48.2% | 100.0% |
130
  | **3-gram** | Word | 15 🏆 | 3.88 | 16 | 100.0% | 100.0% |
131
- | **3-gram** | Subword | 1,528 | 10.58 | 1,986 | 19.4% | 75.4% |
132
- | **4-gram** | Word | 163 | 7.35 | 166 | 62.1% | 100.0% |
133
- | **4-gram** | Subword | 3,131 | 11.61 | 3,878 | 11.9% | 50.9% |
 
 
134
 
135
  ### Top 5 N-grams by Size
136
 
@@ -162,23 +174,33 @@ Below are sample sentences tokenized with each vocabulary size:
162
  | 2 | `in standard roman orthography` | 5 |
163
  | 3 | `written in standard roman` | 5 |
164
  | 4 | `ᑎᐸᐦᐄᑲᓐ ᑎᐸᐦᐄᑲᓐ ᑎᐸᐦᐄᑲᓐ ᑎᐸᐦᐄᑲᓐ` | 4 |
165
- | 5 | `of articles some articles` | 3 |
 
 
 
 
 
 
 
 
 
 
166
 
167
  **2-grams (Subword):**
168
 
169
  | Rank | N-gram | Count |
170
  |------|--------|-------|
171
- | 1 | `i n` | 215 |
172
- | 2 | `, _` | 213 |
173
- | 3 | `_ ᐊ` | 179 |
174
- | 4 | `i k` | 168 |
175
- | 5 | `n _` | 165 |
176
 
177
  **3-grams (Subword):**
178
 
179
  | Rank | N-gram | Count |
180
  |------|--------|-------|
181
- | 1 | `i n _` | 61 |
182
  | 2 | `a n i` | 49 |
183
  | 3 | `w i n` | 48 |
184
  | 4 | `_ k i` | 47 |
@@ -190,16 +212,26 @@ Below are sample sentences tokenized with each vocabulary size:
190
  |------|--------|-------|
191
  | 1 | `w a k _` | 33 |
192
  | 2 | `w i n _` | 27 |
193
- | 3 | `t i o n` | 24 |
194
- | 4 | `k a n i` | 23 |
195
- | 5 | `i k a n` | 22 |
 
 
 
 
 
 
 
 
 
 
196
 
197
 
198
  ### Key Findings
199
 
200
  - **Best Perplexity:** 3-gram (word) with 15
201
  - **Entropy Trend:** Decreases with larger n-grams (more predictable)
202
- - **Coverage:** Top-1000 patterns cover ~51% of corpus
203
  - **Recommendation:** 4-gram or 5-gram for best predictive performance
204
 
205
  ---
@@ -215,14 +247,14 @@ Below are sample sentences tokenized with each vocabulary size:
215
 
216
  | Context | Variant | Avg Entropy | Perplexity | Branching Factor | Unique Contexts | Predictability |
217
  |---------|---------|-------------|------------|------------------|-----------------|----------------|
218
- | **1** | Word | 0.2827 | 1.216 | 1.47 | 1,787 | 71.7% |
219
- | **1** | Subword | 1.9100 | 3.758 | 10.53 | 273 | 0.0% |
220
- | **2** | Word | 0.0424 | 1.030 | 1.05 | 2,607 | 95.8% |
221
- | **2** | Subword | 0.6919 | 1.615 | 2.63 | 2,872 | 30.8% |
222
- | **3** | Word | 0.0178 | 1.012 | 1.02 | 2,724 | 98.2% |
223
- | **3** | Subword | 0.3559 | 1.280 | 1.57 | 7,557 | 64.4% |
224
- | **4** | Word | 0.0086 🏆 | 1.006 | 1.01 | 2,765 | 99.1% |
225
- | **4** | Subword | 0.1591 | 1.117 | 1.21 | 11,842 | 84.1% |
226
 
227
  ### Generated Text Samples (Word-based)
228
 
@@ -230,27 +262,27 @@ Below are text samples generated from each word-based Markov chain model:
230
 
231
  **Context Size 1:**
232
 
233
- 1. `ᐁ ᐊᐧᐃᐢᑮᐦᐃᑲᐣ ᐃᐧᐊ ᐁᔥᐃᐦᑕᒧᐃᐧᐣ ᐋᐸᐦᐄᔥᑌᒡ english and montana some articles in iyuw iyimuun natuashish dia...`
234
- 2. `e kašcihot e wîcit e iskwewit mâk atimwa wes namawîy nataweyihtam cecî cisceyihtâkwaniyic ekw wenâpe...`
235
- 3. `of nonkilling channel on l nehirâmowin qc r s t u v w ᐎ ᐒ`
236
 
237
  **Context Size 2:**
238
 
239
- 1. `some articles in ininiwi išikišwēwin eastern dialect la romaine mingan natashquan pakuashipi and she...`
240
- 2. `articles in lehlueun western dialect list of articles some articles in nīhithawīwin list of articles...`
241
- 3. `ēkwa mīna otaskānitik e ka naskahtamēw nikiskihcēta anihi tahki itēhk pēhtahkik tānpahtiwin ē mic...`
242
 
243
  **Context Size 3:**
244
 
245
- 1. `some articles in nēhiyawēwin âpihtâkosisânak isiwepahki maskisin ᐸᐦᑵᓯᑲᐣ pimîhkân tipahikan itasin...`
246
- 2. `list of articles wikipedias in other native american languages atikamekw avañe aymar choctaw ꮳꮃꭹ c...`
247
- 3. `dialect list of articles some articles in ililîmowin list of articles ᐃᓕᓖᒧᐎᓐ ililîmowin ililîmowin p...`
248
 
249
  **Context Size 4:**
250
 
251
- 1. `dialect list of articles some articles in iyuw iyimuun kawawachikamach dialect list of articles nīhi...`
252
  2. `written in standard roman orthography`
253
- 3. `ᑎᐸᐦᐄᑲᓐ ᑎᐸᐦᐄᑲᓐ ᑎᐸᐦᐄᑲᓐ ᑎᐸᐦᐄᑲᓐ ᑎᐸᐦᐄᑲᓐ ᑎᐸᐦᐄᑲᓐ ᐋᐱᐦᑖᒌᔑᑳᐤ ᐋᐱᐦᑖᑎᐱᔅᑳᐤ 1 05 ᐯᔭᒄ ᑎᐸᐦᐄᑲᓐ ᒦᓐ ᓂᔮᔪ ᒥᓂᑯᔥ ᓂᔮᔪ ᒥᓂᑯᔥ ᒥᔮ...`
254
 
255
 
256
  ### Generated Text Samples (Subword-based)
@@ -259,34 +291,34 @@ Below are text samples generated from each subword-based Markov chain model:
259
 
260
  **Context Size 1:**
261
 
262
- 1. `_ᒉᒀᓐᓂᓕᐅᕝᕙᓪᓗ_ᐃᓐᓂᓂ`
263
- 2. `ik;_ᑲᐤ_(_ᑕᐦᐁᐧᔭᐍᑎ`
264
- 3. `am_ᐁr_ē-nata_ost`
265
 
266
  **Context Size 2:**
267
 
268
- 1. `inēhiyiy-âyot_ayi`
269
- 2. `,_ᐱᔪᓐᓇᖅ_ᖂᑉ_ᒪᓕᒋᐊᓕᖕ`
270
- 3. `_ᐊᑕᐦᑐᒥᒃ_ᑐᒃᓯᓪᓗᓂ_ᐊᓯ`
271
 
272
  **Context Size 3:**
273
 
274
- 1. `in_nešt_mâk_ekwa_a`
275
- 2. `anininisiniw._pask`
276
- 3. `win_okiskān_tipēna`
277
 
278
  **Context Size 4:**
279
 
280
- 1. `wak_tāpihikan_ᐆᒪ_ᐊᐢ`
281
- 2. `win_(statistics_(10`
282
- 3. `tion_métis_federati`
283
 
284
 
285
  ### Key Findings
286
 
287
  - **Best Predictability:** Context-4 (word) with 99.1% predictability
288
  - **Branching Factor:** Decreases with context size (more deterministic)
289
- - **Memory Trade-off:** Larger contexts require more storage (11,842 contexts)
290
  - **Recommendation:** Context-3 or Context-4 for text generation
291
 
292
  ---
@@ -302,9 +334,9 @@ Below are text samples generated from each subword-based Markov chain model:
302
 
303
  | Metric | Value |
304
  |--------|-------|
305
- | Vocabulary Size | 489 |
306
- | Total Tokens | 1,731 |
307
- | Mean Frequency | 3.54 |
308
  | Median Frequency | 2 |
309
  | Frequency Std Dev | 3.40 |
310
 
@@ -312,11 +344,11 @@ Below are text samples generated from each subword-based Markov chain model:
312
 
313
  | Rank | Word | Frequency |
314
  |------|------|-----------|
315
- | 1 | ᐁ | 34 |
316
  | 2 | e | 30 |
317
  | 3 | and | 22 |
318
- | 4 | in | 22 |
319
- | 5 | of | 22 |
320
  | 6 | pîsim | 19 |
321
  | 7 | articles | 18 |
322
  | 8 | cree | 16 |
@@ -327,39 +359,39 @@ Below are text samples generated from each subword-based Markov chain model:
327
 
328
  | Rank | Word | Frequency |
329
  |------|------|-----------|
330
- | 1 | ᐸᑦᑕᖕᓂᑦ | 2 |
331
- | 2 | ordinateur | 2 |
332
- | 3 | demandez | 2 |
333
- | 4 | le | 2 |
334
- | 5 | programme | 2 |
335
- | 6 | eurêka | 2 |
336
- | 7 | culture | 2 |
337
- | 8 | 18 | 2 |
338
- | 9 | août | 2 |
339
  | 10 | ᖃᐅᔨᓴᖅᑎᐅᔪᓄᑦ | 2 |
340
 
341
  ### Zipf's Law Analysis
342
 
343
  | Metric | Value |
344
  |--------|-------|
345
- | Zipf Coefficient | 0.5522 |
346
- | R² (Goodness of Fit) | 0.947702 |
347
  | Adherence Quality | **excellent** |
348
 
349
  ### Coverage Analysis
350
 
351
  | Top N Words | Coverage |
352
  |-------------|----------|
353
- | Top 100 | 47.6% |
354
  | Top 1,000 | 0.0% |
355
  | Top 5,000 | 0.0% |
356
  | Top 10,000 | 0.0% |
357
 
358
  ### Key Findings
359
 
360
- - **Zipf Compliance:** R²=0.9477 indicates excellent adherence to Zipf's law
361
- - **High Frequency Dominance:** Top 100 words cover 47.6% of corpus
362
- - **Long Tail:** -9,511 words needed for remaining 100.0% coverage
363
 
364
  ---
365
  ## 5. Word Embeddings Evaluation
@@ -375,37 +407,38 @@ Below are text samples generated from each subword-based Markov chain model:
375
 
376
  ### 5.1 Cross-Lingual Alignment
377
 
378
- > *Note: Multilingual alignment visualization not available for this language.*
379
 
380
 
381
  ### 5.2 Model Comparison
382
 
383
  | Model | Dimension | Isotropy | Semantic Density | Alignment R@1 | Alignment R@10 |
384
  |-------|-----------|----------|------------------|---------------|----------------|
385
- | **mono_32d** | 32 | 0.0381 🏆 | 0.0000 | N/A | N/A |
386
- | **mono_64d** | 64 | 0.0033 | 0.0000 | N/A | N/A |
387
  | **mono_128d** | 128 | 0.0000 | 0.0000 | N/A | N/A |
 
 
 
388
 
389
  ### Key Findings
390
 
391
- - **Best Isotropy:** mono_32d with 0.0381 (more uniform distribution)
392
  - **Semantic Density:** Average pairwise similarity of 0.0000. Lower values indicate better semantic separation.
393
- - **Alignment Quality:** No aligned models evaluated in this run.
394
  - **Recommendation:** 128d aligned for best cross-lingual performance
395
 
396
  ---
397
  ## 6. Morphological Analysis (Experimental)
398
 
399
- > ⚠️ **Warning:** This language shows low morphological productivity. The statistical signals used for this analysis may be noisy or less reliable than for morphologically rich languages.
400
-
401
  This section presents an automated morphological analysis derived from the statistical divergence between word-level and subword-level models. By analyzing where subword predictability spikes and where word-level coverage fails, we can infer linguistic structures without supervised data.
402
 
403
  ### 6.1 Productivity & Complexity
404
 
405
  | Metric | Value | Interpretation | Recommendation |
406
  |--------|-------|----------------|----------------|
407
- | Productivity Index | **0.000** | Low morphological productivity | ⚠️ Likely unreliable |
408
- | Idiomaticity Gap | **-1.000** | Low formulaic content | - |
409
 
410
  ### 6.2 Affix Inventory (Productive Units)
411
 
@@ -438,7 +471,9 @@ Using **Recursive Hierarchical Substitutability**, we decompose complex words in
438
  ### 6.6 Linguistic Interpretation
439
 
440
  > **Automated Insight:**
441
- The language CR appears to be more isolating or has a highly fixed vocabulary. Word-level models perform nearly as well as subword models, indicating fewer productive morphological processes.
 
 
442
 
443
  ---
444
  ## 7. Summary & Recommendations
@@ -449,7 +484,7 @@ The language CR appears to be more isolating or has a highly fixed vocabulary. W
449
 
450
  | Component | Recommended | Rationale |
451
  |-----------|-------------|-----------|
452
- | Tokenizer | **8k BPE** | Best compression (3.18x) |
453
  | N-gram | **3-gram** | Lowest perplexity (15) |
454
  | Markov | **Context-4** | Highest predictability (99.1%) |
455
  | Embeddings | **100d** | Balanced semantic capture and isotropy |
@@ -665,4 +700,4 @@ MIT License - Free for academic and commercial use.
665
  ---
666
  *Generated by Wikilangs Models Pipeline*
667
 
668
- *Report Date: 2026-01-03 10:19:03*
 
1
  ---
2
  language: cr
3
+ language_name: Cree
4
  language_family: american_algonquian
5
  tags:
6
  - wikilangs
 
10
  - n-gram
11
  - markov
12
  - wikipedia
13
+ - feature-extraction
14
+ - sentence-similarity
15
+ - tokenization
16
+ - n-grams
17
+ - markov-chain
18
+ - text-mining
19
+ - fasttext
20
+ - babelvec
21
+ - vocabulous
22
+ - vocabulary
23
  - monolingual
24
  - family-american_algonquian
25
  license: mit
26
  library_name: wikilangs
27
+ pipeline_tag: text-generation
28
  datasets:
29
  - omarkamali/wikipedia-monthly
30
  dataset_info:
 
33
  metrics:
34
  - name: best_compression_ratio
35
  type: compression
36
+ value: 3.238
37
  - name: best_isotropy
38
  type: isotropy
39
+ value: 0.0354
40
  - name: vocabulary_size
41
  type: vocab
42
  value: 0
43
  generated: 2026-01-03
44
  ---
45
 
46
+ # Cree - Wikilangs Models
47
  ## Comprehensive Research Report & Full Ablation Study
48
 
49
+ This repository contains NLP models trained and evaluated by Wikilangs, specifically on **Cree** Wikipedia data.
50
  We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and word embeddings.
51
 
52
  ## 📋 Repository Contents
 
70
  - [3. Markov Chain Evaluation](#3-markov-chain-evaluation)
71
  - [4. Vocabulary Analysis](#4-vocabulary-analysis)
72
  - [5. Word Embeddings Evaluation](#5-word-embeddings-evaluation)
73
+ - [6. Morphological Analysis (Experimental)](#6--morphological-analysis-experimental)
74
  - [7. Summary & Recommendations](#7-summary--recommendations)
75
  - [Metrics Glossary](#appendix-metrics-glossary--interpretation-guide)
76
  - [Visualizations Index](#visualizations-index)
 
90
 
91
  | Vocab Size | Compression | Avg Token Len | UNK Rate | Total Tokens |
92
  |------------|-------------|---------------|----------|--------------|
93
+ | **8k** | 3.238x 🏆 | 3.24 | 2.7764% | 6,267 |
94
 
95
  ### Tokenization Examples
96
 
97
  Below are sample sentences tokenized with each vocabulary size:
98
 
99
+ **Sample 1:** `ᓀᐦᐃᔭᐁᐧᐃᐧᐣ ᑕᐣᓯᐃᓯᐲᑭᐢᑫᐧᕁ ᓵᓴᕀ ᐳᓂ ᐱᑭᐢᑫᐧᐃᐧᐣ ᐱᐦᒑᔨᕁ ᑳᓇᑕ. ᓵᓴᕀ ᐳᓂ ᐱᑭᐢᑫᐧᐃᐧᐣ ᓇᐊᐧᐨ ᐳᑯ ᒌᑳᐦᑕ...`
100
 
101
  | Vocab | Tokens | Count |
102
  |-------|--------|-------|
103
+ | 8k | `▁ᓀᐦᐃᔭᐁᐧᐃᐧᐣ ▁ᑕᐣᓯ ▁ᑲ ▁ᐃᓯᐲᑭᐢᑫᐧᕁ ▁ᓵᓴᕀ ▁ᐳᓂ ▁ᐱᑭᐢᑫᐧᐃᐧᐣ ▁ᐱᐦᒑᔨᕁ ▁ᑳᓇᑕ . ... (+11 more)` | 21 |
104
 
105
+ **Sample 2:** `ᐊᓐ ᐊᒋᐦᑖᓱᓐ ᐯᔭᒄ ᐃᔑᓂᐦᑳᑌᒡ, ᐋᐸᑎᓐ ᐃᑣᓅᐦᒡ ᐯᔭᒄ ᒉᒀᓐ ᒫᒃ ᐊᐌᓐ᙮ ᐊᓐ ᒫᒃ ᐊᒋᐦᑖᓱᓐ ᐯᔭᒄ, ᐁᐅᑯᓐ ᓃ...`
106
 
107
  | Vocab | Tokens | Count |
108
  |-------|--------|-------|
109
+ | 8k | `▁ᐊᓐ ▁ᐊᒋᐦᑖᓱᓐ ▁ᐯᔭᒄ ▁ᑲ ▁ᐃᔑᓂᐦᑳᑌᒡ , ▁ᐋᐸᑎᓐ ▁ᒉ ▁ᒌ ▁ᐃᑣᓅᐦᒡ ... (+19 more)` | 29 |
110
 
111
+ **Sample 3:** `ᒦᒃᓰᖂ (english : Mexico) ᐊᐢᑭᐩ ᑮᐍᑎᐣ ᐊᒣᕒᐃᑲ ᐆᐦᒋ᙮ ᐊᔨᓯᔨᓂᐘᐠ ᐑᑭᐘᐠ ᐆᒪ ᐊᐢᑭᔭᕽ᙮ </center>`
112
 
113
  | Vocab | Tokens | Count |
114
  |-------|--------|-------|
115
+ | 8k | `▁ᒦᒃᓰᖂ( english ▁:mexico ) ▁ᐊᐢᑭᐩ ▁ᑮᐍᑎᐣ ▁ᐊᒣᕒᐃᑲ ▁ᐆᐦᒋ᙮ ... (+7 more)` | 17 |
116
 
117
 
118
  ### Key Findings
119
 
120
+ - **Best Compression:** 8k achieves 3.238x compression
121
+ - **Lowest UNK Rate:** 8k with 2.7764% unknown tokens
122
  - **Trade-off:** Larger vocabularies improve compression but increase model size
123
  - **Recommendation:** 32k vocabulary provides optimal balance for production use
124
 
 
136
  | N-gram | Variant | Perplexity | Entropy | Unique N-grams | Top-100 Coverage | Top-1000 Coverage |
137
  |--------|---------|------------|---------|----------------|------------------|-------------------|
138
  | **2-gram** | Word | 16 | 4.04 | 17 | 100.0% | 100.0% |
139
+ | **2-gram** | Subword | 473 | 8.89 | 812 | 49.1% | 100.0% |
140
  | **3-gram** | Word | 15 🏆 | 3.88 | 16 | 100.0% | 100.0% |
141
+ | **3-gram** | Subword | 1,468 | 10.52 | 1,902 | 19.8% | 76.9% |
142
+ | **4-gram** | Word | 157 | 7.29 | 160 | 64.3% | 100.0% |
143
+ | **4-gram** | Subword | 2,988 | 11.54 | 3,702 | 12.2% | 52.2% |
144
+ | **5-gram** | Word | 137 | 7.10 | 138 | 73.1% | 100.0% |
145
+ | **5-gram** | Subword | 2,771 | 11.44 | 3,264 | 12.2% | 51.7% |
146
 
147
  ### Top 5 N-grams by Size
148
 
 
174
  | 2 | `in standard roman orthography` | 5 |
175
  | 3 | `written in standard roman` | 5 |
176
  | 4 | `ᑎᐸᐦᐄᑲᓐ ᑎᐸᐦᐄᑲᓐ ᑎᐸᐦᐄᑲᓐ ᑎᐸᐦᐄᑲᓐ` | 4 |
177
+ | 5 | `center for global nonkilling` | 3 |
178
+
179
+ **5-grams (Word):**
180
+
181
+ | Rank | N-gram | Count |
182
+ |------|--------|-------|
183
+ | 1 | `written in standard roman orthography` | 5 |
184
+ | 2 | `list of articles some articles` | 3 |
185
+ | 3 | `of articles some articles in` | 3 |
186
+ | 4 | `dialect list of articles some` | 3 |
187
+ | 5 | `ᑎᐸᐦᐄᑲᓐ ᑎᐸᐦᐄᑲᓐ ᑎᐸᐦᐄᑲᓐ ᑎᐸᐦᐄᑲᓐ ᑎᐸᐦᐄᑲᓐ` | 3 |
188
 
189
  **2-grams (Subword):**
190
 
191
  | Rank | N-gram | Count |
192
  |------|--------|-------|
193
+ | 1 | `i n` | 207 |
194
+ | 2 | `, _` | 202 |
195
+ | 3 | `i k` | 169 |
196
+ | 4 | `_ ᐊ` | 164 |
197
+ | 5 | `i s` | 159 |
198
 
199
  **3-grams (Subword):**
200
 
201
  | Rank | N-gram | Count |
202
  |------|--------|-------|
203
+ | 1 | `i n _` | 58 |
204
  | 2 | `a n i` | 49 |
205
  | 3 | `w i n` | 48 |
206
  | 4 | `_ k i` | 47 |
 
212
  |------|--------|-------|
213
  | 1 | `w a k _` | 33 |
214
  | 2 | `w i n _` | 27 |
215
+ | 3 | `k a n i` | 23 |
216
+ | 4 | `t i o n` | 23 |
217
+ | 5 | `_ o f _` | 22 |
218
+
219
+ **5-grams (Subword):**
220
+
221
+ | Rank | N-gram | Count |
222
+ |------|--------|-------|
223
+ | 1 | `_ a n d _` | 22 |
224
+ | 2 | `a t i o n` | 21 |
225
+ | 3 | `p î s i m` | 20 |
226
+ | 4 | `- p î s i` | 19 |
227
+ | 5 | `a r t i c` | 19 |
228
 
229
 
230
  ### Key Findings
231
 
232
  - **Best Perplexity:** 3-gram (word) with 15
233
  - **Entropy Trend:** Decreases with larger n-grams (more predictable)
234
+ - **Coverage:** Top-1000 patterns cover ~52% of corpus
235
  - **Recommendation:** 4-gram or 5-gram for best predictive performance
236
 
237
  ---
 
247
 
248
  | Context | Variant | Avg Entropy | Perplexity | Branching Factor | Unique Contexts | Predictability |
249
  |---------|---------|-------------|------------|------------------|-----------------|----------------|
250
+ | **1** | Word | 0.2841 | 1.218 | 1.47 | 1,711 | 71.6% |
251
+ | **1** | Subword | 1.8933 | 3.715 | 10.31 | 271 | 0.0% |
252
+ | **2** | Word | 0.0442 | 1.031 | 1.05 | 2,501 | 95.6% |
253
+ | **2** | Subword | 0.6883 | 1.611 | 2.62 | 2,789 | 31.2% |
254
+ | **3** | Word | 0.0186 | 1.013 | 1.02 | 2,617 | 98.1% |
255
+ | **3** | Subword | 0.3514 | 1.276 | 1.56 | 7,299 | 64.9% |
256
+ | **4** | Word | 0.0089 🏆 | 1.006 | 1.01 | 2,657 | 99.1% |
257
+ | **4** | Subword | 0.1579 | 1.116 | 1.21 | 11,392 | 84.2% |
258
 
259
  ### Generated Text Samples (Word-based)
260
 
 
262
 
263
  **Context Size 1:**
264
 
265
+ 1. `ᐁ p q r s ᓰ`
266
+ 2. `e kiskatcik e tašitwâw awesîsac sašimuve nîštam atim nâpeštimw išinihkâtâkaniwiw simpohanin âtayôhkâ...`
267
+ 3. `of articles in ininiwi išikišwēwin eastern dialect western montagnais iso 639 crk location québec an...`
268
 
269
  **Context Size 2:**
270
 
271
+ 1. `some articles in nēhiyawēwin âpihtâkosisânak isiwepahki maskisin ᐸᐦᑵᓯᑲᐣ pimîhkân tipahikan itasin...`
272
+ 2. `articles in iyuw iyimuun natuashish dialect list of articles ᐃᔨᔨᐤ ᐊᔨᒧᐧᐃᓐ iyyû ayimuwin nēhiyawēwin p...`
273
+ 3. `list of articles ᐃᔨᔨᐤ ᐊᔨᒧᐧᐃᓐ iyyû ayimuwin northern dialect chisasibi eastmain waskaganish wemindji ...`
274
 
275
  **Context Size 3:**
276
 
277
+ 1. `some articles in lehlueun western dialect betsiamites mashteuiatsh matimekosh and uashat maliotenam ...`
278
+ 2. `list of articles ᐃᓕᓖᒧᐎᓐ ililîmowin ililîmowin portal english name woods cree iso 639 crk location sa...`
279
+ 3. `dialect list of articles ᐃᓕᓖᒧᐎᓐ ililîmowin ililîmowin portal english name moose cree iso 639 csw loc...`
280
 
281
  **Context Size 4:**
282
 
283
+ 1. `dialect list of articles nīhithawīwin portal english name woods cree iso 639 cwd location manitoba a...`
284
  2. `written in standard roman orthography`
285
+ 3. `ᑎᐸᐦᐄᑲᓐ ᑎᐸᐦᐄᑲᓐ ᑎᐸᐦᐄᑲᓐ ᑎᐸᐦᐄᑲᓐ ᐋᐱᐦᑖᒌᔑᑳᐤ ᐋᐱᐦᑖᑎᐱᔅᑳᐤ 1 05 ᐯᔭᒄ ᑎᐸᐦᐄᑲᓐ ᒦᓐ ᓂᔮᔪ ᒥᓂᑯᔥ ᓂᔮᔪ ᒥᓂᑯᔥ ᒥᔮᐧᐃᐸᔩᐤ ᐯᔭᒄ 1 30`
286
 
287
 
288
  ### Generated Text Samples (Subword-based)
 
291
 
292
  **Context Size 1:**
293
 
294
+ 1. `_ck.._ntahkwiwre`
295
+ 2. `iw._ey_îskānakat`
296
+ 3. `asuét):_ᓅᐦᑭᑫᓂᐤ..`
297
 
298
  **Context Size 2:**
299
 
300
+ 1. `initahtâw._ᑭᒋᒧᐏᐣ_`
301
+ 2. `,_miyis_nawamēwik`
302
+ 3. `ikawahtawāt_kin_o`
303
 
304
  **Context Size 3:**
305
 
306
+ 1. `in_itakwa_é-nipaho`
307
+ 2. `anitināw_ōnahkân_a`
308
+ 3. `winaka_kikamîw-sîp`
309
 
310
  **Context Size 4:**
311
 
312
+ 1. `wak_*`
313
+ 2. `win_ᐊᑎᒽ_ᐯᔭᒄ_ᓀᐦᐃᔭᐍᐏᐣ`
314
+ 3. `tion:_saskapi_qc_y_`
315
 
316
 
317
  ### Key Findings
318
 
319
  - **Best Predictability:** Context-4 (word) with 99.1% predictability
320
  - **Branching Factor:** Decreases with context size (more deterministic)
321
+ - **Memory Trade-off:** Larger contexts require more storage (11,392 contexts)
322
  - **Recommendation:** Context-3 or Context-4 for text generation
323
 
324
  ---
 
334
 
335
  | Metric | Value |
336
  |--------|-------|
337
+ | Vocabulary Size | 468 |
338
+ | Total Tokens | 1,673 |
339
+ | Mean Frequency | 3.57 |
340
  | Median Frequency | 2 |
341
  | Frequency Std Dev | 3.40 |
342
 
 
344
 
345
  | Rank | Word | Frequency |
346
  |------|------|-----------|
347
+ | 1 | ᐁ | 31 |
348
  | 2 | e | 30 |
349
  | 3 | and | 22 |
350
+ | 4 | of | 22 |
351
+ | 5 | in | 21 |
352
  | 6 | pîsim | 19 |
353
  | 7 | articles | 18 |
354
  | 8 | cree | 16 |
 
359
 
360
  | Rank | Word | Frequency |
361
  |------|------|-----------|
362
+ | 1 | ᑯᓐᓄᑦ | 2 |
363
+ | 2 | ᐊᒻᒪᐃᓛᒃ | 2 |
364
+ | 3 | ᐊᑎᕐᒥᒃ | 2 |
365
+ | 4 | ᖃᕆᑕᐅᔭᕐᒧᑦ | 2 |
366
+ | 5 | ᐅᖃᐅᓯᕐᒥᒃ | 2 |
367
+ | 6 | ᐊᔾᔨᐅᖏᑦᑐᒥᒃ | 2 |
368
+ | 7 | ᑖᓐᓇ | 2 |
369
+ | 8 | ᑕᐃᓐᓇ | 2 |
370
+ | 9 | ᖃᕆᑕᐅᔭᒃᑯᑦ | 2 |
371
  | 10 | ᖃᐅᔨᓴᖅᑎᐅᔪᓄᑦ | 2 |
372
 
373
  ### Zipf's Law Analysis
374
 
375
  | Metric | Value |
376
  |--------|-------|
377
+ | Zipf Coefficient | 0.5578 |
378
+ | R² (Goodness of Fit) | 0.947960 |
379
  | Adherence Quality | **excellent** |
380
 
381
  ### Coverage Analysis
382
 
383
  | Top N Words | Coverage |
384
  |-------------|----------|
385
+ | Top 100 | 48.8% |
386
  | Top 1,000 | 0.0% |
387
  | Top 5,000 | 0.0% |
388
  | Top 10,000 | 0.0% |
389
 
390
  ### Key Findings
391
 
392
+ - **Zipf Compliance:** R²=0.9480 indicates excellent adherence to Zipf's law
393
+ - **High Frequency Dominance:** Top 100 words cover 48.8% of corpus
394
+ - **Long Tail:** -9,532 words needed for remaining 100.0% coverage
395
 
396
  ---
397
  ## 5. Word Embeddings Evaluation
 
407
 
408
  ### 5.1 Cross-Lingual Alignment
409
 
410
+ ![Multilingual t-SNE](visualizations/embedding_tsne_multilingual.png)
411
 
412
 
413
  ### 5.2 Model Comparison
414
 
415
  | Model | Dimension | Isotropy | Semantic Density | Alignment R@1 | Alignment R@10 |
416
  |-------|-----------|----------|------------------|---------------|----------------|
417
+ | **mono_32d** | 32 | 0.0354 | 0.0000 | N/A | N/A |
418
+ | **mono_64d** | 64 | 0.0038 | 0.0000 | N/A | N/A |
419
  | **mono_128d** | 128 | 0.0000 | 0.0000 | N/A | N/A |
420
+ | **aligned_32d** | 32 | 0.0354 🏆 | 0.0000 | 0.0000 | 0.0000 |
421
+ | **aligned_64d** | 64 | 0.0038 | 0.0000 | 0.0000 | 0.0000 |
422
+ | **aligned_128d** | 128 | 0.0000 | 0.0000 | 0.0000 | 0.0000 |
423
 
424
  ### Key Findings
425
 
426
+ - **Best Isotropy:** aligned_32d with 0.0354 (more uniform distribution)
427
  - **Semantic Density:** Average pairwise similarity of 0.0000. Lower values indicate better semantic separation.
428
+ - **Alignment Quality:** Aligned models evaluated but achieved 0% recall.
429
  - **Recommendation:** 128d aligned for best cross-lingual performance
430
 
431
  ---
432
  ## 6. Morphological Analysis (Experimental)
433
 
 
 
434
  This section presents an automated morphological analysis derived from the statistical divergence between word-level and subword-level models. By analyzing where subword predictability spikes and where word-level coverage fails, we can infer linguistic structures without supervised data.
435
 
436
  ### 6.1 Productivity & Complexity
437
 
438
  | Metric | Value | Interpretation | Recommendation |
439
  |--------|-------|----------------|----------------|
440
+ | Productivity Index | **5.000** | High morphological productivity | Reliable analysis |
441
+ | Idiomaticity Gap | **0.933** | High formulaic/idiomatic content | - |
442
 
443
  ### 6.2 Affix Inventory (Productive Units)
444
 
 
471
  ### 6.6 Linguistic Interpretation
472
 
473
  > **Automated Insight:**
474
+ The language Cree shows high morphological productivity. The subword models are significantly more efficient than word models, suggesting a rich system of affixation or compounding.
475
+
476
+ > **Note on Idiomaticity:** The high Idiomaticity Gap suggests a large number of frequent multi-word expressions or formulaic sequences that are statistically distinct from their component parts.
477
 
478
  ---
479
  ## 7. Summary & Recommendations
 
484
 
485
  | Component | Recommended | Rationale |
486
  |-----------|-------------|-----------|
487
+ | Tokenizer | **8k BPE** | Best compression (3.24x) |
488
  | N-gram | **3-gram** | Lowest perplexity (15) |
489
  | Markov | **Context-4** | Highest predictability (99.1%) |
490
  | Embeddings | **100d** | Balanced semantic capture and isotropy |
 
700
  ---
701
  *Generated by Wikilangs Models Pipeline*
702
 
703
+ *Report Date: 2026-01-03 20:39:39*
models/embeddings/aligned/cr_128d.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5dacb587c7d15197d345442364b1889a8b7a457b453f7df9253e97673b4fb352
3
+ size 1024067754
models/embeddings/aligned/cr_128d.meta.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lang": "cr", "dim": 128, "max_seq_len": 512, "is_aligned": true}
models/embeddings/aligned/cr_128d.projection.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b6ebec184c4af672ce0f571958b634809f46f6a1ea2dc7e6f8f2e7f53555387
3
+ size 65664
models/embeddings/aligned/cr_128d_metadata.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "language": "cr",
3
+ "dimension": 128,
4
+ "version": "aligned",
5
+ "hub_language": "en",
6
+ "seed_vocab_size": 39,
7
+ "vocab_size": 65
8
+ }
models/embeddings/aligned/cr_32d.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e6a077873b64cad1912b66853d07d4f890bc6606e6b08083d4bfd3153121129
3
+ size 256017834
models/embeddings/aligned/cr_32d.meta.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lang": "cr", "dim": 32, "max_seq_len": 512, "is_aligned": true}
models/embeddings/aligned/cr_32d.projection.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b14cf327498f9c92f06cc00df8284bfbdfd03069966ad20b3ada72dc5ce02488
3
+ size 4224
models/embeddings/aligned/cr_32d_metadata.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "language": "cr",
3
+ "dimension": 32,
4
+ "version": "aligned",
5
+ "hub_language": "en",
6
+ "seed_vocab_size": 39,
7
+ "vocab_size": 65
8
+ }
models/embeddings/aligned/cr_64d.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd5960af306a49bc16d12139a057de3daae86c9d43e31642ee7dbd085eb553b7
3
+ size 512034474
models/embeddings/aligned/cr_64d.meta.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lang": "cr", "dim": 64, "max_seq_len": 512, "is_aligned": true}
models/embeddings/aligned/cr_64d.projection.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d846b3fae902e708babece2e6c2e3c62275450ddb7096ca337af53dcf63a96b6
3
+ size 16512
models/embeddings/aligned/cr_64d_metadata.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "language": "cr",
3
+ "dimension": 64,
4
+ "version": "aligned",
5
+ "hub_language": "en",
6
+ "seed_vocab_size": 39,
7
+ "vocab_size": 65
8
+ }
models/embeddings/monolingual/cr_128d.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:322799eab6ee84e7d28a0f76de24efb5b70f9e1fcd0eef88c9b61829ce272397
3
  size 1024067754
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5dacb587c7d15197d345442364b1889a8b7a457b453f7df9253e97673b4fb352
3
  size 1024067754
models/embeddings/monolingual/cr_32d.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:095584a1fc693fcab2c3fd7f0ac831d533933eb58ce565c3b082866e303db9f6
3
  size 256017834
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e6a077873b64cad1912b66853d07d4f890bc6606e6b08083d4bfd3153121129
3
  size 256017834
models/embeddings/monolingual/cr_64d.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fdd5c40a0696cfa8faa26c56bda982ab2a5efbbc0bc7d30fcea9c944b571b542
3
  size 512034474
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd5960af306a49bc16d12139a057de3daae86c9d43e31642ee7dbd085eb553b7
3
  size 512034474
models/subword_markov/cr_markov_ctx1_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d336a63f8ad5e082086c17ec41db170270495929a8dd340a6abcd2e5998bc0e6
3
- size 19362
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d58185c9c750453fec2b1be4bea145ca19c8154e9c1851c770fcfe60e0945619
3
+ size 19062
models/subword_markov/cr_markov_ctx1_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 1,
3
  "variant": "subword",
4
  "language": "cr",
5
- "unique_contexts": 273,
6
- "total_transitions": 21066
7
  }
 
2
  "context_size": 1,
3
  "variant": "subword",
4
  "language": "cr",
5
+ "unique_contexts": 271,
6
+ "total_transitions": 20269
7
  }
models/subword_markov/cr_markov_ctx2_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:88cd25c549a06a864e6af9c0a79d29d91815a704038e65f2ce7d2c821349c095
3
- size 54769
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f22b06562e4d08f3c27092ecc548a0cae5a9446cc2ae3e54f2073e8b8900e95
3
+ size 52808
models/subword_markov/cr_markov_ctx2_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 2,
3
  "variant": "subword",
4
  "language": "cr",
5
- "unique_contexts": 2872,
6
- "total_transitions": 21041
7
  }
 
2
  "context_size": 2,
3
  "variant": "subword",
4
  "language": "cr",
5
+ "unique_contexts": 2789,
6
+ "total_transitions": 20244
7
  }
models/subword_markov/cr_markov_ctx3_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:493e48b8388399c0693ca1644064c6aff29ebd8768e02dc362bd13a9beb9923c
3
- size 109170
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f2ad975814a144c1719a36efe4086171171e4415b69c07c0cfc7589d7b44408b
3
+ size 105153
models/subword_markov/cr_markov_ctx3_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 3,
3
  "variant": "subword",
4
  "language": "cr",
5
- "unique_contexts": 7557,
6
- "total_transitions": 21016
7
  }
 
2
  "context_size": 3,
3
  "variant": "subword",
4
  "language": "cr",
5
+ "unique_contexts": 7299,
6
+ "total_transitions": 20219
7
  }
models/subword_markov/cr_markov_ctx4_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:77ef47f052b70e08ed38cb704ef21e780c36aa884a719e5d4963f672dd6af637
3
- size 165104
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fbe0924ddd06b05b7652bd329b4be5d86de19b757a42631954cd9abc0a1910f8
3
+ size 158381
models/subword_markov/cr_markov_ctx4_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 4,
3
  "variant": "subword",
4
  "language": "cr",
5
- "unique_contexts": 11842,
6
- "total_transitions": 20991
7
  }
 
2
  "context_size": 4,
3
  "variant": "subword",
4
  "language": "cr",
5
+ "unique_contexts": 11392,
6
+ "total_transitions": 20194
7
  }
models/subword_ngram/cr_2gram_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4ec4f4bbb6a07a46393df0aae6d9e7264c3913c983048c199b187ccf1637509f
3
- size 10150
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:437302b4fde849c0752032e34257d718792b347ef32757c168db42d3025abc26
3
+ size 9822
models/subword_ngram/cr_2gram_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "n": 2,
3
  "variant": "subword",
4
  "language": "cr",
5
- "unique_ngrams": 848,
6
- "total_ngrams": 21066
7
  }
 
2
  "n": 2,
3
  "variant": "subword",
4
  "language": "cr",
5
+ "unique_ngrams": 812,
6
+ "total_ngrams": 20269
7
  }
models/subword_ngram/cr_3gram_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3a341e935103c45c5dd7be03811b9df507bbbcb941614f0a4d3949665defe332
3
- size 22181
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7980b5aac9cba90ebe083cb0bfd2b3416465c2a52b83cbcf3a73434e722de7a0
3
+ size 21356
models/subword_ngram/cr_3gram_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "n": 3,
3
  "variant": "subword",
4
  "language": "cr",
5
- "unique_ngrams": 1986,
6
- "total_ngrams": 21041
7
  }
 
2
  "n": 3,
3
  "variant": "subword",
4
  "language": "cr",
5
+ "unique_ngrams": 1902,
6
+ "total_ngrams": 20244
7
  }
models/subword_ngram/cr_4gram_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:36e06026aa7388f2d7e58a4f337a1ff541dac38696cfa56eaaaec940804c17a1
3
- size 46333
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2bcb275f3772e5f6b8bc913b0b382afa84dac97f36e76b7d03ee8a6db106e56
3
+ size 44325
models/subword_ngram/cr_4gram_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "n": 4,
3
  "variant": "subword",
4
  "language": "cr",
5
- "unique_ngrams": 3878,
6
- "total_ngrams": 21016
7
  }
 
2
  "n": 4,
3
  "variant": "subword",
4
  "language": "cr",
5
+ "unique_ngrams": 3702,
6
+ "total_ngrams": 20219
7
  }
models/subword_ngram/cr_5gram_subword.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:18b2af82413c2d5e0b7124bc18273388641db7427c102a2465c939e7d69ffc67
3
+ size 42412
models/subword_ngram/cr_5gram_subword_metadata.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "n": 5,
3
+ "variant": "subword",
4
+ "language": "cr",
5
+ "unique_ngrams": 3264,
6
+ "total_ngrams": 20194
7
+ }
models/tokenizer/cr_tokenizer_8k.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a0973f38231d8be6cc17f5bb11b4adcec87ca7e994a2d91d59f4a7f15ea4655f
3
- size 379309
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aafdacc6f2d991f561954af6f998ab1116a7904c0d6408e0e6f25d2ce7b6f625
3
+ size 379259
models/tokenizer/cr_tokenizer_8k.vocab CHANGED
The diff for this file is too large to render. See raw diff
 
models/vocabulary/cr_vocabulary.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c9a339a558f9a91e507cf4c6a9afd7657b8a650a75b8594311ae35047326f22c
3
- size 10707
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4da6445f4ebee272af36858a529222db9605c42d9f0ad70060cc98b59cfaf5ee
3
+ size 10298
models/vocabulary/cr_vocabulary_metadata.json CHANGED
@@ -1,15 +1,15 @@
1
  {
2
  "language": "cr",
3
- "vocabulary_size": 489,
4
  "variant": "full",
5
  "statistics": {
6
- "type_token_ratio": 0.5915817165406116,
7
  "coverage": {
8
- "top_100": 0.2709634988490628,
9
- "top_1000": 0.7372574810917462
10
  },
11
- "hapax_count": 1310,
12
- "hapax_ratio": 0.7281823235130628,
13
  "total_documents": 25
14
  }
15
  }
 
1
  {
2
  "language": "cr",
3
+ "vocabulary_size": 468,
4
  "variant": "full",
5
  "statistics": {
6
+ "type_token_ratio": 0.5884562841530054,
7
  "coverage": {
8
+ "top_100": 0.2786885245901639,
9
+ "top_1000": 0.7530737704918032
10
  },
11
+ "hapax_count": 1255,
12
+ "hapax_ratio": 0.7283807312826466,
13
  "total_documents": 25
14
  }
15
  }
models/word_markov/cr_markov_ctx1_word.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1b9cad226d4f9247d3628f1a8d45e964b5ac7aee3588940f10b795c7db079363
3
- size 51656
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:67ec91aad046d21a2680c190c002d6bbdf3e246b18b9dffb01c01c60fdd45417
3
+ size 49544
models/word_markov/cr_markov_ctx1_word_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 1,
3
  "variant": "word",
4
  "language": "cr",
5
- "unique_contexts": 1787,
6
- "total_transitions": 3016
7
  }
 
2
  "context_size": 1,
3
  "variant": "word",
4
  "language": "cr",
5
+ "unique_contexts": 1711,
6
+ "total_transitions": 2903
7
  }
models/word_markov/cr_markov_ctx2_word.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b1f017105092b235e6c58f3c540d7b81c2ec84d6afd12ad9d42b7f06f32711cb
3
- size 68727
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da940dffbcc51f50754424866b5a37358c2ebc38b4236083ea24a6f1806004df
3
+ size 65639
models/word_markov/cr_markov_ctx2_word_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 2,
3
  "variant": "word",
4
  "language": "cr",
5
- "unique_contexts": 2607,
6
- "total_transitions": 2991
7
  }
 
2
  "context_size": 2,
3
  "variant": "word",
4
  "language": "cr",
5
+ "unique_contexts": 2501,
6
+ "total_transitions": 2878
7
  }
models/word_markov/cr_markov_ctx3_word.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ba438b38b527c374838426729a544c3af06986da131dcd7b6efc96c6174fdf64
3
- size 78351
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5fd696554eca1524675c807015f1565335c8e8d03c64184b7283aecaf4877ac
3
+ size 75064
models/word_markov/cr_markov_ctx3_word_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 3,
3
  "variant": "word",
4
  "language": "cr",
5
- "unique_contexts": 2724,
6
- "total_transitions": 2966
7
  }
 
2
  "context_size": 3,
3
  "variant": "word",
4
  "language": "cr",
5
+ "unique_contexts": 2617,
6
+ "total_transitions": 2853
7
  }
models/word_markov/cr_markov_ctx4_word.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:de165eca955a9457fb7d104a709440b37d8fe11272266a6eaf29d8aca13bcb96
3
- size 86120
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:241b5162742d910baedb70c153556b238227f4b436220e840d48f82f5b69800a
3
+ size 82423
models/word_markov/cr_markov_ctx4_word_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 4,
3
  "variant": "word",
4
  "language": "cr",
5
- "unique_contexts": 2765,
6
- "total_transitions": 2941
7
  }
 
2
  "context_size": 4,
3
  "variant": "word",
4
  "language": "cr",
5
+ "unique_contexts": 2657,
6
+ "total_transitions": 2828
7
  }
models/word_ngram/cr_2gram_word_metadata.json CHANGED
@@ -3,5 +3,5 @@
3
  "variant": "word",
4
  "language": "cr",
5
  "unique_ngrams": 17,
6
- "total_ngrams": 3016
7
  }
 
3
  "variant": "word",
4
  "language": "cr",
5
  "unique_ngrams": 17,
6
+ "total_ngrams": 2903
7
  }
models/word_ngram/cr_3gram_word_metadata.json CHANGED
@@ -3,5 +3,5 @@
3
  "variant": "word",
4
  "language": "cr",
5
  "unique_ngrams": 16,
6
- "total_ngrams": 2991
7
  }
 
3
  "variant": "word",
4
  "language": "cr",
5
  "unique_ngrams": 16,
6
+ "total_ngrams": 2878
7
  }
models/word_ngram/cr_4gram_word.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8c1ac43c0e3b24b9b058c232c3156ef3f4650c2f5098663b05aff7249ed3efd3
3
- size 5830
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf88b51861c7b3e616572bab59d3354c7c36ad29a0cdfcbfb5afb5e47dca5e8c
3
+ size 5691
models/word_ngram/cr_4gram_word_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "n": 4,
3
  "variant": "word",
4
  "language": "cr",
5
- "unique_ngrams": 166,
6
- "total_ngrams": 2966
7
  }
 
2
  "n": 4,
3
  "variant": "word",
4
  "language": "cr",
5
+ "unique_ngrams": 160,
6
+ "total_ngrams": 2853
7
  }
models/word_ngram/cr_5gram_word.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70760bb1c7ef26efafbcc96f57e199a10946cfebe8a2d04b08e9a1105d4bfa70
3
+ size 5647