omarkamali commited on
Commit
1ca515d
·
verified ·
1 Parent(s): c982f3a

Upload all models and assets for bi (latest)

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +1 -0
  2. README.md +184 -147
  3. models/embeddings/aligned/bi_128d.bin +3 -0
  4. models/embeddings/aligned/bi_128d.meta.json +1 -0
  5. models/embeddings/aligned/bi_128d.projection.npy +3 -0
  6. models/embeddings/aligned/bi_128d_metadata.json +8 -0
  7. models/embeddings/aligned/bi_32d.bin +3 -0
  8. models/embeddings/aligned/bi_32d.meta.json +1 -0
  9. models/embeddings/aligned/bi_32d.projection.npy +3 -0
  10. models/embeddings/aligned/bi_32d_metadata.json +8 -0
  11. models/embeddings/aligned/bi_64d.bin +3 -0
  12. models/embeddings/aligned/bi_64d.meta.json +1 -0
  13. models/embeddings/aligned/bi_64d.projection.npy +3 -0
  14. models/embeddings/aligned/bi_64d_metadata.json +8 -0
  15. models/embeddings/monolingual/bi_128d.bin +2 -2
  16. models/embeddings/monolingual/bi_128d_metadata.json +1 -1
  17. models/embeddings/monolingual/bi_32d.bin +2 -2
  18. models/embeddings/monolingual/bi_32d_metadata.json +1 -1
  19. models/embeddings/monolingual/bi_64d.bin +2 -2
  20. models/embeddings/monolingual/bi_64d_metadata.json +1 -1
  21. models/subword_markov/bi_markov_ctx1_subword.parquet +2 -2
  22. models/subword_markov/bi_markov_ctx1_subword_metadata.json +2 -2
  23. models/subword_markov/bi_markov_ctx2_subword.parquet +2 -2
  24. models/subword_markov/bi_markov_ctx2_subword_metadata.json +2 -2
  25. models/subword_markov/bi_markov_ctx3_subword.parquet +2 -2
  26. models/subword_markov/bi_markov_ctx3_subword_metadata.json +2 -2
  27. models/subword_markov/bi_markov_ctx4_subword.parquet +2 -2
  28. models/subword_markov/bi_markov_ctx4_subword_metadata.json +2 -2
  29. models/subword_ngram/bi_2gram_subword.parquet +2 -2
  30. models/subword_ngram/bi_2gram_subword_metadata.json +2 -2
  31. models/subword_ngram/bi_3gram_subword.parquet +2 -2
  32. models/subword_ngram/bi_3gram_subword_metadata.json +2 -2
  33. models/subword_ngram/bi_4gram_subword.parquet +2 -2
  34. models/subword_ngram/bi_4gram_subword_metadata.json +2 -2
  35. models/subword_ngram/bi_5gram_subword.parquet +3 -0
  36. models/subword_ngram/bi_5gram_subword_metadata.json +7 -0
  37. models/tokenizer/bi_tokenizer_16k.model +2 -2
  38. models/tokenizer/bi_tokenizer_16k.vocab +0 -0
  39. models/tokenizer/bi_tokenizer_8k.model +2 -2
  40. models/tokenizer/bi_tokenizer_8k.vocab +0 -0
  41. models/vocabulary/bi_vocabulary.parquet +2 -2
  42. models/vocabulary/bi_vocabulary_metadata.json +8 -8
  43. models/word_markov/bi_markov_ctx1_word.parquet +2 -2
  44. models/word_markov/bi_markov_ctx1_word_metadata.json +2 -2
  45. models/word_markov/bi_markov_ctx2_word.parquet +2 -2
  46. models/word_markov/bi_markov_ctx2_word_metadata.json +2 -2
  47. models/word_markov/bi_markov_ctx3_word.parquet +2 -2
  48. models/word_markov/bi_markov_ctx3_word_metadata.json +2 -2
  49. models/word_markov/bi_markov_ctx4_word.parquet +2 -2
  50. models/word_markov/bi_markov_ctx4_word_metadata.json +2 -2
.gitattributes CHANGED
@@ -40,3 +40,4 @@ visualizations/tsne_sentences.png filter=lfs diff=lfs merge=lfs -text
40
  visualizations/tsne_words.png filter=lfs diff=lfs merge=lfs -text
41
  visualizations/zipf_law.png filter=lfs diff=lfs merge=lfs -text
42
  visualizations/ngram_coverage.png filter=lfs diff=lfs merge=lfs -text
 
 
40
  visualizations/tsne_words.png filter=lfs diff=lfs merge=lfs -text
41
  visualizations/zipf_law.png filter=lfs diff=lfs merge=lfs -text
42
  visualizations/ngram_coverage.png filter=lfs diff=lfs merge=lfs -text
43
+ visualizations/embedding_tsne_multilingual.png filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  language: bi
3
- language_name: BI
4
  language_family: germanic_west_anglofrisian
5
  tags:
6
  - wikilangs
@@ -10,11 +10,21 @@ tags:
10
  - n-gram
11
  - markov
12
  - wikipedia
 
 
 
 
 
 
 
 
 
 
13
  - monolingual
14
  - family-germanic_west_anglofrisian
15
  license: mit
16
  library_name: wikilangs
17
- pipeline_tag: feature-extraction
18
  datasets:
19
  - omarkamali/wikipedia-monthly
20
  dataset_info:
@@ -23,20 +33,20 @@ dataset_info:
23
  metrics:
24
  - name: best_compression_ratio
25
  type: compression
26
- value: 4.443
27
  - name: best_isotropy
28
  type: isotropy
29
- value: 0.0388
30
  - name: vocabulary_size
31
  type: vocab
32
  value: 0
33
  generated: 2026-01-03
34
  ---
35
 
36
- # BI - Wikilangs Models
37
  ## Comprehensive Research Report & Full Ablation Study
38
 
39
- This repository contains NLP models trained and evaluated by Wikilangs, specifically on **BI** Wikipedia data.
40
  We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and word embeddings.
41
 
42
  ## 📋 Repository Contents
@@ -60,7 +70,7 @@ We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and
60
  - [3. Markov Chain Evaluation](#3-markov-chain-evaluation)
61
  - [4. Vocabulary Analysis](#4-vocabulary-analysis)
62
  - [5. Word Embeddings Evaluation](#5-word-embeddings-evaluation)
63
- - [6. Morphological Analysis (Experimental)](#6-morphological-analysis)
64
  - [7. Summary & Recommendations](#7-summary--recommendations)
65
  - [Metrics Glossary](#appendix-metrics-glossary--interpretation-guide)
66
  - [Visualizations Index](#visualizations-index)
@@ -80,39 +90,39 @@ We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and
80
 
81
  | Vocab Size | Compression | Avg Token Len | UNK Rate | Total Tokens |
82
  |------------|-------------|---------------|----------|--------------|
83
- | **8k** | 4.032x | 4.05 | 0.1444% | 47,092 |
84
- | **16k** | 4.443x 🏆 | 4.47 | 0.1591% | 42,734 |
85
 
86
  ### Tokenization Examples
87
 
88
  Below are sample sentences tokenized with each vocabulary size:
89
 
90
- **Sample 1:** `Copenhagen (toktok Denmak: København), hem i kapitol blong Denmak. Long yia popu...`
91
 
92
  | Vocab | Tokens | Count |
93
  |-------|--------|-------|
94
- | 8k | `▁copenhagen( toktok denmak : ▁københavn ), hemikapitol ... (+20 more)` | 30 |
95
- | 16k | `▁copenhagen( toktok denmak : ▁københavn ), hemikapitol ... (+20 more)` | 30 |
96
 
97
- **Sample 2:** `Emily Elizabeth Dickinson (10 Desemba 15 May em i bin wan poet blong Amerika. ...`
98
 
99
  | Vocab | Tokens | Count |
100
  |-------|--------|-------|
101
- | 8k | `▁em il yelizabethdick ins on( 1 0 ... (+19 more)` | 29 |
102
- | 16k | `▁emilyelizabethdickinson( 1 0desemba ▁–1 ... (+15 more)` | 25 |
103
 
104
- **Sample 3:** `Narafala kaen blong spot long Vanuatu i stap pleiplei tru long kaontri long yumi...`
105
 
106
  | Vocab | Tokens | Count |
107
  |-------|--------|-------|
108
- | 8k | `▁narafala ▁kaenblong ▁spot ▁longvanuatu ▁istap ▁pleiplei ▁tru ... (+7 more)` | 17 |
109
- | 16k | `▁narafalakaenblong ▁spotlong ▁vanuatu ▁istappleipleitru ... (+7 more)` | 17 |
110
 
111
 
112
  ### Key Findings
113
 
114
- - **Best Compression:** 16k achieves 4.443x compression
115
- - **Lowest UNK Rate:** 8k with 0.1444% unknown tokens
116
  - **Trade-off:** Larger vocabularies improve compression but increase model size
117
  - **Recommendation:** 32k vocabulary provides optimal balance for production use
118
 
@@ -129,12 +139,14 @@ Below are sample sentences tokenized with each vocabulary size:
129
 
130
  | N-gram | Variant | Perplexity | Entropy | Unique N-grams | Top-100 Coverage | Top-1000 Coverage |
131
  |--------|---------|------------|---------|----------------|------------------|-------------------|
132
- | **2-gram** | Word | 362 | 8.50 | 1,049 | 58.8% | 98.9% |
133
- | **2-gram** | Subword | 209 🏆 | 7.71 | 983 | 73.7% | 100.0% |
134
- | **3-gram** | Word | 496 | 8.95 | 1,408 | 53.1% | 92.0% |
135
- | **3-gram** | Subword | 1,182 | 10.21 | 5,848 | 38.2% | 79.4% |
136
- | **4-gram** | Word | 887 | 9.79 | 2,457 | 43.9% | 77.4% |
137
- | **4-gram** | Subword | 3,532 | 11.79 | 19,225 | 28.5% | 58.2% |
 
 
138
 
139
  ### Top 5 N-grams by Size
140
 
@@ -142,68 +154,88 @@ Below are sample sentences tokenized with each vocabulary size:
142
 
143
  | Rank | N-gram | Count |
144
  |------|--------|-------|
145
- | 1 | `hem i` | 738 |
146
- | 2 | `stet blong` | 729 |
147
- | 3 | `em i` | 617 |
148
- | 4 | `blong amerika` | 598 |
149
- | 5 | `blong yunaeted` | 535 |
150
 
151
  **3-grams (Word):**
152
 
153
  | Rank | N-gram | Count |
154
  |------|--------|-------|
155
- | 1 | `stet blong amerika` | 583 |
156
- | 2 | `yunaeted stet blong` | 479 |
157
- | 3 | `blong yunaeted stet` | 479 |
158
- | 4 | `blong singsing blong` | 292 |
159
  | 5 | `blong hem i` | 259 |
160
 
161
  **4-grams (Word):**
162
 
163
  | Rank | N-gram | Count |
164
  |------|--------|-------|
165
- | 1 | `yunaeted stet blong amerika` | 477 |
166
- | 2 | `blong yunaeted stet blong` | 470 |
167
  | 3 | `akta blong yunaeted stet` | 210 |
168
- | 4 | `woman blong singsing blong` | 182 |
169
  | 5 | `blong singsing blong japan` | 150 |
170
 
 
 
 
 
 
 
 
 
 
 
171
  **2-grams (Subword):**
172
 
173
  | Rank | N-gram | Count |
174
  |------|--------|-------|
175
- | 1 | `o n` | 9,093 |
176
- | 2 | `n g` | 8,780 |
177
- | 3 | `l o` | 8,027 |
178
- | 4 | `g _` | 7,936 |
179
- | 5 | `_ b` | 7,059 |
180
 
181
  **3-grams (Subword):**
182
 
183
  | Rank | N-gram | Count |
184
  |------|--------|-------|
185
- | 1 | `n g _` | 7,795 |
186
- | 2 | `o n g` | 7,296 |
187
- | 3 | `l o n` | 7,257 |
188
- | 4 | `_ b l` | 5,277 |
189
- | 5 | `b l o` | 5,252 |
190
 
191
  **4-grams (Subword):**
192
 
193
  | Rank | N-gram | Count |
194
  |------|--------|-------|
195
- | 1 | `o n g _` | 7,200 |
196
- | 2 | `l o n g` | 7,191 |
197
- | 3 | `_ b l o` | 5,238 |
198
- | 4 | `b l o n` | 5,015 |
199
- | 5 | `_ l o n` | 2,153 |
 
 
 
 
 
 
 
 
 
 
200
 
201
 
202
  ### Key Findings
203
 
204
- - **Best Perplexity:** 2-gram (subword) with 209
205
  - **Entropy Trend:** Decreases with larger n-grams (more predictable)
206
- - **Coverage:** Top-1000 patterns cover ~58% of corpus
207
  - **Recommendation:** 4-gram or 5-gram for best predictive performance
208
 
209
  ---
@@ -219,14 +251,14 @@ Below are sample sentences tokenized with each vocabulary size:
219
 
220
  | Context | Variant | Avg Entropy | Perplexity | Branching Factor | Unique Contexts | Predictability |
221
  |---------|---------|-------------|------------|------------------|-----------------|----------------|
222
- | **1** | Word | 0.5840 | 1.499 | 3.04 | 8,338 | 41.6% |
223
- | **1** | Subword | 0.9602 | 1.946 | 6.50 | 364 | 4.0% |
224
- | **2** | Word | 0.1997 | 1.148 | 1.41 | 24,957 | 80.0% |
225
- | **2** | Subword | 0.9911 | 1.988 | 5.10 | 2,361 | 0.9% |
226
- | **3** | Word | 0.0755 | 1.054 | 1.13 | 34,724 | 92.4% |
227
- | **3** | Subword | 0.7964 | 1.737 | 3.17 | 12,016 | 20.4% |
228
- | **4** | Word | 0.0328 🏆 | 1.023 | 1.06 | 38,736 | 96.7% |
229
- | **4** | Subword | 0.4627 | 1.378 | 1.90 | 38,018 | 53.7% |
230
 
231
  ### Generated Text Samples (Word-based)
232
 
@@ -234,27 +266,27 @@ Below are text samples generated from each word-based Markov chain model:
234
 
235
  **Context Size 1:**
236
 
237
- 1. `blong olgeta mo yu ol disaepol blong dover long wol plante fasin blong court i wan`
238
- 2. `i bin wan strongfala win if you s 84 913 km2 populaesen blong stet blong hem`
239
- 3. `long saed blong tekem carbondioxde mo wanwan aelan gaua o aoba hem i wokem long milly`
240
 
241
  **Context Size 2:**
242
 
243
- 1. `hem i stap insaet long solwota everi man i save sindaon o silip long hem islam relijon`
244
- 2. `stet blong philippines blong stet blong amerika blong stet blong amerika blong stet blong amerika mo...`
245
- 3. `em i bin ded 8 septemba em i woman blong singsing blong japan man blong singsing blong`
246
 
247
  **Context Size 3:**
248
 
249
- 1. `blong yunaeted stet blong amerika model akta blong pornografi blong ajentina em i stap popiula from ...`
250
- 2. `yunaeted stet blong amerika akta blong yunaeted stet blong amerika blong yunaeted stet blong amerika...`
251
- 3. `blong singsing blong japan thumb anna iriyama man blong singsing blong kanada man blong singsing blo...`
252
 
253
  **Context Size 4:**
254
 
255
- 1. `blong yunaeted stet blong amerika blong stet blong yunaeted stet blong amerika blong yunaeted stet b...`
256
- 2. `yunaeted stet blong amerika blong stet blong yunaeted stet blong amerika blong yunaeted stet blong a...`
257
- 3. `akta blong yunaeted stet blong amerika akta blong yunaeted stet blong amerika blong stet blong yunae...`
258
 
259
 
260
  ### Generated Text Samples (Subword-based)
@@ -263,34 +295,34 @@ Below are text samples generated from each subword-based Markov chain model:
263
 
264
  **Context Size 1:**
265
 
266
- 1. `_dimo_ste_lon_i_`
267
- 2. `a_blong_bl_19_s_`
268
- 3. `ngstang_yulolem:`
269
 
270
  **Context Size 2:**
271
 
272
- 1. `ong_300px_12_3_44`
273
- 2. `ng_st_boetexanblo`
274
- 3. `long_prol_no,_рос`
275
 
276
  **Context Size 3:**
277
 
278
- 1. `ng_amerika._akta_b`
279
- 2. `ong_savela_taeland`
280
- 3. `long_amerika_maura`
281
 
282
  **Context Size 4:**
283
 
284
- 1. `ong_amerika._praem_`
285
- 2. `long_not_prize_nigh`
286
- 3. `_blong_21_man_blong`
287
 
288
 
289
  ### Key Findings
290
 
291
- - **Best Predictability:** Context-4 (word) with 96.7% predictability
292
  - **Branching Factor:** Decreases with context size (more deterministic)
293
- - **Memory Trade-off:** Larger contexts require more storage (38,018 contexts)
294
  - **Recommendation:** Context-3 or Context-4 for text generation
295
 
296
  ---
@@ -306,48 +338,48 @@ Below are text samples generated from each subword-based Markov chain model:
306
 
307
  | Metric | Value |
308
  |--------|-------|
309
- | Vocabulary Size | 3,117 |
310
- | Total Tokens | 48,872 |
311
- | Mean Frequency | 15.68 |
312
  | Median Frequency | 3 |
313
- | Frequency Std Dev | 124.49 |
314
 
315
  ### Most Common Words
316
 
317
  | Rank | Word | Frequency |
318
  |------|------|-----------|
319
- | 1 | blong | 5,014 |
320
- | 2 | i | 3,182 |
321
- | 3 | long | 2,146 |
322
- | 4 | mo | 1,031 |
323
- | 5 | hem | 1,008 |
324
- | 6 | ol | 886 |
325
- | 7 | wan | 875 |
326
- | 8 | stet | 840 |
327
- | 9 | amerika | 673 |
328
- | 10 | em | 660 |
329
 
330
  ### Least Common Words (from vocabulary)
331
 
332
  | Rank | Word | Frequency |
333
  |------|------|-----------|
334
- | 1 | lotta | 2 |
335
- | 2 | continua | 2 |
336
- | 3 | ekshumesen | 2 |
337
- | 4 | suspension | 2 |
338
- | 5 | fulwan | 2 |
339
- | 6 | konfirm | 2 |
340
- | 7 | trial | 2 |
341
- | 8 | window | 2 |
342
- | 9 | piazza | 2 |
343
- | 10 | fontana | 2 |
344
 
345
  ### Zipf's Law Analysis
346
 
347
  | Metric | Value |
348
  |--------|-------|
349
- | Zipf Coefficient | 1.0400 |
350
- | R² (Goodness of Fit) | 0.989215 |
351
  | Adherence Quality | **excellent** |
352
 
353
  ### Coverage Analysis
@@ -361,9 +393,9 @@ Below are text samples generated from each subword-based Markov chain model:
361
 
362
  ### Key Findings
363
 
364
- - **Zipf Compliance:** R²=0.9892 indicates excellent adherence to Zipf's law
365
  - **High Frequency Dominance:** Top 100 words cover 62.1% of corpus
366
- - **Long Tail:** -6,883 words needed for remaining 100.0% coverage
367
 
368
  ---
369
  ## 5. Word Embeddings Evaluation
@@ -379,37 +411,40 @@ Below are text samples generated from each subword-based Markov chain model:
379
 
380
  ### 5.1 Cross-Lingual Alignment
381
 
382
- > *Note: Multilingual alignment visualization not available for this language.*
 
 
383
 
384
 
385
  ### 5.2 Model Comparison
386
 
387
  | Model | Dimension | Isotropy | Semantic Density | Alignment R@1 | Alignment R@10 |
388
  |-------|-----------|----------|------------------|---------------|----------------|
389
- | **mono_32d** | 32 | 0.0388 🏆 | 0.6777 | N/A | N/A |
390
- | **mono_64d** | 64 | 0.0097 | 0.6676 | N/A | N/A |
391
- | **mono_128d** | 128 | 0.0021 | 0.6720 | N/A | N/A |
 
 
 
392
 
393
  ### Key Findings
394
 
395
- - **Best Isotropy:** mono_32d with 0.0388 (more uniform distribution)
396
- - **Semantic Density:** Average pairwise similarity of 0.6724. Lower values indicate better semantic separation.
397
- - **Alignment Quality:** No aligned models evaluated in this run.
398
  - **Recommendation:** 128d aligned for best cross-lingual performance
399
 
400
  ---
401
  ## 6. Morphological Analysis (Experimental)
402
 
403
- > ⚠️ **Warning:** This language shows low morphological productivity. The statistical signals used for this analysis may be noisy or less reliable than for morphologically rich languages.
404
-
405
  This section presents an automated morphological analysis derived from the statistical divergence between word-level and subword-level models. By analyzing where subword predictability spikes and where word-level coverage fails, we can infer linguistic structures without supervised data.
406
 
407
  ### 6.1 Productivity & Complexity
408
 
409
  | Metric | Value | Interpretation | Recommendation |
410
  |--------|-------|----------------|----------------|
411
- | Productivity Index | **0.000** | Low morphological productivity | ⚠️ Likely unreliable |
412
- | Idiomaticity Gap | **-1.000** | Low formulaic content | - |
413
 
414
  ### 6.2 Affix Inventory (Productive Units)
415
 
@@ -422,9 +457,9 @@ These are the most productive prefixes and suffixes identified by sampling the v
422
  #### Productive Suffixes
423
  | Suffix | Examples |
424
  |--------|----------|
425
- | `-en` | ren, disisen, citizen |
426
- | `-an` | givhan, kirgistan, wan |
427
- | `-em` | shoem, wokem, blem |
428
 
429
  ### 6.3 Bound Stems (Lexical Roots)
430
 
@@ -432,7 +467,7 @@ Bound stems are high-frequency subword units that are semantically cohesive but
432
 
433
  | Stem | Cohesion | Substitutability | Examples |
434
  |------|----------|------------------|----------|
435
- | `amba` | 1.38x | 8 contexts | ambae, namba, bambae |
436
 
437
  ### 6.4 Affix Compatibility (Co-occurrence)
438
 
@@ -450,23 +485,25 @@ Using **Recursive Hierarchical Substitutability**, we decompose complex words in
450
  | republican | **`republic-an`** | 4.5 | `republic` |
451
  | andastanem | **`andast-an-em`** | 3.0 | `andast` |
452
  | niutesteman | **`niutest-em-an`** | 3.0 | `niutest` |
453
- | kirgistan | **`kirgist-an`** | 1.5 | `kirgist` |
454
- | valencian | **`valenci-an`** | 1.5 | `valenci` |
455
- | singaotem | **`singaot-em`** | 1.5 | `singaot` |
456
- | defdefren | **`defdefr-en`** | 1.5 | `defdefr` |
457
- | melanesian | **`melanesi-an`** | 1.5 | `melanesi` |
458
- | konstitusen | **`konstitus-en`** | 1.5 | `konstitus` |
459
- | komposisen | **`komposis-en`** | 1.5 | `komposis` |
460
- | smithsonian | **`smithsoni-an`** | 1.5 | `smithsoni` |
461
- | kompitisen | **`kompitis-en`** | 1.5 | `kompitis` |
462
- | bisnesman | **`bisnesm-an`** | 1.5 | `bisnesm` |
463
- | protestan | **`protest-an`** | 1.5 | `protest` |
464
  | ekshumesen | **`ekshumes-en`** | 1.5 | `ekshumes` |
 
 
 
 
 
 
465
 
466
  ### 6.6 Linguistic Interpretation
467
 
468
  > **Automated Insight:**
469
- The language BI appears to be more isolating or has a highly fixed vocabulary. Word-level models perform nearly as well as subword models, indicating fewer productive morphological processes.
 
 
470
 
471
  ---
472
  ## 7. Summary & Recommendations
@@ -478,8 +515,8 @@ The language BI appears to be more isolating or has a highly fixed vocabulary. W
478
  | Component | Recommended | Rationale |
479
  |-----------|-------------|-----------|
480
  | Tokenizer | **16k BPE** | Best compression (4.44x) |
481
- | N-gram | **2-gram** | Lowest perplexity (209) |
482
- | Markov | **Context-4** | Highest predictability (96.7%) |
483
  | Embeddings | **100d** | Balanced semantic capture and isotropy |
484
 
485
 
@@ -693,4 +730,4 @@ MIT License - Free for academic and commercial use.
693
  ---
694
  *Generated by Wikilangs Models Pipeline*
695
 
696
- *Report Date: 2026-01-03 07:17:54*
 
1
  ---
2
  language: bi
3
+ language_name: Bislama
4
  language_family: germanic_west_anglofrisian
5
  tags:
6
  - wikilangs
 
10
  - n-gram
11
  - markov
12
  - wikipedia
13
+ - feature-extraction
14
+ - sentence-similarity
15
+ - tokenization
16
+ - n-grams
17
+ - markov-chain
18
+ - text-mining
19
+ - fasttext
20
+ - babelvec
21
+ - vocabulous
22
+ - vocabulary
23
  - monolingual
24
  - family-germanic_west_anglofrisian
25
  license: mit
26
  library_name: wikilangs
27
+ pipeline_tag: text-generation
28
  datasets:
29
  - omarkamali/wikipedia-monthly
30
  dataset_info:
 
33
  metrics:
34
  - name: best_compression_ratio
35
  type: compression
36
+ value: 4.441
37
  - name: best_isotropy
38
  type: isotropy
39
+ value: 0.0691
40
  - name: vocabulary_size
41
  type: vocab
42
  value: 0
43
  generated: 2026-01-03
44
  ---
45
 
46
+ # Bislama - Wikilangs Models
47
  ## Comprehensive Research Report & Full Ablation Study
48
 
49
+ This repository contains NLP models trained and evaluated by Wikilangs, specifically on **Bislama** Wikipedia data.
50
  We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and word embeddings.
51
 
52
  ## 📋 Repository Contents
 
70
  - [3. Markov Chain Evaluation](#3-markov-chain-evaluation)
71
  - [4. Vocabulary Analysis](#4-vocabulary-analysis)
72
  - [5. Word Embeddings Evaluation](#5-word-embeddings-evaluation)
73
+ - [6. Morphological Analysis (Experimental)](#6--morphological-analysis-experimental)
74
  - [7. Summary & Recommendations](#7-summary--recommendations)
75
  - [Metrics Glossary](#appendix-metrics-glossary--interpretation-guide)
76
  - [Visualizations Index](#visualizations-index)
 
90
 
91
  | Vocab Size | Compression | Avg Token Len | UNK Rate | Total Tokens |
92
  |------------|-------------|---------------|----------|--------------|
93
+ | **8k** | 4.034x | 4.06 | 0.1436% | 45,948 |
94
+ | **16k** | 4.441x 🏆 | 4.46 | 0.1581% | 41,742 |
95
 
96
  ### Tokenization Examples
97
 
98
  Below are sample sentences tokenized with each vocabulary size:
99
 
100
+ **Sample 1:** `Spiro Theodore "Ted" Agnew (9 Novemba 17 Septemba em i politikis blong Yunaete...`
101
 
102
  | Vocab | Tokens | Count |
103
  |-------|--------|-------|
104
+ | 8k | `▁spi ro theodore" ted "agnew( 9 novemba ... (+19 more)` | 29 |
105
+ | 16k | `▁spirotheodore" ted "agnew( 9 novemba ▁– ... (+18 more)` | 28 |
106
 
107
+ **Sample 2:** `Xi Jinping (boen i hed blong stet blong Jaena. blong Stet blong Jaena`
108
 
109
  | Vocab | Tokens | Count |
110
  |-------|--------|-------|
111
+ | 8k | `▁xi ▁jinping ▁( boen ihed ▁blong ▁stetblong ▁jaena ... (+5 more)` | 15 |
112
+ | 16k | `▁xijinping( boen i ▁hed ▁blongstet ▁blongjaena ... (+5 more)` | 15 |
113
 
114
+ **Sample 3:** `Miori Ichikawa (boen 12 Februari em i bin woman blong singsing blong Japan. woma...`
115
 
116
  | Vocab | Tokens | Count |
117
  |-------|--------|-------|
118
+ | 8k | `▁mi oriich ika wa( boen1 2 ... (+16 more)` | 26 |
119
+ | 16k | `▁mioriichikawa( boen1 2februariemi ... (+13 more)` | 23 |
120
 
121
 
122
  ### Key Findings
123
 
124
+ - **Best Compression:** 16k achieves 4.441x compression
125
+ - **Lowest UNK Rate:** 8k with 0.1436% unknown tokens
126
  - **Trade-off:** Larger vocabularies improve compression but increase model size
127
  - **Recommendation:** 32k vocabulary provides optimal balance for production use
128
 
 
139
 
140
  | N-gram | Variant | Perplexity | Entropy | Unique N-grams | Top-100 Coverage | Top-1000 Coverage |
141
  |--------|---------|------------|---------|----------------|------------------|-------------------|
142
+ | **2-gram** | Word | 362 | 8.50 | 1,045 | 58.8% | 99.0% |
143
+ | **2-gram** | Subword | 208 🏆 | 7.70 | 976 | 73.9% | 100.0% |
144
+ | **3-gram** | Word | 494 | 8.95 | 1,403 | 53.1% | 92.1% |
145
+ | **3-gram** | Subword | 1,176 | 10.20 | 5,825 | 38.3% | 79.5% |
146
+ | **4-gram** | Word | 875 | 9.77 | 2,432 | 44.2% | 77.7% |
147
+ | **4-gram** | Subword | 3,512 | 11.78 | 19,179 | 28.6% | 58.3% |
148
+ | **5-gram** | Word | 727 | 9.51 | 1,831 | 46.0% | 82.2% |
149
+ | **5-gram** | Subword | 5,192 | 12.34 | 26,363 | 25.9% | 52.6% |
150
 
151
  ### Top 5 N-grams by Size
152
 
 
154
 
155
  | Rank | N-gram | Count |
156
  |------|--------|-------|
157
+ | 1 | `hem i` | 741 |
158
+ | 2 | `stet blong` | 731 |
159
+ | 3 | `em i` | 611 |
160
+ | 4 | `blong amerika` | 599 |
161
+ | 5 | `blong yunaeted` | 537 |
162
 
163
  **3-grams (Word):**
164
 
165
  | Rank | N-gram | Count |
166
  |------|--------|-------|
167
+ | 1 | `stet blong amerika` | 585 |
168
+ | 2 | `blong yunaeted stet` | 481 |
169
+ | 3 | `yunaeted stet blong` | 481 |
170
+ | 4 | `blong singsing blong` | 291 |
171
  | 5 | `blong hem i` | 259 |
172
 
173
  **4-grams (Word):**
174
 
175
  | Rank | N-gram | Count |
176
  |------|--------|-------|
177
+ | 1 | `yunaeted stet blong amerika` | 479 |
178
+ | 2 | `blong yunaeted stet blong` | 472 |
179
  | 3 | `akta blong yunaeted stet` | 210 |
180
+ | 4 | `woman blong singsing blong` | 181 |
181
  | 5 | `blong singsing blong japan` | 150 |
182
 
183
+ **5-grams (Word):**
184
+
185
+ | Rank | N-gram | Count |
186
+ |------|--------|-------|
187
+ | 1 | `blong yunaeted stet blong amerika` | 471 |
188
+ | 2 | `akta blong yunaeted stet blong` | 210 |
189
+ | 3 | `woman blong singsing blong japan` | 129 |
190
+ | 4 | `em i woman blong singsing` | 100 |
191
+ | 5 | `i woman blong singsing blong` | 96 |
192
+
193
  **2-grams (Subword):**
194
 
195
  | Rank | N-gram | Count |
196
  |------|--------|-------|
197
+ | 1 | `o n` | 9,097 |
198
+ | 2 | `n g` | 8,801 |
199
+ | 3 | `l o` | 8,033 |
200
+ | 4 | `g _` | 7,960 |
201
+ | 5 | `_ b` | 7,074 |
202
 
203
  **3-grams (Subword):**
204
 
205
  | Rank | N-gram | Count |
206
  |------|--------|-------|
207
+ | 1 | `n g _` | 7,816 |
208
+ | 2 | `o n g` | 7,315 |
209
+ | 3 | `l o n` | 7,271 |
210
+ | 4 | `_ b l` | 5,295 |
211
+ | 5 | `b l o` | 5,265 |
212
 
213
  **4-grams (Subword):**
214
 
215
  | Rank | N-gram | Count |
216
  |------|--------|-------|
217
+ | 1 | `o n g _` | 7,216 |
218
+ | 2 | `l o n g` | 7,207 |
219
+ | 3 | `_ b l o` | 5,255 |
220
+ | 4 | `b l o n` | 5,031 |
221
+ | 5 | `_ l o n` | 2,154 |
222
+
223
+ **5-grams (Subword):**
224
+
225
+ | Rank | N-gram | Count |
226
+ |------|--------|-------|
227
+ | 1 | `l o n g _` | 7,179 |
228
+ | 2 | `b l o n g` | 5,030 |
229
+ | 3 | `_ b l o n` | 5,028 |
230
+ | 4 | `_ l o n g` | 2,151 |
231
+ | 5 | `e m _ i _` | 1,374 |
232
 
233
 
234
  ### Key Findings
235
 
236
+ - **Best Perplexity:** 2-gram (subword) with 208
237
  - **Entropy Trend:** Decreases with larger n-grams (more predictable)
238
+ - **Coverage:** Top-1000 patterns cover ~53% of corpus
239
  - **Recommendation:** 4-gram or 5-gram for best predictive performance
240
 
241
  ---
 
251
 
252
  | Context | Variant | Avg Entropy | Perplexity | Branching Factor | Unique Contexts | Predictability |
253
  |---------|---------|-------------|------------|------------------|-----------------|----------------|
254
+ | **1** | Word | 0.5784 | 1.493 | 3.02 | 8,408 | 42.2% |
255
+ | **1** | Subword | 0.9577 | 1.942 | 6.51 | 362 | 4.2% |
256
+ | **2** | Word | 0.1997 | 1.148 | 1.41 | 25,020 | 80.0% |
257
+ | **2** | Subword | 0.9916 | 1.988 | 5.13 | 2,350 | 0.8% |
258
+ | **3** | Word | 0.0750 | 1.053 | 1.13 | 34,806 | 92.5% |
259
+ | **3** | Subword | 0.7944 | 1.734 | 3.18 | 12,029 | 20.6% |
260
+ | **4** | Word | 0.0323 🏆 | 1.023 | 1.05 | 38,812 | 96.8% |
261
+ | **4** | Subword | 0.4624 | 1.378 | 1.90 | 38,112 | 53.8% |
262
 
263
  ### Generated Text Samples (Word-based)
264
 
 
266
 
267
  **Context Size 1:**
268
 
269
+ 1. `blong miusik grup i praem minista blong pasifik tu kristianiti islam jeinisim i praem minista blong`
270
+ 2. `i stap wetem graon kavremap 29 septemba hem hemi sapraesm ol pipol likem kakae we i`
271
+ 3. `long septemba i stap mekem afta blong et et i wan fruit kakae we ol komposisen`
272
 
273
  **Context Size 2:**
274
 
275
+ 1. `hem i wan miusik grup stet blong philippines blong stet blong amerika man blong singsing blong japan`
276
+ 2. `stet blong peru bik kaontri long saot blong yurop we i stap araon 860 090 external links`
277
+ 3. `em i bin transletem niu testeman i kam mo watchem kustom danis wetem good fren pipol`
278
 
279
  **Context Size 3:**
280
 
281
+ 1. `yunaeted stet blong amerika akta blong yunaeted stet blong amerika risos long internet www vilnius l...`
282
+ 2. `blong yunaeted stet blong amerika blong yunaeted stet blong amerika akta blong yunaeted stet blong a...`
283
+ 3. `blong singsing blong taelan woman blong singsing blong japan woman blong singsing blong japan man bl...`
284
 
285
  **Context Size 4:**
286
 
287
+ 1. `blong yunaeted stet blong amerika akta blong yunaeted stet blong amerika blong stet blong yunaeted s...`
288
+ 2. `yunaeted stet blong amerika bara lyle crist images of america alliance arcadia publishing s 41 isbn ...`
289
+ 3. `akta blong yunaeted stet blong amerika akta blong yunaeted stet blong amerika akta blong yunaeted st...`
290
 
291
 
292
  ### Generated Text Samples (Subword-based)
 
295
 
296
  **Context Size 1:**
297
 
298
+ 1. `_stakthae_m_blon`
299
+ 2. `ak_25paryulgraju`
300
+ 3. `ng_lons_i_we_d_p`
301
 
302
  **Context Size 2:**
303
 
304
+ 1. `ong_yun_wosing_i_`
305
+ 2. `ng_noasol_ww.cita`
306
+ 3. `long_en_lon_i_sol`
307
 
308
  **Context Size 3:**
309
 
310
+ 1. `ng_nara_(cano_red_`
311
+ 2. `ong_wan_blong_mius`
312
+ 3. `long_(long_blong_y`
313
 
314
  **Context Size 4:**
315
 
316
+ 1. `ong_nolej,_televis_`
317
+ 2. `long_gud_fasin_muha`
318
+ 3. `_blong_stet_blong_s`
319
 
320
 
321
  ### Key Findings
322
 
323
+ - **Best Predictability:** Context-4 (word) with 96.8% predictability
324
  - **Branching Factor:** Decreases with context size (more deterministic)
325
+ - **Memory Trade-off:** Larger contexts require more storage (38,112 contexts)
326
  - **Recommendation:** Context-3 or Context-4 for text generation
327
 
328
  ---
 
338
 
339
  | Metric | Value |
340
  |--------|-------|
341
+ | Vocabulary Size | 3,106 |
342
+ | Total Tokens | 48,839 |
343
+ | Mean Frequency | 15.72 |
344
  | Median Frequency | 3 |
345
+ | Frequency Std Dev | 125.16 |
346
 
347
  ### Most Common Words
348
 
349
  | Rank | Word | Frequency |
350
  |------|------|-----------|
351
+ | 1 | blong | 5,030 |
352
+ | 2 | i | 3,201 |
353
+ | 3 | long | 2,145 |
354
+ | 4 | mo | 1,056 |
355
+ | 5 | hem | 1,010 |
356
+ | 6 | ol | 899 |
357
+ | 7 | wan | 870 |
358
+ | 8 | stet | 842 |
359
+ | 9 | amerika | 672 |
360
+ | 10 | em | 654 |
361
 
362
  ### Least Common Words (from vocabulary)
363
 
364
  | Rank | Word | Frequency |
365
  |------|------|-----------|
366
+ | 1 | ftps | 2 |
367
+ | 2 | sftp | 2 |
368
+ | 3 | operating | 2 |
369
+ | 4 | guide | 2 |
370
+ | 5 | spesifikesen | 2 |
371
+ | 6 | firewall | 2 |
372
+ | 7 | sapot | 2 |
373
+ | 8 | lesin | 2 |
374
+ | 9 | sanem | 2 |
375
+ | 10 | extended | 2 |
376
 
377
  ### Zipf's Law Analysis
378
 
379
  | Metric | Value |
380
  |--------|-------|
381
+ | Zipf Coefficient | 1.0402 |
382
+ | R² (Goodness of Fit) | 0.989274 |
383
  | Adherence Quality | **excellent** |
384
 
385
  ### Coverage Analysis
 
393
 
394
  ### Key Findings
395
 
396
+ - **Zipf Compliance:** R²=0.9893 indicates excellent adherence to Zipf's law
397
  - **High Frequency Dominance:** Top 100 words cover 62.1% of corpus
398
+ - **Long Tail:** -6,894 words needed for remaining 100.0% coverage
399
 
400
  ---
401
  ## 5. Word Embeddings Evaluation
 
411
 
412
  ### 5.1 Cross-Lingual Alignment
413
 
414
+ ![Alignment Quality](visualizations/embedding_alignment_quality.png)
415
+
416
+ ![Multilingual t-SNE](visualizations/embedding_tsne_multilingual.png)
417
 
418
 
419
  ### 5.2 Model Comparison
420
 
421
  | Model | Dimension | Isotropy | Semantic Density | Alignment R@1 | Alignment R@10 |
422
  |-------|-----------|----------|------------------|---------------|----------------|
423
+ | **mono_32d** | 32 | 0.0691 🏆 | 0.6642 | N/A | N/A |
424
+ | **mono_64d** | 64 | 0.0097 | 0.6595 | N/A | N/A |
425
+ | **mono_128d** | 128 | 0.0022 | 0.6755 | N/A | N/A |
426
+ | **aligned_32d** | 32 | 0.0691 | 0.6741 | 0.0060 | 0.0420 |
427
+ | **aligned_64d** | 64 | 0.0097 | 0.6519 | 0.0080 | 0.0860 |
428
+ | **aligned_128d** | 128 | 0.0022 | 0.6801 | 0.0200 | 0.0920 |
429
 
430
  ### Key Findings
431
 
432
+ - **Best Isotropy:** mono_32d with 0.0691 (more uniform distribution)
433
+ - **Semantic Density:** Average pairwise similarity of 0.6675. Lower values indicate better semantic separation.
434
+ - **Alignment Quality:** Aligned models achieve up to 2.0% R@1 in cross-lingual retrieval.
435
  - **Recommendation:** 128d aligned for best cross-lingual performance
436
 
437
  ---
438
  ## 6. Morphological Analysis (Experimental)
439
 
 
 
440
  This section presents an automated morphological analysis derived from the statistical divergence between word-level and subword-level models. By analyzing where subword predictability spikes and where word-level coverage fails, we can infer linguistic structures without supervised data.
441
 
442
  ### 6.1 Productivity & Complexity
443
 
444
  | Metric | Value | Interpretation | Recommendation |
445
  |--------|-------|----------------|----------------|
446
+ | Productivity Index | **5.000** | High morphological productivity | Reliable analysis |
447
+ | Idiomaticity Gap | **0.564** | High formulaic/idiomatic content | - |
448
 
449
  ### 6.2 Affix Inventory (Productive Units)
450
 
 
457
  #### Productive Suffixes
458
  | Suffix | Examples |
459
  |--------|----------|
460
+ | `-en` | warren, truiden, paten |
461
+ | `-em` | katem, raonem, sanem |
462
+ | `-an` | ejukesan, busan, giaman |
463
 
464
  ### 6.3 Bound Stems (Lexical Roots)
465
 
 
467
 
468
  | Stem | Cohesion | Substitutability | Examples |
469
  |------|----------|------------------|----------|
470
+ | `amba` | 1.40x | 8 contexts | ambae, namba, stamba |
471
 
472
  ### 6.4 Affix Compatibility (Co-occurrence)
473
 
 
485
  | republican | **`republic-an`** | 4.5 | `republic` |
486
  | andastanem | **`andast-an-em`** | 3.0 | `andast` |
487
  | niutesteman | **`niutest-em-an`** | 3.0 | `niutest` |
488
+ | komunikesen | **`komunikes-en`** | 1.5 | `komunikes` |
489
+ | oganaesesen | **`oganaeses-en`** | 1.5 | `oganaeses` |
490
+ | sustreksen | **`sustreks-en`** | 1.5 | `sustreks` |
491
+ | vaespresiden | **`vaespresid-en`** | 1.5 | `vaespresid` |
492
+ | populesen | **`popules-en`** | 1.5 | `popules` |
 
 
 
 
 
 
493
  | ekshumesen | **`ekshumes-en`** | 1.5 | `ekshumes` |
494
+ | komposisen | **`komposis-en`** | 1.5 | `komposis` |
495
+ | konstitusen | **`konstitus-en`** | 1.5 | `konstitus` |
496
+ | sébastien | **`sébasti-en`** | 1.5 | `sébasti` |
497
+ | austronesian | **`austronesi-an`** | 1.5 | `austronesi` |
498
+ | divelopem | **`divelop-em`** | 1.5 | `divelop` |
499
+ | christian | **`christi-an`** | 1.5 | `christi` |
500
 
501
  ### 6.6 Linguistic Interpretation
502
 
503
  > **Automated Insight:**
504
+ The language Bislama shows high morphological productivity. The subword models are significantly more efficient than word models, suggesting a rich system of affixation or compounding.
505
+
506
+ > **Note on Idiomaticity:** The high Idiomaticity Gap suggests a large number of frequent multi-word expressions or formulaic sequences that are statistically distinct from their component parts.
507
 
508
  ---
509
  ## 7. Summary & Recommendations
 
515
  | Component | Recommended | Rationale |
516
  |-----------|-------------|-----------|
517
  | Tokenizer | **16k BPE** | Best compression (4.44x) |
518
+ | N-gram | **2-gram** | Lowest perplexity (208) |
519
+ | Markov | **Context-4** | Highest predictability (96.8%) |
520
  | Embeddings | **100d** | Balanced semantic capture and isotropy |
521
 
522
 
 
730
  ---
731
  *Generated by Wikilangs Models Pipeline*
732
 
733
+ *Report Date: 2026-01-03 18:57:38*
models/embeddings/aligned/bi_128d.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b04089d63affbd01946ea22712a42b9270a9e9c1e3642b2c4557f09e64125f45
3
+ size 1025093310
models/embeddings/aligned/bi_128d.meta.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lang": "bi", "dim": 128, "max_seq_len": 512, "is_aligned": true}
models/embeddings/aligned/bi_128d.projection.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a0187132ed142b2493533fcad026c5d7be9bcd7cd03af41e20ed204911939b9
3
+ size 65664
models/embeddings/aligned/bi_128d_metadata.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "language": "bi",
3
+ "dimension": 128,
4
+ "version": "aligned",
5
+ "hub_language": "en",
6
+ "seed_vocab_size": 565,
7
+ "vocab_size": 1052
8
+ }
models/embeddings/aligned/bi_32d.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a1b8bc649eeb41d1b7c45409caed59f6f1e6fc8c3a618675db998527999e92d
3
+ size 256285374
models/embeddings/aligned/bi_32d.meta.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lang": "bi", "dim": 32, "max_seq_len": 512, "is_aligned": true}
models/embeddings/aligned/bi_32d.projection.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af2e9ecddbabbf795c354094a776879f0e3b6edcdaff3d42ad65641f2bf82a11
3
+ size 4224
models/embeddings/aligned/bi_32d_metadata.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "language": "bi",
3
+ "dimension": 32,
4
+ "version": "aligned",
5
+ "hub_language": "en",
6
+ "seed_vocab_size": 565,
7
+ "vocab_size": 1052
8
+ }
models/embeddings/aligned/bi_64d.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bfe3f902fbea272ab483848ef0bfa3a8f447352ec83746e7036947e7d16daefc
3
+ size 512554686
models/embeddings/aligned/bi_64d.meta.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lang": "bi", "dim": 64, "max_seq_len": 512, "is_aligned": true}
models/embeddings/aligned/bi_64d.projection.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0fbe174e9560c7369b8587d1eec8c4597aa4afff653cdea5e66616f728f25bc
3
+ size 16512
models/embeddings/aligned/bi_64d_metadata.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "language": "bi",
3
+ "dimension": 64,
4
+ "version": "aligned",
5
+ "hub_language": "en",
6
+ "seed_vocab_size": 565,
7
+ "vocab_size": 1052
8
+ }
models/embeddings/monolingual/bi_128d.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5ba043557491d9bd866f4880ed63951e9cedd258275bb5ab99cc6a4588a48418
3
- size 1025087100
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b04089d63affbd01946ea22712a42b9270a9e9c1e3642b2c4557f09e64125f45
3
+ size 1025093310
models/embeddings/monolingual/bi_128d_metadata.json CHANGED
@@ -11,5 +11,5 @@
11
  "encoding_method": "rope",
12
  "dim": 128
13
  },
14
- "vocab_size": 1046
15
  }
 
11
  "encoding_method": "rope",
12
  "dim": 128
13
  },
14
+ "vocab_size": 1052
15
  }
models/embeddings/monolingual/bi_32d.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:77f00c5cf1334339a3726d39a681bacc6c8c5b7d226141cb5bd2ec7d3d93731c
3
- size 256283772
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a1b8bc649eeb41d1b7c45409caed59f6f1e6fc8c3a618675db998527999e92d
3
+ size 256285374
models/embeddings/monolingual/bi_32d_metadata.json CHANGED
@@ -11,5 +11,5 @@
11
  "encoding_method": "rope",
12
  "dim": 32
13
  },
14
- "vocab_size": 1046
15
  }
 
11
  "encoding_method": "rope",
12
  "dim": 32
13
  },
14
+ "vocab_size": 1052
15
  }
models/embeddings/monolingual/bi_64d.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:64be2d287f8ece0f9696fe2372c048f2ea1c8dc331d051d642a50efe1563bc8b
3
- size 512551548
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bfe3f902fbea272ab483848ef0bfa3a8f447352ec83746e7036947e7d16daefc
3
+ size 512554686
models/embeddings/monolingual/bi_64d_metadata.json CHANGED
@@ -11,5 +11,5 @@
11
  "encoding_method": "rope",
12
  "dim": 64
13
  },
14
- "vocab_size": 1046
15
  }
 
11
  "encoding_method": "rope",
12
  "dim": 64
13
  },
14
+ "vocab_size": 1052
15
  }
models/subword_markov/bi_markov_ctx1_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2e9372210a853979f0e4b07d4830a256cffa139a1fece177ff9c6038eed35df5
3
- size 22700
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8524f1a9c10520cc90998cb812247a37e806c4685ba4e3d908cfe064d56fe25
3
+ size 22675
models/subword_markov/bi_markov_ctx1_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 1,
3
  "variant": "subword",
4
  "language": "bi",
5
- "unique_contexts": 364,
6
- "total_transitions": 308852
7
  }
 
2
  "context_size": 1,
3
  "variant": "subword",
4
  "language": "bi",
5
+ "unique_contexts": 362,
6
+ "total_transitions": 308808
7
  }
models/subword_markov/bi_markov_ctx2_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6ca6b0d701fee7db455da7e30950385f2806f14cec4f336ce82c4779e09621b4
3
- size 91544
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34494d2c3fd7277869199119468c66a51dacc18e713456fcb18965e6b21e05a3
3
+ size 89295
models/subword_markov/bi_markov_ctx2_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 2,
3
  "variant": "subword",
4
  "language": "bi",
5
- "unique_contexts": 2361,
6
- "total_transitions": 307409
7
  }
 
2
  "context_size": 2,
3
  "variant": "subword",
4
  "language": "bi",
5
+ "unique_contexts": 2350,
6
+ "total_transitions": 307368
7
  }
models/subword_markov/bi_markov_ctx3_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:68a8c133f845c55867a0fe53a9d2c945b085e56ce9e1e895558840e0a31ab900
3
- size 279694
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:57d3e82e14b72150fc4f4cc52d7d27341f4386a1f1c03172174f9aed47b3cd22
3
+ size 274739
models/subword_markov/bi_markov_ctx3_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 3,
3
  "variant": "subword",
4
  "language": "bi",
5
- "unique_contexts": 12016,
6
- "total_transitions": 305966
7
  }
 
2
  "context_size": 3,
3
  "variant": "subword",
4
  "language": "bi",
5
+ "unique_contexts": 12029,
6
+ "total_transitions": 305928
7
  }
models/subword_markov/bi_markov_ctx4_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6e769ec2bcec378d6b983577c7020c20977d3c068c15135f0bc9e18b6c7e264f
3
- size 620331
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24f23097b8ef424ce0be18fd51ceab87432f9a9ea51a0dc91ba21933142679f3
3
+ size 614238
models/subword_markov/bi_markov_ctx4_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 4,
3
  "variant": "subword",
4
  "language": "bi",
5
- "unique_contexts": 38018,
6
- "total_transitions": 304523
7
  }
 
2
  "context_size": 4,
3
  "variant": "subword",
4
  "language": "bi",
5
+ "unique_contexts": 38112,
6
+ "total_transitions": 304488
7
  }
models/subword_ngram/bi_2gram_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4527bcf666ffd2dd1c578cfc5840d02a71199e369b0c8a627659726d39e176f6
3
- size 14130
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e4829e48479c7969991243f36dfea6f4e6d74ea2f630ea2043c13d2ca2c89511
3
+ size 13975
models/subword_ngram/bi_2gram_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "n": 2,
3
  "variant": "subword",
4
  "language": "bi",
5
- "unique_ngrams": 983,
6
- "total_ngrams": 308852
7
  }
 
2
  "n": 2,
3
  "variant": "subword",
4
  "language": "bi",
5
+ "unique_ngrams": 976,
6
+ "total_ngrams": 308808
7
  }
models/subword_ngram/bi_3gram_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b80fdffb1d982a5003ad057c3864a9f483b951ec85c6662f9c20d596201a26ff
3
- size 63627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f5fd82b92dfc6d4356afce05270d72ac4e39277a4cfecdc7ed558590eaed9ee
3
+ size 63337
models/subword_ngram/bi_3gram_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "n": 3,
3
  "variant": "subword",
4
  "language": "bi",
5
- "unique_ngrams": 5848,
6
- "total_ngrams": 307409
7
  }
 
2
  "n": 3,
3
  "variant": "subword",
4
  "language": "bi",
5
+ "unique_ngrams": 5825,
6
+ "total_ngrams": 307368
7
  }
models/subword_ngram/bi_4gram_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c4ebfe1944f1ab6833eb266e7debe2291ba5a0381a209e1ff6ac58781421a2c2
3
- size 221014
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:20e7ff332ff6bbddf062a55c5f6828a679b47b4c79847b4ba098c8a3eccea847
3
+ size 219083
models/subword_ngram/bi_4gram_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "n": 4,
3
  "variant": "subword",
4
  "language": "bi",
5
- "unique_ngrams": 19225,
6
- "total_ngrams": 305966
7
  }
 
2
  "n": 4,
3
  "variant": "subword",
4
  "language": "bi",
5
+ "unique_ngrams": 19179,
6
+ "total_ngrams": 305928
7
  }
models/subword_ngram/bi_5gram_subword.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de55d59bf119f4b44768dc3b459ec322e227c6b8d17fb1a1a538913bae9a97fe
3
+ size 319412
models/subword_ngram/bi_5gram_subword_metadata.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "n": 5,
3
+ "variant": "subword",
4
+ "language": "bi",
5
+ "unique_ngrams": 26363,
6
+ "total_ngrams": 304488
7
+ }
models/tokenizer/bi_tokenizer_16k.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cb93c36553e5bad9dc7098178cff3fc9a58ec9892899f622abc48bd993de075f
3
- size 506618
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e828f5e2c1b1fc98dac5a27a95c31b59be52a8402143d113689a11719d8ace9a
3
+ size 506724
models/tokenizer/bi_tokenizer_16k.vocab CHANGED
The diff for this file is too large to render. See raw diff
 
models/tokenizer/bi_tokenizer_8k.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6a4efeb4ad5e3a8b678a5c79bd0331f01122c9bb189bbfd294c1172390d5a05c
3
- size 370520
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28f725c6cb47e38ae99fa39137b89f2abb928cc14a1ee5d46ef16a3017e64106
3
+ size 370903
models/tokenizer/bi_tokenizer_8k.vocab CHANGED
The diff for this file is too large to render. See raw diff
 
models/vocabulary/bi_vocabulary.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:87f70e9b39cbf89b9a192034747f166c231ea9bf458bb2112e79e97ae15e247a
3
- size 50916
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33e06d16f52a22d02ec5ebd57923482baa956d867f4a4e6664539ddda180960e
3
+ size 50487
models/vocabulary/bi_vocabulary_metadata.json CHANGED
@@ -1,16 +1,16 @@
1
  {
2
  "language": "bi",
3
- "vocabulary_size": 3117,
4
  "variant": "full",
5
  "statistics": {
6
- "type_token_ratio": 0.15550018456995202,
7
  "coverage": {
8
- "top_100": 0.5605020302694721,
9
- "top_1000": 0.7978589885566629,
10
- "top_5000": 0.9367847914359543
11
  },
12
- "hapax_count": 5308,
13
- "hapax_ratio": 0.6300296735905044,
14
- "total_documents": 1443
15
  }
16
  }
 
1
  {
2
  "language": "bi",
3
+ "vocabulary_size": 3106,
4
  "variant": "full",
5
  "statistics": {
6
+ "type_token_ratio": 0.15657562289065527,
7
  "coverage": {
8
+ "top_100": 0.5595227117643804,
9
+ "top_1000": 0.7971893845784999,
10
+ "top_5000": 0.9356361691533113
11
  },
12
+ "hapax_count": 5384,
13
+ "hapax_ratio": 0.6341578327444052,
14
+ "total_documents": 1440
15
  }
16
  }
models/word_markov/bi_markov_ctx1_word.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5bcf4783672108d39279e5b6f88455605e79e8765b7ffe3f48acaac9ccda1c56
3
- size 230383
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ecb13023f55478121529681449b7ba5a4ee949d8ecb72247965a859fcffcc5e2
3
+ size 231420
models/word_markov/bi_markov_ctx1_word_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 1,
3
  "variant": "word",
4
  "language": "bi",
5
- "unique_contexts": 8338,
6
- "total_transitions": 52737
7
  }
 
2
  "context_size": 1,
3
  "variant": "word",
4
  "language": "bi",
5
+ "unique_contexts": 8408,
6
+ "total_transitions": 52783
7
  }
models/word_markov/bi_markov_ctx2_word.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c04695f69e6a2692e750e9bff816affd14058dc6384bd028ad4dea65463972ae
3
- size 440900
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a773d109fde8766031a00b61fe2c016e512310440b2793f45f0c3b3ccf3d501
3
+ size 441478
models/word_markov/bi_markov_ctx2_word_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 2,
3
  "variant": "word",
4
  "language": "bi",
5
- "unique_contexts": 24957,
6
- "total_transitions": 51294
7
  }
 
2
  "context_size": 2,
3
  "variant": "word",
4
  "language": "bi",
5
+ "unique_contexts": 25020,
6
+ "total_transitions": 51343
7
  }
models/word_markov/bi_markov_ctx3_word.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1acfd63f57a3dbfe43f8279cf4b22712ba4b1311bd026f0bab886fd886ead857
3
- size 595208
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fca25f8469c5430353b0b9ba3c70930045f3bc7297b5e8d5418d50fa61aa1094
3
+ size 596165
models/word_markov/bi_markov_ctx3_word_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 3,
3
  "variant": "word",
4
  "language": "bi",
5
- "unique_contexts": 34724,
6
- "total_transitions": 49851
7
  }
 
2
  "context_size": 3,
3
  "variant": "word",
4
  "language": "bi",
5
+ "unique_contexts": 34806,
6
+ "total_transitions": 49903
7
  }
models/word_markov/bi_markov_ctx4_word.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:19edc02c0845063d5a953f889ccfbfd26a2b1ec828f00bb70ee81b9c16697c22
3
- size 659846
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6503dd4b9dff191c13926ba12557bf311d6b5db8c6f259a39a282c2b106887a
3
+ size 663890
models/word_markov/bi_markov_ctx4_word_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 4,
3
  "variant": "word",
4
  "language": "bi",
5
- "unique_contexts": 38736,
6
- "total_transitions": 48408
7
  }
 
2
  "context_size": 4,
3
  "variant": "word",
4
  "language": "bi",
5
+ "unique_contexts": 38812,
6
+ "total_transitions": 48463
7
  }