omarkamali commited on
Commit
bea2839
·
verified ·
1 Parent(s): 1352f13

Upload all models and assets for bdr (latest)

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +1 -0
  2. README.md +171 -134
  3. models/embeddings/aligned/bdr_128d.bin +3 -0
  4. models/embeddings/aligned/bdr_128d.meta.json +1 -0
  5. models/embeddings/aligned/bdr_128d.projection.npy +3 -0
  6. models/embeddings/aligned/bdr_128d_metadata.json +8 -0
  7. models/embeddings/aligned/bdr_32d.bin +3 -0
  8. models/embeddings/aligned/bdr_32d.meta.json +1 -0
  9. models/embeddings/aligned/bdr_32d.projection.npy +3 -0
  10. models/embeddings/aligned/bdr_32d_metadata.json +8 -0
  11. models/embeddings/aligned/bdr_64d.bin +3 -0
  12. models/embeddings/aligned/bdr_64d.meta.json +1 -0
  13. models/embeddings/aligned/bdr_64d.projection.npy +3 -0
  14. models/embeddings/aligned/bdr_64d_metadata.json +8 -0
  15. models/embeddings/monolingual/bdr_128d.bin +2 -2
  16. models/embeddings/monolingual/bdr_128d_metadata.json +1 -1
  17. models/embeddings/monolingual/bdr_32d.bin +2 -2
  18. models/embeddings/monolingual/bdr_32d_metadata.json +1 -1
  19. models/embeddings/monolingual/bdr_64d.bin +2 -2
  20. models/embeddings/monolingual/bdr_64d_metadata.json +1 -1
  21. models/subword_markov/bdr_markov_ctx1_subword.parquet +2 -2
  22. models/subword_markov/bdr_markov_ctx1_subword_metadata.json +2 -2
  23. models/subword_markov/bdr_markov_ctx2_subword.parquet +2 -2
  24. models/subword_markov/bdr_markov_ctx2_subword_metadata.json +2 -2
  25. models/subword_markov/bdr_markov_ctx3_subword.parquet +2 -2
  26. models/subword_markov/bdr_markov_ctx3_subword_metadata.json +2 -2
  27. models/subword_markov/bdr_markov_ctx4_subword.parquet +2 -2
  28. models/subword_markov/bdr_markov_ctx4_subword_metadata.json +2 -2
  29. models/subword_ngram/bdr_2gram_subword.parquet +2 -2
  30. models/subword_ngram/bdr_2gram_subword_metadata.json +2 -2
  31. models/subword_ngram/bdr_3gram_subword.parquet +2 -2
  32. models/subword_ngram/bdr_3gram_subword_metadata.json +2 -2
  33. models/subword_ngram/bdr_4gram_subword.parquet +2 -2
  34. models/subword_ngram/bdr_4gram_subword_metadata.json +2 -2
  35. models/subword_ngram/bdr_5gram_subword.parquet +3 -0
  36. models/subword_ngram/bdr_5gram_subword_metadata.json +7 -0
  37. models/tokenizer/bdr_tokenizer_8k.model +2 -2
  38. models/tokenizer/bdr_tokenizer_8k.vocab +0 -0
  39. models/vocabulary/bdr_vocabulary.parquet +2 -2
  40. models/vocabulary/bdr_vocabulary_metadata.json +7 -7
  41. models/word_markov/bdr_markov_ctx1_word.parquet +2 -2
  42. models/word_markov/bdr_markov_ctx1_word_metadata.json +2 -2
  43. models/word_markov/bdr_markov_ctx2_word.parquet +2 -2
  44. models/word_markov/bdr_markov_ctx2_word_metadata.json +2 -2
  45. models/word_markov/bdr_markov_ctx3_word.parquet +2 -2
  46. models/word_markov/bdr_markov_ctx3_word_metadata.json +2 -2
  47. models/word_markov/bdr_markov_ctx4_word.parquet +2 -2
  48. models/word_markov/bdr_markov_ctx4_word_metadata.json +2 -2
  49. models/word_ngram/bdr_2gram_word.parquet +2 -2
  50. models/word_ngram/bdr_2gram_word_metadata.json +1 -1
.gitattributes CHANGED
@@ -38,3 +38,4 @@ visualizations/performance_dashboard.png filter=lfs diff=lfs merge=lfs -text
38
  visualizations/tsne_sentences.png filter=lfs diff=lfs merge=lfs -text
39
  visualizations/tsne_words.png filter=lfs diff=lfs merge=lfs -text
40
  visualizations/zipf_law.png filter=lfs diff=lfs merge=lfs -text
 
 
38
  visualizations/tsne_sentences.png filter=lfs diff=lfs merge=lfs -text
39
  visualizations/tsne_words.png filter=lfs diff=lfs merge=lfs -text
40
  visualizations/zipf_law.png filter=lfs diff=lfs merge=lfs -text
41
+ visualizations/embedding_tsne_multilingual.png filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  language: bdr
3
- language_name: BDR
4
  language_family: austronesian_other
5
  tags:
6
  - wikilangs
@@ -10,11 +10,21 @@ tags:
10
  - n-gram
11
  - markov
12
  - wikipedia
 
 
 
 
 
 
 
 
 
 
13
  - monolingual
14
  - family-austronesian_other
15
  license: mit
16
  library_name: wikilangs
17
- pipeline_tag: feature-extraction
18
  datasets:
19
  - omarkamali/wikipedia-monthly
20
  dataset_info:
@@ -23,20 +33,20 @@ dataset_info:
23
  metrics:
24
  - name: best_compression_ratio
25
  type: compression
26
- value: 4.792
27
  - name: best_isotropy
28
  type: isotropy
29
- value: 0.0482
30
  - name: vocabulary_size
31
  type: vocab
32
  value: 0
33
  generated: 2026-01-03
34
  ---
35
 
36
- # BDR - Wikilangs Models
37
  ## Comprehensive Research Report & Full Ablation Study
38
 
39
- This repository contains NLP models trained and evaluated by Wikilangs, specifically on **BDR** Wikipedia data.
40
  We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and word embeddings.
41
 
42
  ## 📋 Repository Contents
@@ -60,7 +70,7 @@ We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and
60
  - [3. Markov Chain Evaluation](#3-markov-chain-evaluation)
61
  - [4. Vocabulary Analysis](#4-vocabulary-analysis)
62
  - [5. Word Embeddings Evaluation](#5-word-embeddings-evaluation)
63
- - [6. Morphological Analysis (Experimental)](#6-morphological-analysis)
64
  - [7. Summary & Recommendations](#7-summary--recommendations)
65
  - [Metrics Glossary](#appendix-metrics-glossary--interpretation-guide)
66
  - [Visualizations Index](#visualizations-index)
@@ -80,35 +90,35 @@ We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and
80
 
81
  | Vocab Size | Compression | Avg Token Len | UNK Rate | Total Tokens |
82
  |------------|-------------|---------------|----------|--------------|
83
- | **8k** | 4.792x 🏆 | 4.81 | 0.1661% | 33,107 |
84
 
85
  ### Tokenization Examples
86
 
87
  Below are sample sentences tokenized with each vocabulary size:
88
 
89
- **Sample 1:** `Nimbug iyono indu' manuk nuut ngentelo ta' keteraan manuk lain.`
90
 
91
  | Vocab | Tokens | Count |
92
  |-------|--------|-------|
93
- | 8k | `▁nimbugiyonoindu 'manuknuutngentelota 'keteraan ... (+3 more)` | 13 |
94
 
95
- **Sample 2:** `Raja iyo no' dangan jomo kuleh kuasa diom pemerintah dikau kerajaan.Endo rojo pi...`
96
 
97
  | Vocab | Tokens | Count |
98
  |-------|--------|-------|
99
- | 8k | `▁rajaiyono 'dangan ▁jomokulehkuasadiom ▁pemerintah ... (+20 more)` | 30 |
100
 
101
- **Sample 3:** `Para-para iyo no tempat ngena segala barang enjata rak`
102
 
103
  | Vocab | Tokens | Count |
104
  |-------|--------|-------|
105
- | 8k | `▁para - para iyonotempatngenasegalabarangenjata ... (+1 more)` | 11 |
106
 
107
 
108
  ### Key Findings
109
 
110
- - **Best Compression:** 8k achieves 4.792x compression
111
- - **Lowest UNK Rate:** 8k with 0.1661% unknown tokens
112
  - **Trade-off:** Larger vocabularies improve compression but increase model size
113
  - **Recommendation:** 32k vocabulary provides optimal balance for production use
114
 
@@ -126,11 +136,13 @@ Below are sample sentences tokenized with each vocabulary size:
126
  | N-gram | Variant | Perplexity | Entropy | Unique N-grams | Top-100 Coverage | Top-1000 Coverage |
127
  |--------|---------|------------|---------|----------------|------------------|-------------------|
128
  | **2-gram** | Word | 287 | 8.16 | 401 | 53.3% | 100.0% |
129
- | **2-gram** | Subword | 181 🏆 | 7.50 | 597 | 77.1% | 100.0% |
130
- | **3-gram** | Word | 221 | 7.79 | 271 | 59.6% | 100.0% |
131
- | **3-gram** | Subword | 1,140 | 10.15 | 3,421 | 32.8% | 85.1% |
132
- | **4-gram** | Word | 273 | 8.09 | 346 | 51.1% | 100.0% |
133
- | **4-gram** | Subword | 4,426 | 12.11 | 11,413 | 17.0% | 52.6% |
 
 
134
 
135
  ### Top 5 N-grams by Size
136
 
@@ -139,7 +151,7 @@ Below are sample sentences tokenized with each vocabulary size:
139
  | Rank | N-gram | Count |
140
  |------|--------|-------|
141
  | 1 | `tungan metelak` | 162 |
142
- | 2 | `iyo no` | 138 |
143
  | 3 | `iyo noh` | 69 |
144
  | 4 | `iyo tu` | 68 |
145
  | 5 | `bioso ni` | 45 |
@@ -164,42 +176,62 @@ Below are sample sentences tokenized with each vocabulary size:
164
  | 4 | `iyo no endangan jomo` | 12 |
165
  | 5 | `no endangan jomo politik` | 12 |
166
 
 
 
 
 
 
 
 
 
 
 
167
  **2-grams (Subword):**
168
 
169
  | Rank | N-gram | Count |
170
  |------|--------|-------|
171
- | 1 | `a n` | 5,437 |
172
- | 2 | `n _` | 3,734 |
173
- | 3 | `n g` | 3,473 |
174
- | 4 | `i _` | 3,019 |
175
- | 5 | `_ t` | 2,998 |
176
 
177
  **3-grams (Subword):**
178
 
179
  | Rank | N-gram | Count |
180
  |------|--------|-------|
181
- | 1 | `a n _` | 2,443 |
182
- | 2 | `a n g` | 1,577 |
183
- | 3 | `n g _` | 1,357 |
184
- | 4 | `_ t a` | 1,076 |
185
- | 5 | `_ n i` | 987 |
186
 
187
  **4-grams (Subword):**
188
 
189
  | Rank | N-gram | Count |
190
  |------|--------|-------|
191
- | 1 | `a n g _` | 910 |
192
- | 2 | `_ n i _` | 650 |
193
- | 3 | `_ i y o` | 643 |
194
- | 4 | `n g a n` | 619 |
195
- | 5 | `g a n _` | 579 |
 
 
 
 
 
 
 
 
 
 
196
 
197
 
198
  ### Key Findings
199
 
200
- - **Best Perplexity:** 2-gram (subword) with 181
201
  - **Entropy Trend:** Decreases with larger n-grams (more predictable)
202
- - **Coverage:** Top-1000 patterns cover ~53% of corpus
203
  - **Recommendation:** 4-gram or 5-gram for best predictive performance
204
 
205
  ---
@@ -215,14 +247,14 @@ Below are sample sentences tokenized with each vocabulary size:
215
 
216
  | Context | Variant | Avg Entropy | Perplexity | Branching Factor | Unique Contexts | Predictability |
217
  |---------|---------|-------------|------------|------------------|-----------------|----------------|
218
- | **1** | Word | 0.8053 | 1.747 | 3.60 | 5,241 | 19.5% |
219
- | **1** | Subword | 1.4652 | 2.761 | 11.14 | 104 | 0.0% |
220
- | **2** | Word | 0.1666 | 1.122 | 1.26 | 18,592 | 83.3% |
221
- | **2** | Subword | 1.1951 | 2.290 | 5.74 | 1,154 | 0.0% |
222
- | **3** | Word | 0.0377 | 1.026 | 1.05 | 22,996 | 96.2% |
223
- | **3** | Subword | 0.7985 | 1.739 | 3.15 | 6,603 | 20.2% |
224
- | **4** | Word | 0.0104 🏆 | 1.007 | 1.01 | 23,590 | 99.0% |
225
- | **4** | Subword | 0.5443 | 1.458 | 2.09 | 20,699 | 45.6% |
226
 
227
  ### Generated Text Samples (Word-based)
228
 
@@ -230,27 +262,27 @@ Below are text samples generated from each word-based Markov chain model:
230
 
231
  **Context Size 1:**
232
 
233
- 1. `ni ta lok kuah engko tangsi selegubdi tu terhasil moko dangan pelego dendo malaysia beliau tu`
234
- 2. `tu tungan ni un duo ni mediam tepung buas tak sekul tena tana amun pinapi enggo`
235
- 3. `iyo boi nilego oleg ni jomo yang bok ni pan akan buan raya kota belud tu`
236
 
237
  **Context Size 2:**
238
 
239
- 1. `iyo no nyaun preskripsi toos bineli ta farmasi atau mediam kadai yang nyaun sebarang halangan engko ...`
240
- 2. `iyo noh kui tradisional jomo mitu sabah kui tu bentuk ni dokon indung jari engko binuat lua`
241
- 3. `iyo tu boi ni urus le ni gua a masi un sampai betiru terutama ni sembiang pardu`
242
 
243
  **Context Size 3:**
244
 
245
- 1. `ma na ni teko ta tampat tungan setemu tapi jomo tenemuan ai no lumaan`
246
- 2. `dewan undangan negeri sabah ta kewasan tempasuk lua tungan metelak politik malaysia di pertua laat a...`
247
- 3. `undangan negeri sabah betiru`
248
 
249
  **Context Size 4:**
250
 
251
- 1. `dewan undangan negeri sabah dun lua september tu anggota pertubuhan kebangsaan melayu bersatu malays...`
252
- 2. `sama ma na ni ai ngemban matai`
253
- 3. `ahli dewan undangan negeri sabah dewan undangan negeri sabah dewan undangan negeri sabah ta kewasan ...`
254
 
255
 
256
  ### Generated Text Samples (Subword-based)
@@ -259,34 +291,34 @@ Below are text samples generated from each subword-based Markov chain model:
259
 
260
  **Context Size 1:**
261
 
262
- 1. `_njano-bo_cseria`
263
- 2. `anegal_8_t_bu"_b`
264
- 3. `ngim_nd_bo_isaup`
265
 
266
  **Context Size 2:**
267
 
268
- 1. `an_kain_tamuḥamal`
269
- 2. `n_jom_no_turi_mud`
270
- 3. `ngko_ta_tang_boi_`
271
 
272
  **Context Size 3:**
273
 
274
- 1. `an_ni_ana'_nakasal`
275
- 2. `ang_jomo_untuan_ta`
276
- 3. `ng_teali_pulo_ko'_`
277
 
278
  **Context Size 4:**
279
 
280
- 1. `ang_sefalopod_lua'_`
281
- 2. `_ni_denga_septembag`
282
- 3. `_iyo_no_telia_punya`
283
 
284
 
285
  ### Key Findings
286
 
287
  - **Best Predictability:** Context-4 (word) with 99.0% predictability
288
  - **Branching Factor:** Decreases with context size (more deterministic)
289
- - **Memory Trade-off:** Larger contexts require more storage (20,699 contexts)
290
  - **Recommendation:** Context-3 or Context-4 for text generation
291
 
292
  ---
@@ -302,25 +334,25 @@ Below are text samples generated from each subword-based Markov chain model:
302
 
303
  | Metric | Value |
304
  |--------|-------|
305
- | Vocabulary Size | 2,342 |
306
- | Total Tokens | 23,366 |
307
- | Mean Frequency | 9.98 |
308
  | Median Frequency | 3 |
309
- | Frequency Std Dev | 33.27 |
310
 
311
  ### Most Common Words
312
 
313
  | Rank | Word | Frequency |
314
  |------|------|-----------|
315
- | 1 | ni | 760 |
316
- | 2 | tu | 584 |
317
- | 3 | iyo | 549 |
318
- | 4 | ta | 455 |
319
- | 5 | yang | 382 |
320
- | 6 | boi | 354 |
321
  | 7 | pan | 303 |
322
  | 8 | kok | 280 |
323
- | 9 | jomo | 275 |
324
  | 10 | tungan | 250 |
325
 
326
  ### Least Common Words (from vocabulary)
@@ -342,24 +374,24 @@ Below are text samples generated from each subword-based Markov chain model:
342
 
343
  | Metric | Value |
344
  |--------|-------|
345
- | Zipf Coefficient | 0.9532 |
346
- | R² (Goodness of Fit) | 0.984280 |
347
  | Adherence Quality | **excellent** |
348
 
349
  ### Coverage Analysis
350
 
351
  | Top N Words | Coverage |
352
  |-------------|----------|
353
- | Top 100 | 45.5% |
354
- | Top 1,000 | 85.6% |
355
  | Top 5,000 | 0.0% |
356
  | Top 10,000 | 0.0% |
357
 
358
  ### Key Findings
359
 
360
  - **Zipf Compliance:** R²=0.9843 indicates excellent adherence to Zipf's law
361
- - **High Frequency Dominance:** Top 100 words cover 45.5% of corpus
362
- - **Long Tail:** -7,658 words needed for remaining 100.0% coverage
363
 
364
  ---
365
  ## 5. Word Embeddings Evaluation
@@ -375,37 +407,40 @@ Below are text samples generated from each subword-based Markov chain model:
375
 
376
  ### 5.1 Cross-Lingual Alignment
377
 
378
- > *Note: Multilingual alignment visualization not available for this language.*
 
 
379
 
380
 
381
  ### 5.2 Model Comparison
382
 
383
  | Model | Dimension | Isotropy | Semantic Density | Alignment R@1 | Alignment R@10 |
384
  |-------|-----------|----------|------------------|---------------|----------------|
385
- | **mono_32d** | 32 | 0.0482 🏆 | 0.8825 | N/A | N/A |
386
- | **mono_64d** | 64 | 0.0132 | 0.9050 | N/A | N/A |
387
- | **mono_128d** | 128 | 0.0053 | 0.9273 | N/A | N/A |
 
 
 
388
 
389
  ### Key Findings
390
 
391
- - **Best Isotropy:** mono_32d with 0.0482 (more uniform distribution)
392
- - **Semantic Density:** Average pairwise similarity of 0.9049. Lower values indicate better semantic separation.
393
- - **Alignment Quality:** No aligned models evaluated in this run.
394
  - **Recommendation:** 128d aligned for best cross-lingual performance
395
 
396
  ---
397
  ## 6. Morphological Analysis (Experimental)
398
 
399
- > ⚠️ **Warning:** This language shows low morphological productivity. The statistical signals used for this analysis may be noisy or less reliable than for morphologically rich languages.
400
-
401
  This section presents an automated morphological analysis derived from the statistical divergence between word-level and subword-level models. By analyzing where subword predictability spikes and where word-level coverage fails, we can infer linguistic structures without supervised data.
402
 
403
  ### 6.1 Productivity & Complexity
404
 
405
  | Metric | Value | Interpretation | Recommendation |
406
  |--------|-------|----------------|----------------|
407
- | Productivity Index | **0.000** | Low morphological productivity | ⚠️ Likely unreliable |
408
- | Idiomaticity Gap | **-1.000** | Low formulaic content | - |
409
 
410
  ### 6.2 Affix Inventory (Productive Units)
411
 
@@ -414,21 +449,21 @@ These are the most productive prefixes and suffixes identified by sampling the v
414
  #### Productive Prefixes
415
  | Prefix | Examples |
416
  |--------|----------|
417
- | `-pe` | petaling, pekakas, peketa |
418
- | `-se` | sejak, seniram, sejati |
419
- | `-ke` | kenangan, kerita, keratas |
420
- | `-te` | tehe, tempoh, tetiak |
421
- | `-me` | meruma, menurut, melioro |
422
- | `-be` | berukuran, berfikir, benua |
423
 
424
  #### Productive Suffixes
425
  | Suffix | Examples |
426
  |--------|----------|
427
- | `-n` | kumpulan, regisin, haiwan |
428
- | `-an` | kumpulan, haiwan, berukuran |
429
- | `-ng` | petaling, kantung, ngulang |
430
- | `-ang` | ngulang, manang, sayang |
431
- | `-ah` | tah, fatimah, umrah |
432
 
433
  ### 6.3 Bound Stems (Lexical Roots)
434
 
@@ -443,16 +478,16 @@ This table shows which prefixes and suffixes most frequently co-occur on the sam
443
 
444
  | Prefix | Suffix | Frequency | Examples |
445
  |--------|--------|-----------|----------|
446
- | `-pe` | `-n` | 55 words | pelan, pentaran |
447
- | `-pe` | `-an` | 49 words | pelan, pentaran |
448
- | `-ke` | `-n` | 42 words | kenangan, keteraan |
449
- | `-ke` | `-an` | 36 words | kenangan, keteraan |
450
- | `-se` | `-n` | 11 words | sebahagian, selain |
451
- | `-te` | `-n` | 9 words | temban, tenomon |
452
- | `-se` | `-ng` | 9 words | sedong, sepanjang |
453
- | `-me` | `-n` | 9 words | mesimpon, meluman |
454
- | `-pe` | `-ng` | 7 words | petaling, pelancong |
455
- | `-be` | `-n` | 7 words | berukuran, been |
456
 
457
  ### 6.5 Recursive Morpheme Segmentation
458
 
@@ -461,25 +496,27 @@ Using **Recursive Hierarchical Substitutability**, we decompose complex words in
461
  | Word | Suggested Split | Confidence | Stem |
462
  |------|-----------------|------------|------|
463
  | kebenyakan | **`ke-be-nyak-an`** | 7.5 | `nyak` |
464
- | kebangsaan | **`ke-bangsa-an`** | 6.0 | `bangsa` |
465
- | kelebihan | **`ke-lebih-an`** | 6.0 | `lebih` |
466
- | kelahiran | **`ke-lahir-an`** | 6.0 | `lahir` |
467
  | keramaian | **`ke-ramai-an`** | 6.0 | `ramai` |
468
  | kepulauan | **`ke-pulau-an`** | 6.0 | `pulau` |
469
- | kebudayaan | **`ke-budaya-an`** | 6.0 | `budaya` |
470
  | keputeraan | **`ke-putera-an`** | 6.0 | `putera` |
471
- | sedembila | **`se-dembila`** | 4.5 | `dembila` |
472
- | perpisahan | **`pe-rpis-ah-an`** | 4.5 | `rpis` |
473
- | keselamatan | **`ke-se-lamat-an`** | 4.5 | `lamat` |
474
  | pernikahan | **`pe-rnik-ah-an`** | 4.5 | `rnik` |
475
- | perjuangan | **`pe-rjua-ng-an`** | 4.5 | `rjua` |
 
 
476
  | kemerdekaan | **`ke-me-rdeka-an`** | 4.5 | `rdeka` |
477
- | kepelbagaian | **`ke-pe-lbagai-an`** | 4.5 | `lbagai` |
 
478
 
479
  ### 6.6 Linguistic Interpretation
480
 
481
  > **Automated Insight:**
482
- The language BDR appears to be more isolating or has a highly fixed vocabulary. Word-level models perform nearly as well as subword models, indicating fewer productive morphological processes.
 
 
483
 
484
  ---
485
  ## 7. Summary & Recommendations
@@ -490,8 +527,8 @@ The language BDR appears to be more isolating or has a highly fixed vocabulary.
490
 
491
  | Component | Recommended | Rationale |
492
  |-----------|-------------|-----------|
493
- | Tokenizer | **8k BPE** | Best compression (4.79x) |
494
- | N-gram | **2-gram** | Lowest perplexity (181) |
495
  | Markov | **Context-4** | Highest predictability (99.0%) |
496
  | Embeddings | **100d** | Balanced semantic capture and isotropy |
497
 
@@ -706,4 +743,4 @@ MIT License - Free for academic and commercial use.
706
  ---
707
  *Generated by Wikilangs Models Pipeline*
708
 
709
- *Report Date: 2026-01-03 06:44:23*
 
1
  ---
2
  language: bdr
3
+ language_name: West Coast Bajau
4
  language_family: austronesian_other
5
  tags:
6
  - wikilangs
 
10
  - n-gram
11
  - markov
12
  - wikipedia
13
+ - feature-extraction
14
+ - sentence-similarity
15
+ - tokenization
16
+ - n-grams
17
+ - markov-chain
18
+ - text-mining
19
+ - fasttext
20
+ - babelvec
21
+ - vocabulous
22
+ - vocabulary
23
  - monolingual
24
  - family-austronesian_other
25
  license: mit
26
  library_name: wikilangs
27
+ pipeline_tag: text-generation
28
  datasets:
29
  - omarkamali/wikipedia-monthly
30
  dataset_info:
 
33
  metrics:
34
  - name: best_compression_ratio
35
  type: compression
36
+ value: 4.803
37
  - name: best_isotropy
38
  type: isotropy
39
+ value: 0.0390
40
  - name: vocabulary_size
41
  type: vocab
42
  value: 0
43
  generated: 2026-01-03
44
  ---
45
 
46
+ # West Coast Bajau - Wikilangs Models
47
  ## Comprehensive Research Report & Full Ablation Study
48
 
49
+ This repository contains NLP models trained and evaluated by Wikilangs, specifically on **West Coast Bajau** Wikipedia data.
50
  We analyze tokenizers, n-gram models, Markov chains, vocabulary statistics, and word embeddings.
51
 
52
  ## 📋 Repository Contents
 
70
  - [3. Markov Chain Evaluation](#3-markov-chain-evaluation)
71
  - [4. Vocabulary Analysis](#4-vocabulary-analysis)
72
  - [5. Word Embeddings Evaluation](#5-word-embeddings-evaluation)
73
+ - [6. Morphological Analysis (Experimental)](#6--morphological-analysis-experimental)
74
  - [7. Summary & Recommendations](#7-summary--recommendations)
75
  - [Metrics Glossary](#appendix-metrics-glossary--interpretation-guide)
76
  - [Visualizations Index](#visualizations-index)
 
90
 
91
  | Vocab Size | Compression | Avg Token Len | UNK Rate | Total Tokens |
92
  |------------|-------------|---------------|----------|--------------|
93
+ | **8k** | 4.803x 🏆 | 4.82 | 0.1461% | 32,844 |
94
 
95
  ### Tokenization Examples
96
 
97
  Below are sample sentences tokenized with each vocabulary size:
98
 
99
+ **Sample 1:** `Dugal tu io akan bungkar pedih ni amun niak mangam buas dembangi , Dugal tu baya...`
100
 
101
  | Vocab | Tokens | Count |
102
  |-------|--------|-------|
103
+ | 8k | `▁dugaltuio ▁akanbungkarpedihniamun ▁niakmangam ... (+15 more)` | 25 |
104
 
105
+ **Sample 2:** `Bul (Ling Melayu: Bola) iyo dembua barang pinakai untuk besukan`
106
 
107
  | Vocab | Tokens | Count |
108
  |-------|--------|-------|
109
+ | 8k | `▁bul( ling melayu :bola )iyodembuabarang ... (+3 more)` | 13 |
110
 
111
+ **Sample 3:** `Tupi sungku tu sejenis tupi tradisional jomo sama. Tupi sungku pinakai untuk nge...`
112
 
113
  | Vocab | Tokens | Count |
114
  |-------|--------|-------|
115
+ | 8k | `▁tupi ▁sungkutusejenistupitradisionaljomosama . tupi ... (+10 more)` | 20 |
116
 
117
 
118
  ### Key Findings
119
 
120
+ - **Best Compression:** 8k achieves 4.803x compression
121
+ - **Lowest UNK Rate:** 8k with 0.1461% unknown tokens
122
  - **Trade-off:** Larger vocabularies improve compression but increase model size
123
  - **Recommendation:** 32k vocabulary provides optimal balance for production use
124
 
 
136
  | N-gram | Variant | Perplexity | Entropy | Unique N-grams | Top-100 Coverage | Top-1000 Coverage |
137
  |--------|---------|------------|---------|----------------|------------------|-------------------|
138
  | **2-gram** | Word | 287 | 8.16 | 401 | 53.3% | 100.0% |
139
+ | **2-gram** | Subword | 180 | 7.49 | 593 | 77.1% | 100.0% |
140
+ | **3-gram** | Word | 219 | 7.78 | 269 | 59.9% | 100.0% |
141
+ | **3-gram** | Subword | 1,136 | 10.15 | 3,407 | 32.8% | 85.1% |
142
+ | **4-gram** | Word | 272 | 8.09 | 345 | 51.2% | 100.0% |
143
+ | **4-gram** | Subword | 4,404 | 12.10 | 11,348 | 17.0% | 52.7% |
144
+ | **5-gram** | Word | 110 🏆 | 6.78 | 144 | 81.2% | 100.0% |
145
+ | **5-gram** | Subword | 8,815 | 13.11 | 18,466 | 12.6% | 37.5% |
146
 
147
  ### Top 5 N-grams by Size
148
 
 
151
  | Rank | N-gram | Count |
152
  |------|--------|-------|
153
  | 1 | `tungan metelak` | 162 |
154
+ | 2 | `iyo no` | 137 |
155
  | 3 | `iyo noh` | 69 |
156
  | 4 | `iyo tu` | 68 |
157
  | 5 | `bioso ni` | 45 |
 
176
  | 4 | `iyo no endangan jomo` | 12 |
177
  | 5 | `no endangan jomo politik` | 12 |
178
 
179
+ **5-grams (Word):**
180
+
181
+ | Rank | N-gram | Count |
182
+ |------|--------|-------|
183
+ | 1 | `no endangan jomo politik ta` | 12 |
184
+ | 2 | `iyo no endangan jomo politik` | 12 |
185
+ | 3 | `beliau tu ahli dewan undangan` | 11 |
186
+ | 4 | `malaysia beliau tu ahli dewan` | 11 |
187
+ | 5 | `ahli dewan undangan negeri sabah` | 11 |
188
+
189
  **2-grams (Subword):**
190
 
191
  | Rank | N-gram | Count |
192
  |------|--------|-------|
193
+ | 1 | `a n` | 5,403 |
194
+ | 2 | `n _` | 3,715 |
195
+ | 3 | `n g` | 3,458 |
196
+ | 4 | `i _` | 3,000 |
197
+ | 5 | `_ t` | 2,981 |
198
 
199
  **3-grams (Subword):**
200
 
201
  | Rank | N-gram | Count |
202
  |------|--------|-------|
203
+ | 1 | `a n _` | 2,433 |
204
+ | 2 | `a n g` | 1,567 |
205
+ | 3 | `n g _` | 1,349 |
206
+ | 4 | `_ t a` | 1,073 |
207
+ | 5 | `_ n i` | 983 |
208
 
209
  **4-grams (Subword):**
210
 
211
  | Rank | N-gram | Count |
212
  |------|--------|-------|
213
+ | 1 | `a n g _` | 903 |
214
+ | 2 | `_ n i _` | 648 |
215
+ | 3 | `_ i y o` | 641 |
216
+ | 4 | `n g a n` | 618 |
217
+ | 5 | `g a n _` | 578 |
218
+
219
+ **5-grams (Subword):**
220
+
221
+ | Rank | N-gram | Count |
222
+ |------|--------|-------|
223
+ | 1 | `n g a n _` | 565 |
224
+ | 2 | `_ i y o _` | 504 |
225
+ | 3 | `y a n g _` | 410 |
226
+ | 4 | `_ y a n g` | 370 |
227
+ | 5 | `_ t a ' _` | 355 |
228
 
229
 
230
  ### Key Findings
231
 
232
+ - **Best Perplexity:** 5-gram (word) with 110
233
  - **Entropy Trend:** Decreases with larger n-grams (more predictable)
234
+ - **Coverage:** Top-1000 patterns cover ~37% of corpus
235
  - **Recommendation:** 4-gram or 5-gram for best predictive performance
236
 
237
  ---
 
247
 
248
  | Context | Variant | Avg Entropy | Perplexity | Branching Factor | Unique Contexts | Predictability |
249
  |---------|---------|-------------|------------|------------------|-----------------|----------------|
250
+ | **1** | Word | 0.8062 | 1.749 | 3.61 | 5,206 | 19.4% |
251
+ | **1** | Subword | 1.4517 | 2.735 | 11.24 | 101 | 0.0% |
252
+ | **2** | Word | 0.1664 | 1.122 | 1.26 | 18,482 | 83.4% |
253
+ | **2** | Subword | 1.2091 | 2.312 | 5.80 | 1,130 | 0.0% |
254
+ | **3** | Word | 0.0377 | 1.026 | 1.05 | 22,853 | 96.2% |
255
+ | **3** | Subword | 0.8020 | 1.744 | 3.16 | 6,542 | 19.8% |
256
+ | **4** | Word | 0.0104 🏆 | 1.007 | 1.01 | 23,441 | 99.0% |
257
+ | **4** | Subword | 0.5453 | 1.459 | 2.09 | 20,556 | 45.5% |
258
 
259
  ### Generated Text Samples (Word-based)
260
 
 
262
 
263
  **Context Size 1:**
264
 
265
+ 1. `ni tak sekolah kebangsanaan puteri ngerujuk ta dikau bumbung laat atau pan buli kinurban iyo no`
266
+ 2. `tu pan kuleh status teralap malaysia diom arena seni soro tungan metelak politik malaysia iko pinaka...`
267
+ 3. `iyo menjadi budaya bajau sama ngeruo elau ule bedagang tradisi boi penenakan tak taun iyo pan`
268
 
269
  **Context Size 2:**
270
 
271
+ 1. `iyo no dangan jomo mediom menjogo keselamatan ko kestabilan masyarakat nuut ta diom undang undang lu...`
272
+ 2. `iyo noh tun dr hasmah binti haji mohamad ali nganak 12 julai hasmah iyono doktor dendo yang`
273
+ 3. `iyo tu pelego pemuzik kok pelakun dendo malaysia iyo tekilo kok watak ni lua kawasan asahan sumatera`
274
 
275
  **Context Size 3:**
276
 
277
+ 1. `ma na ni dediki bana`
278
+ 2. `dewan undangan negeri sabah dewan undangan negeri sabah betiru`
279
+ 3. `undangan negeri sabah boi nilantik lua 8 oktober beliau betiru ngentan jewatan ketua parti islam se ...`
280
 
281
  **Context Size 4:**
282
 
283
+ 1. `dewan undangan negeri sabah betiru`
284
+ 2. `sama ma na ni telampau oyo antawa oyo bana`
285
+ 3. `ahli dewan undangan negeri sabah dewan undangan negeri sabah ta kewasan matunggong lua tungan metela...`
286
 
287
 
288
  ### Generated Text Samples (Subword-based)
 
291
 
292
  **Context Size 1:**
293
 
294
+ 1. `_amagri_dintutha`
295
+ 2. `a'_tandartino_il`
296
+ 3. `ntan_bik_di_a_bi`
297
 
298
  **Context Size 2:**
299
 
300
+ 1. `angerangerebini_p`
301
+ 2. `n_tau_us_amungine`
302
+ 3. `ng._tan_fa_langha`
303
 
304
  **Context Size 3:**
305
 
306
+ 1. `an_ole_ta'_mapas_d`
307
+ 2. `ang_ma'na_tang_di_`
308
+ 3. `ng_un_pan_atley_ma`
309
 
310
  **Context Size 4:**
311
 
312
+ 1. `ang_semek_regisin,_`
313
+ 2. `_ni_untuk_pelbagas_`
314
+ 3. `_iyo_no_un_pakai_bi`
315
 
316
 
317
  ### Key Findings
318
 
319
  - **Best Predictability:** Context-4 (word) with 99.0% predictability
320
  - **Branching Factor:** Decreases with context size (more deterministic)
321
+ - **Memory Trade-off:** Larger contexts require more storage (20,556 contexts)
322
  - **Recommendation:** Context-3 or Context-4 for text generation
323
 
324
  ---
 
334
 
335
  | Metric | Value |
336
  |--------|-------|
337
+ | Vocabulary Size | 2,333 |
338
+ | Total Tokens | 23,243 |
339
+ | Mean Frequency | 9.96 |
340
  | Median Frequency | 3 |
341
+ | Frequency Std Dev | 33.24 |
342
 
343
  ### Most Common Words
344
 
345
  | Rank | Word | Frequency |
346
  |------|------|-----------|
347
+ | 1 | ni | 758 |
348
+ | 2 | tu | 583 |
349
+ | 3 | iyo | 548 |
350
+ | 4 | ta | 452 |
351
+ | 5 | yang | 381 |
352
+ | 6 | boi | 353 |
353
  | 7 | pan | 303 |
354
  | 8 | kok | 280 |
355
+ | 9 | jomo | 273 |
356
  | 10 | tungan | 250 |
357
 
358
  ### Least Common Words (from vocabulary)
 
374
 
375
  | Metric | Value |
376
  |--------|-------|
377
+ | Zipf Coefficient | 0.9534 |
378
+ | R² (Goodness of Fit) | 0.984288 |
379
  | Adherence Quality | **excellent** |
380
 
381
  ### Coverage Analysis
382
 
383
  | Top N Words | Coverage |
384
  |-------------|----------|
385
+ | Top 100 | 45.6% |
386
+ | Top 1,000 | 85.7% |
387
  | Top 5,000 | 0.0% |
388
  | Top 10,000 | 0.0% |
389
 
390
  ### Key Findings
391
 
392
  - **Zipf Compliance:** R²=0.9843 indicates excellent adherence to Zipf's law
393
+ - **High Frequency Dominance:** Top 100 words cover 45.6% of corpus
394
+ - **Long Tail:** -7,667 words needed for remaining 100.0% coverage
395
 
396
  ---
397
  ## 5. Word Embeddings Evaluation
 
407
 
408
  ### 5.1 Cross-Lingual Alignment
409
 
410
+ ![Alignment Quality](visualizations/embedding_alignment_quality.png)
411
+
412
+ ![Multilingual t-SNE](visualizations/embedding_tsne_multilingual.png)
413
 
414
 
415
  ### 5.2 Model Comparison
416
 
417
  | Model | Dimension | Isotropy | Semantic Density | Alignment R@1 | Alignment R@10 |
418
  |-------|-----------|----------|------------------|---------------|----------------|
419
+ | **mono_32d** | 32 | 0.0390 🏆 | 0.8823 | N/A | N/A |
420
+ | **mono_64d** | 64 | 0.0242 | 0.9395 | N/A | N/A |
421
+ | **mono_128d** | 128 | 0.0071 | 0.9393 | N/A | N/A |
422
+ | **aligned_32d** | 32 | 0.0390 | 0.8940 | 0.0078 | 0.0667 |
423
+ | **aligned_64d** | 64 | 0.0242 | 0.9410 | 0.0039 | 0.0627 |
424
+ | **aligned_128d** | 128 | 0.0071 | 0.9401 | 0.0039 | 0.0627 |
425
 
426
  ### Key Findings
427
 
428
+ - **Best Isotropy:** mono_32d with 0.0390 (more uniform distribution)
429
+ - **Semantic Density:** Average pairwise similarity of 0.9227. Lower values indicate better semantic separation.
430
+ - **Alignment Quality:** Aligned models achieve up to 0.8% R@1 in cross-lingual retrieval.
431
  - **Recommendation:** 128d aligned for best cross-lingual performance
432
 
433
  ---
434
  ## 6. Morphological Analysis (Experimental)
435
 
 
 
436
  This section presents an automated morphological analysis derived from the statistical divergence between word-level and subword-level models. By analyzing where subword predictability spikes and where word-level coverage fails, we can infer linguistic structures without supervised data.
437
 
438
  ### 6.1 Productivity & Complexity
439
 
440
  | Metric | Value | Interpretation | Recommendation |
441
  |--------|-------|----------------|----------------|
442
+ | Productivity Index | **2.767** | High morphological productivity | Reliable analysis |
443
+ | Idiomaticity Gap | **1.669** | High formulaic/idiomatic content | - |
444
 
445
  ### 6.2 Affix Inventory (Productive Units)
446
 
 
449
  #### Productive Prefixes
450
  | Prefix | Examples |
451
  |--------|----------|
452
+ | `-pe` | petaling, pekakasan, pesat |
453
+ | `-se` | sepanjang, serupo, seri |
454
+ | `-ke` | keempat, kempen, kemudahan |
455
+ | `-te` | tena, teposok, terbaik |
456
+ | `-me` | melodi, mencakup, menduo |
457
+ | `-be` | bege, begiang, been |
458
 
459
  #### Productive Suffixes
460
  | Suffix | Examples |
461
  |--------|----------|
462
+ | `-n` | sangkan, intan, kempen |
463
+ | `-an` | sangkan, intan, pekakasan |
464
+ | `-ng` | suang, sepanjang, petaling |
465
+ | `-ang` | suang, sepanjang, begiang |
466
+ | `-ah` | tanah, buah, majalah |
467
 
468
  ### 6.3 Bound Stems (Lexical Roots)
469
 
 
478
 
479
  | Prefix | Suffix | Frequency | Examples |
480
  |--------|--------|-----------|----------|
481
+ | `-pe` | `-n` | 55 words | pekakasan, pernikahan |
482
+ | `-pe` | `-an` | 49 words | pekakasan, pernikahan |
483
+ | `-ke` | `-n` | 42 words | kempen, kemudahan |
484
+ | `-ke` | `-an` | 36 words | kemudahan, keadilan |
485
+ | `-se` | `-n` | 11 words | semimon, sembilan |
486
+ | `-se` | `-ng` | 9 words | sepanjang, sesambung |
487
+ | `-te` | `-n` | 9 words | temban, teniman |
488
+ | `-me` | `-n` | 9 words | mekitoon, mesakan |
489
+ | `-pe` | `-ng` | 7 words | petaling, perang |
490
+ | `-se` | `-an` | 7 words | sembilan, sebahagian |
491
 
492
  ### 6.5 Recursive Morpheme Segmentation
493
 
 
496
  | Word | Suggested Split | Confidence | Stem |
497
  |------|-----------------|------------|------|
498
  | kebenyakan | **`ke-be-nyak-an`** | 7.5 | `nyak` |
499
+ | kebudayaan | **`ke-budaya-an`** | 6.0 | `budaya` |
 
 
500
  | keramaian | **`ke-ramai-an`** | 6.0 | `ramai` |
501
  | kepulauan | **`ke-pulau-an`** | 6.0 | `pulau` |
 
502
  | keputeraan | **`ke-putera-an`** | 6.0 | `putera` |
503
+ | kelahiran | **`ke-lahir-an`** | 6.0 | `lahir` |
504
+ | kelebihan | **`ke-lebih-an`** | 6.0 | `lebih` |
505
+ | kebangsaan | **`ke-bangsa-an`** | 6.0 | `bangsa` |
506
  | pernikahan | **`pe-rnik-ah-an`** | 4.5 | `rnik` |
507
+ | pertandingan | **`pe-rtandi-ng-an`** | 4.5 | `rtandi` |
508
+ | pelancongan | **`pe-lanco-ng-an`** | 4.5 | `lanco` |
509
+ | persembahan | **`pe-rsemb-ah-an`** | 4.5 | `rsemb` |
510
  | kemerdekaan | **`ke-me-rdeka-an`** | 4.5 | `rdeka` |
511
+ | keselamatan | **`ke-se-lamat-an`** | 4.5 | `lamat` |
512
+ | sedembila | **`se-dembila`** | 4.5 | `dembila` |
513
 
514
  ### 6.6 Linguistic Interpretation
515
 
516
  > **Automated Insight:**
517
+ The language West Coast Bajau shows high morphological productivity. The subword models are significantly more efficient than word models, suggesting a rich system of affixation or compounding.
518
+
519
+ > **Note on Idiomaticity:** The high Idiomaticity Gap suggests a large number of frequent multi-word expressions or formulaic sequences that are statistically distinct from their component parts.
520
 
521
  ---
522
  ## 7. Summary & Recommendations
 
527
 
528
  | Component | Recommended | Rationale |
529
  |-----------|-------------|-----------|
530
+ | Tokenizer | **8k BPE** | Best compression (4.80x) |
531
+ | N-gram | **5-gram** | Lowest perplexity (110) |
532
  | Markov | **Context-4** | Highest predictability (99.0%) |
533
  | Embeddings | **100d** | Balanced semantic capture and isotropy |
534
 
 
743
  ---
744
  *Generated by Wikilangs Models Pipeline*
745
 
746
+ *Report Date: 2026-01-03 18:34:47*
models/embeddings/aligned/bdr_128d.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e41667680a8bc0cadec67edf4743451f35cd7995f62085ad35d96463f5772de
3
+ size 1024832531
models/embeddings/aligned/bdr_128d.meta.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lang": "bdr", "dim": 128, "max_seq_len": 512, "is_aligned": true}
models/embeddings/aligned/bdr_128d.projection.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a12fadca82ce6dfea45d130790fcac9cd31a5b472f332630040e1cfe4a2787f
3
+ size 65664
models/embeddings/aligned/bdr_128d_metadata.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "language": "bdr",
3
+ "dimension": 128,
4
+ "version": "aligned",
5
+ "hub_language": "en",
6
+ "seed_vocab_size": 255,
7
+ "vocab_size": 801
8
+ }
models/embeddings/aligned/bdr_32d.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ae1bb7c338b11bc2a28dfc0e13abbbd2a39098370cf28ad78b0514a6f5863aa
3
+ size 256217363
models/embeddings/aligned/bdr_32d.meta.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lang": "bdr", "dim": 32, "max_seq_len": 512, "is_aligned": true}
models/embeddings/aligned/bdr_32d.projection.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aad3c75607eb6c5b251c400c7ff7627a3c005d45d6cb24c0abcd80758e3be665
3
+ size 4224
models/embeddings/aligned/bdr_32d_metadata.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "language": "bdr",
3
+ "dimension": 32,
4
+ "version": "aligned",
5
+ "hub_language": "en",
6
+ "seed_vocab_size": 255,
7
+ "vocab_size": 801
8
+ }
models/embeddings/aligned/bdr_64d.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37213f3cd2a0a803f1b3c410bc6e46b64095e2d36a2dc927c279189c10fdde99
3
+ size 512422419
models/embeddings/aligned/bdr_64d.meta.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lang": "bdr", "dim": 64, "max_seq_len": 512, "is_aligned": true}
models/embeddings/aligned/bdr_64d.projection.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ff9f1c468e4c40cecffd704552c1ed2039f271a861bc14cd8ba44252ed0c8dd
3
+ size 16512
models/embeddings/aligned/bdr_64d_metadata.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "language": "bdr",
3
+ "dimension": 64,
4
+ "version": "aligned",
5
+ "hub_language": "en",
6
+ "seed_vocab_size": 255,
7
+ "vocab_size": 801
8
+ }
models/embeddings/monolingual/bdr_128d.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:14901cc49f6d41fceff30e160e6f74a2d3e4c31128383c0eb0c581df8f55667d
3
- size 1024837728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e41667680a8bc0cadec67edf4743451f35cd7995f62085ad35d96463f5772de
3
+ size 1024832531
models/embeddings/monolingual/bdr_128d_metadata.json CHANGED
@@ -11,5 +11,5 @@
11
  "encoding_method": "rope",
12
  "dim": 128
13
  },
14
- "vocab_size": 806
15
  }
 
11
  "encoding_method": "rope",
12
  "dim": 128
13
  },
14
+ "vocab_size": 801
15
  }
models/embeddings/monolingual/bdr_32d.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fa045fafcb5ac078ce6dc442fa1420ea920b282300bd6a2013469a40763dfb5e
3
- size 256218720
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ae1bb7c338b11bc2a28dfc0e13abbbd2a39098370cf28ad78b0514a6f5863aa
3
+ size 256217363
models/embeddings/monolingual/bdr_32d_metadata.json CHANGED
@@ -11,5 +11,5 @@
11
  "encoding_method": "rope",
12
  "dim": 32
13
  },
14
- "vocab_size": 806
15
  }
 
11
  "encoding_method": "rope",
12
  "dim": 32
13
  },
14
+ "vocab_size": 801
15
  }
models/embeddings/monolingual/bdr_64d.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0f57494e6d97c810c174b928ec3c054ba096b896edd93f5bb2951c51ea1469fb
3
- size 512425056
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37213f3cd2a0a803f1b3c410bc6e46b64095e2d36a2dc927c279189c10fdde99
3
+ size 512422419
models/embeddings/monolingual/bdr_64d_metadata.json CHANGED
@@ -11,5 +11,5 @@
11
  "encoding_method": "rope",
12
  "dim": 64
13
  },
14
- "vocab_size": 806
15
  }
 
11
  "encoding_method": "rope",
12
  "dim": 64
13
  },
14
+ "vocab_size": 801
15
  }
models/subword_markov/bdr_markov_ctx1_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:90dd99baf68b20403d569b8e5f8d37366e4ddf2960b74590e45c40e1931d3cf1
3
- size 12465
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c01103c97f395e9b1a6df414c90c77b0441b957516d54a843bb55d5ae0454da
3
+ size 12256
models/subword_markov/bdr_markov_ctx1_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 1,
3
  "variant": "subword",
4
  "language": "bdr",
5
- "unique_contexts": 104,
6
- "total_transitions": 158102
7
  }
 
2
  "context_size": 1,
3
  "variant": "subword",
4
  "language": "bdr",
5
+ "unique_contexts": 101,
6
+ "total_transitions": 157193
7
  }
models/subword_markov/bdr_markov_ctx2_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:de0458067f06b19cd92069895f072cb47f6cdac7b1aec47edc3f54439b9dc8fb
3
- size 51058
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a1eaf53753bc5cd77e26c762c70b525953aea8bbbdaba88d81f8b89f3cad7d8
3
+ size 50308
models/subword_markov/bdr_markov_ctx2_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 2,
3
  "variant": "subword",
4
  "language": "bdr",
5
- "unique_contexts": 1154,
6
- "total_transitions": 157564
7
  }
 
2
  "context_size": 2,
3
  "variant": "subword",
4
  "language": "bdr",
5
+ "unique_contexts": 1130,
6
+ "total_transitions": 156655
7
  }
models/subword_markov/bdr_markov_ctx3_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8a603938c76585d0058fba20159b63f48cc8de1a081d60c9e812d5ef4cc856c2
3
- size 147977
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:125c046ee06d3fd97f3d071738c1eee6377f5c7c4013bd62eef4fb484d2c9ccc
3
+ size 146590
models/subword_markov/bdr_markov_ctx3_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 3,
3
  "variant": "subword",
4
  "language": "bdr",
5
- "unique_contexts": 6603,
6
- "total_transitions": 157026
7
  }
 
2
  "context_size": 3,
3
  "variant": "subword",
4
  "language": "bdr",
5
+ "unique_contexts": 6542,
6
+ "total_transitions": 156117
7
  }
models/subword_markov/bdr_markov_ctx4_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:40b65aa668a1cff884dd7d1981f09619017109f7d40edc49b7916cb9d523100b
3
- size 355053
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1cfe96d25d60cd6b2f2a9c4d6224d5dc2d7e34f5199adc174e0e948de40a415
3
+ size 353233
models/subword_markov/bdr_markov_ctx4_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 4,
3
  "variant": "subword",
4
  "language": "bdr",
5
- "unique_contexts": 20699,
6
- "total_transitions": 156488
7
  }
 
2
  "context_size": 4,
3
  "variant": "subword",
4
  "language": "bdr",
5
+ "unique_contexts": 20556,
6
+ "total_transitions": 155579
7
  }
models/subword_ngram/bdr_2gram_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f89d4cd1bab1bda41e8fb6e1bcbdcfb402307503bc759166623f23f6684cb7d8
3
- size 9761
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:81ad86f878dbd31142b31c0bceefa29fec8ddff13564e7a1537d9e13f941884e
3
+ size 9737
models/subword_ngram/bdr_2gram_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "n": 2,
3
  "variant": "subword",
4
  "language": "bdr",
5
- "unique_ngrams": 597,
6
- "total_ngrams": 158102
7
  }
 
2
  "n": 2,
3
  "variant": "subword",
4
  "language": "bdr",
5
+ "unique_ngrams": 593,
6
+ "total_ngrams": 157193
7
  }
models/subword_ngram/bdr_3gram_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:82c353ab6740086eface298a08842c1494816358218a9e1924c16d5cc9298c3d
3
- size 37547
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5536c1c7357dd2f396f0e7af3a16bbf82d7e0abefe4045ab9afe2432147f21ae
3
+ size 37388
models/subword_ngram/bdr_3gram_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "n": 3,
3
  "variant": "subword",
4
  "language": "bdr",
5
- "unique_ngrams": 3421,
6
- "total_ngrams": 157564
7
  }
 
2
  "n": 3,
3
  "variant": "subword",
4
  "language": "bdr",
5
+ "unique_ngrams": 3407,
6
+ "total_ngrams": 156655
7
  }
models/subword_ngram/bdr_4gram_subword.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:00c9461926041d02b2401848dda26b82b5b87974371b657a4eeb9a268d495f8c
3
- size 128743
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:805cf63298b535adbcf60a5db766aade000901bf269c45ce290338156be8f197
3
+ size 129044
models/subword_ngram/bdr_4gram_subword_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "n": 4,
3
  "variant": "subword",
4
  "language": "bdr",
5
- "unique_ngrams": 11413,
6
- "total_ngrams": 157026
7
  }
 
2
  "n": 4,
3
  "variant": "subword",
4
  "language": "bdr",
5
+ "unique_ngrams": 11348,
6
+ "total_ngrams": 156117
7
  }
models/subword_ngram/bdr_5gram_subword.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f4e8e2a8a2fd55877f905703b6d37187254f576327870646a2a80da1780f039
3
+ size 224851
models/subword_ngram/bdr_5gram_subword_metadata.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "n": 5,
3
+ "variant": "subword",
4
+ "language": "bdr",
5
+ "unique_ngrams": 18466,
6
+ "total_ngrams": 155579
7
+ }
models/tokenizer/bdr_tokenizer_8k.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:60470ba72369b5d6feb2a6ddd8122b4eb757065597901f8d33f1f47db9d61152
3
- size 373260
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3257481d465f4f6cee30948577989924f9bf988a0865d4b1d66ac5beb8c0a6c
3
+ size 373305
models/tokenizer/bdr_tokenizer_8k.vocab CHANGED
The diff for this file is too large to render. See raw diff
 
models/vocabulary/bdr_vocabulary.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a58092d4262f6aa43f6194c3af31f93b2c60e2e1f1457610992d6848f4156246
3
- size 38048
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b427705d359f9ebf3fa546a3c56857b1b319cd6cc0cfdb8bec8e9e95e7f20b9
3
+ size 37799
models/vocabulary/bdr_vocabulary_metadata.json CHANGED
@@ -1,16 +1,16 @@
1
  {
2
  "language": "bdr",
3
- "vocabulary_size": 2342,
4
  "variant": "full",
5
  "statistics": {
6
- "type_token_ratio": 0.20221606648199447,
7
  "coverage": {
8
- "top_100": 0.4034075816795052,
9
- "top_1000": 0.7593442871779305,
10
- "top_5000": 0.9875156528668463
11
  },
12
- "hapax_count": 2987,
13
- "hapax_ratio": 0.5605179208106587,
14
  "total_documents": 538
15
  }
16
  }
 
1
  {
2
  "language": "bdr",
3
+ "vocabulary_size": 2333,
4
  "variant": "full",
5
  "statistics": {
6
+ "type_token_ratio": 0.2020302243932224,
7
  "coverage": {
8
+ "top_100": 0.4047855289268814,
9
+ "top_1000": 0.760418256754694,
10
+ "top_5000": 0.9887803388795604
11
  },
12
+ "hapax_count": 2961,
13
+ "hapax_ratio": 0.5593124291650926,
14
  "total_documents": 538
15
  }
16
  }
models/word_markov/bdr_markov_ctx1_word.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:50da15bb2d596e55b1af74da418be234ae20d7994f359665a73457c179ace2c4
3
- size 149999
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93344d32ca7f52ac3cf06047cffe71756e8f7a067baa2b2d0e581bff40c4ad9b
3
+ size 149528
models/word_markov/bdr_markov_ctx1_word_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 1,
3
  "variant": "word",
4
  "language": "bdr",
5
- "unique_contexts": 5241,
6
- "total_transitions": 25815
7
  }
 
2
  "context_size": 1,
3
  "variant": "word",
4
  "language": "bdr",
5
+ "unique_contexts": 5206,
6
+ "total_transitions": 25666
7
  }
models/word_markov/bdr_markov_ctx2_word.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a70c0745d64f93f951581ccf1e770bc0198416279a90b1c76af931bc397bb65f
3
- size 308032
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2ff731f7444897501c9da118612008f21992d71fa9e8d4c5c82bd2db7ebb4c6
3
+ size 306258
models/word_markov/bdr_markov_ctx2_word_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 2,
3
  "variant": "word",
4
  "language": "bdr",
5
- "unique_contexts": 18592,
6
- "total_transitions": 25277
7
  }
 
2
  "context_size": 2,
3
  "variant": "word",
4
  "language": "bdr",
5
+ "unique_contexts": 18482,
6
+ "total_transitions": 25128
7
  }
models/word_markov/bdr_markov_ctx3_word.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:223447078bb309ba5d758145f41c4571806689ffaf1efe8aef015c68c49c7526
3
- size 388681
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b86db7e1621c9ba0690146e1820f26862e9beaaa79b8cfbf63842988dbb2e4e7
3
+ size 386401
models/word_markov/bdr_markov_ctx3_word_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 3,
3
  "variant": "word",
4
  "language": "bdr",
5
- "unique_contexts": 22996,
6
- "total_transitions": 24739
7
  }
 
2
  "context_size": 3,
3
  "variant": "word",
4
  "language": "bdr",
5
+ "unique_contexts": 22853,
6
+ "total_transitions": 24590
7
  }
models/word_markov/bdr_markov_ctx4_word.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:66c6e307074a65e79883d656ec156325dab1be18c7c2082eab0c9665ad6e3133
3
- size 429090
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dff2b6fb681e66f38ede9f1c2614a2c163ec702f2199b843e0e84e335a3600ce
3
+ size 426318
models/word_markov/bdr_markov_ctx4_word_metadata.json CHANGED
@@ -2,6 +2,6 @@
2
  "context_size": 4,
3
  "variant": "word",
4
  "language": "bdr",
5
- "unique_contexts": 23590,
6
- "total_transitions": 24201
7
  }
 
2
  "context_size": 4,
3
  "variant": "word",
4
  "language": "bdr",
5
+ "unique_contexts": 23441,
6
+ "total_transitions": 24052
7
  }
models/word_ngram/bdr_2gram_word.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3f7dbbf4075f6494f4e948bc0417ec1b327b33cb3676c74c44c5cf7b15c76e93
3
- size 7711
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc048e89e93b12320521aa74196bfb21637e0a035925b52b7ef01bae1000a530
3
+ size 7714
models/word_ngram/bdr_2gram_word_metadata.json CHANGED
@@ -3,5 +3,5 @@
3
  "variant": "word",
4
  "language": "bdr",
5
  "unique_ngrams": 401,
6
- "total_ngrams": 25815
7
  }
 
3
  "variant": "word",
4
  "language": "bdr",
5
  "unique_ngrams": 401,
6
+ "total_ngrams": 25666
7
  }