Clara Isabel Meister commited on
Commit
9037773
·
0 Parent(s):

adding tokenizers

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +60 -0
  2. byte_128k/tokenizers/bpe.json +3 -0
  3. byte_128k/tokenizers/langspec_soft_ar_em_probs.json +0 -0
  4. byte_128k/tokenizers/langspec_soft_cy_em_probs.json +0 -0
  5. byte_128k/tokenizers/langspec_soft_de_em_probs.json +0 -0
  6. byte_128k/tokenizers/langspec_soft_el_em_probs.json +0 -0
  7. byte_128k/tokenizers/langspec_soft_en_em_probs.json +0 -0
  8. byte_128k/tokenizers/langspec_soft_es_em_probs.json +0 -0
  9. byte_128k/tokenizers/langspec_soft_ga_em_probs.json +0 -0
  10. byte_128k/tokenizers/langspec_soft_gn_em_probs.json +0 -0
  11. byte_128k/tokenizers/langspec_soft_hi_em_probs.json +0 -0
  12. byte_128k/tokenizers/langspec_soft_id_em_probs.json +0 -0
  13. byte_128k/tokenizers/langspec_soft_ku_em_probs.json +0 -0
  14. byte_128k/tokenizers/langspec_soft_mg_em_probs.json +0 -0
  15. byte_128k/tokenizers/langspec_soft_pl_em_probs.json +0 -0
  16. byte_128k/tokenizers/langspec_soft_ru_em_probs.json +0 -0
  17. byte_128k/tokenizers/langspec_soft_sw_em_probs.json +0 -0
  18. byte_128k/tokenizers/langspec_soft_tr_em_probs.json +0 -0
  19. byte_128k/tokenizers/langspec_soft_vi_em_probs.json +0 -0
  20. byte_128k/tokenizers/langspec_soft_yo_em_probs.json +0 -0
  21. byte_128k/tokenizers/langspec_soft_zh_em_probs.json +0 -0
  22. byte_128k/tokenizers/multigramlm_soft.json +0 -0
  23. byte_128k/tokenizers/multigramlm_soft_em_0.pkl +3 -0
  24. byte_128k/tokenizers/multigramlm_soft_em_1.pkl +3 -0
  25. byte_128k/tokenizers/multigramlm_soft_em_2.pkl +3 -0
  26. byte_128k/tokenizers/standard_unigramlm_hard.json +0 -0
  27. byte_128k/tokenizers/standard_unigramlm_hard_em_0.pkl +3 -0
  28. byte_128k/tokenizers/standard_unigramlm_hard_em_1.pkl +3 -0
  29. byte_128k/tokenizers/standard_unigramlm_hard_em_2.pkl +3 -0
  30. byte_128k/tokenizers/standard_unigramlm_soft.json +0 -0
  31. byte_128k/tokenizers/standard_unigramlm_soft_em_0.pkl +3 -0
  32. byte_128k/tokenizers/standard_unigramlm_soft_em_1.pkl +3 -0
  33. byte_128k/tokenizers/standard_unigramlm_soft_em_2.pkl +3 -0
  34. byte_128k/tokenizers/tokmix.json +0 -0
  35. byte_128k/tokenizers/tokmix_ar_tokenizer.json +0 -0
  36. byte_128k/tokenizers/tokmix_ar_tokenizer_em_0.pkl +3 -0
  37. byte_128k/tokenizers/tokmix_ar_tokenizer_em_1.pkl +3 -0
  38. byte_128k/tokenizers/tokmix_ar_tokenizer_em_2.pkl +3 -0
  39. byte_128k/tokenizers/tokmix_cy_tokenizer.json +0 -0
  40. byte_128k/tokenizers/tokmix_de_tokenizer.json +0 -0
  41. byte_128k/tokenizers/tokmix_de_tokenizer_em_0.pkl +3 -0
  42. byte_128k/tokenizers/tokmix_de_tokenizer_em_1.pkl +3 -0
  43. byte_128k/tokenizers/tokmix_de_tokenizer_em_2.pkl +3 -0
  44. byte_128k/tokenizers/tokmix_el_tokenizer.json +0 -0
  45. byte_128k/tokenizers/tokmix_en_tokenizer.json +0 -0
  46. byte_128k/tokenizers/tokmix_es_tokenizer.json +0 -0
  47. byte_128k/tokenizers/tokmix_ga_tokenizer.json +0 -0
  48. byte_128k/tokenizers/tokmix_gn_tokenizer.json +0 -0
  49. byte_128k/tokenizers/tokmix_hi_tokenizer.json +0 -0
  50. byte_128k/tokenizers/tokmix_id_tokenizer.json +0 -0
.gitattributes ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.lz4 filter=lfs diff=lfs merge=lfs -text
12
+ *.mds filter=lfs diff=lfs merge=lfs -text
13
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
14
+ *.model filter=lfs diff=lfs merge=lfs -text
15
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
16
+ *.npy filter=lfs diff=lfs merge=lfs -text
17
+ *.npz filter=lfs diff=lfs merge=lfs -text
18
+ *.onnx filter=lfs diff=lfs merge=lfs -text
19
+ *.ot filter=lfs diff=lfs merge=lfs -text
20
+ *.parquet filter=lfs diff=lfs merge=lfs -text
21
+ *.pb filter=lfs diff=lfs merge=lfs -text
22
+ *.pickle filter=lfs diff=lfs merge=lfs -text
23
+ *.pkl filter=lfs diff=lfs merge=lfs -text
24
+ *.pt filter=lfs diff=lfs merge=lfs -text
25
+ *.pth filter=lfs diff=lfs merge=lfs -text
26
+ *.rar filter=lfs diff=lfs merge=lfs -text
27
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
28
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
29
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
30
+ *.tar filter=lfs diff=lfs merge=lfs -text
31
+ *.tflite filter=lfs diff=lfs merge=lfs -text
32
+ *.tgz filter=lfs diff=lfs merge=lfs -text
33
+ *.wasm filter=lfs diff=lfs merge=lfs -text
34
+ *.xz filter=lfs diff=lfs merge=lfs -text
35
+ *.zip filter=lfs diff=lfs merge=lfs -text
36
+ *.zst filter=lfs diff=lfs merge=lfs -text
37
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
38
+ # Audio files - uncompressed
39
+ *.pcm filter=lfs diff=lfs merge=lfs -text
40
+ *.sam filter=lfs diff=lfs merge=lfs -text
41
+ *.raw filter=lfs diff=lfs merge=lfs -text
42
+ # Audio files - compressed
43
+ *.aac filter=lfs diff=lfs merge=lfs -text
44
+ *.flac filter=lfs diff=lfs merge=lfs -text
45
+ *.mp3 filter=lfs diff=lfs merge=lfs -text
46
+ *.ogg filter=lfs diff=lfs merge=lfs -text
47
+ *.wav filter=lfs diff=lfs merge=lfs -text
48
+ # Image files - uncompressed
49
+ *.bmp filter=lfs diff=lfs merge=lfs -text
50
+ *.gif filter=lfs diff=lfs merge=lfs -text
51
+ *.png filter=lfs diff=lfs merge=lfs -text
52
+ *.tiff filter=lfs diff=lfs merge=lfs -text
53
+ # Image files - compressed
54
+ *.jpg filter=lfs diff=lfs merge=lfs -text
55
+ *.jpeg filter=lfs diff=lfs merge=lfs -text
56
+ *.webp filter=lfs diff=lfs merge=lfs -text
57
+ # Video files - compressed
58
+ *.mp4 filter=lfs diff=lfs merge=lfs -text
59
+ *.webm filter=lfs diff=lfs merge=lfs -text
60
+ byte_128k/tokenizers/bpe.json filter=lfs diff=lfs merge=lfs -text
byte_128k/tokenizers/bpe.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae62cf50a1a5396d42477556c2bf7445da404627c8ec08bcef86a2ac88e7bdb9
3
+ size 10703232
byte_128k/tokenizers/langspec_soft_ar_em_probs.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/langspec_soft_cy_em_probs.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/langspec_soft_de_em_probs.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/langspec_soft_el_em_probs.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/langspec_soft_en_em_probs.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/langspec_soft_es_em_probs.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/langspec_soft_ga_em_probs.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/langspec_soft_gn_em_probs.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/langspec_soft_hi_em_probs.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/langspec_soft_id_em_probs.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/langspec_soft_ku_em_probs.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/langspec_soft_mg_em_probs.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/langspec_soft_pl_em_probs.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/langspec_soft_ru_em_probs.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/langspec_soft_sw_em_probs.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/langspec_soft_tr_em_probs.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/langspec_soft_vi_em_probs.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/langspec_soft_yo_em_probs.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/langspec_soft_zh_em_probs.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/multigramlm_soft.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/multigramlm_soft_em_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6624167d4567d297b0f01c6afb1d512706f100fbf392c806a3523e1f4dd4a9b4
3
+ size 25929173
byte_128k/tokenizers/multigramlm_soft_em_1.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e562fbb82bf81799241d3bafdba17d8e880c962681b02d82bb7d6b28d36beb8
3
+ size 19871977
byte_128k/tokenizers/multigramlm_soft_em_2.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f34b3fc75b8b105c257f39d7847d41919663788b6458ab0301f43de81f70269
3
+ size 14046099
byte_128k/tokenizers/standard_unigramlm_hard.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/standard_unigramlm_hard_em_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b31e8b0ddd4b65333c8d76cb9ededc4b85eff2a4e06c89a26f397f7d05d579f
3
+ size 25929173
byte_128k/tokenizers/standard_unigramlm_hard_em_1.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe4e86930b9b9eb9d9218e33a9ceeef46aa9e7468b361cc221caa9a140050070
3
+ size 19818731
byte_128k/tokenizers/standard_unigramlm_hard_em_2.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:863a6fbc4b1aa069da4e6f0939f235516874e7d459a8cc7d8187e1f50787d1fa
3
+ size 14010846
byte_128k/tokenizers/standard_unigramlm_soft.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/standard_unigramlm_soft_em_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c94ea23558c1fb4eb6a6b11124829d56fa49220285ffc9f045504a6868b0dcb
3
+ size 25929173
byte_128k/tokenizers/standard_unigramlm_soft_em_1.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ff5ee745bd1b0fc5df0c22ddc078977dfad8cd359bd237520697fb041bc1f5d
3
+ size 19915464
byte_128k/tokenizers/standard_unigramlm_soft_em_2.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10122eda2325d87472270bfba6f8756bb684952d969135ccadabe5726ea7a115
3
+ size 14108820
byte_128k/tokenizers/tokmix.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/tokmix_ar_tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/tokmix_ar_tokenizer_em_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be0327a53ab859fd514152aabaf7b73b065de64a6d45208332b209670e2474d7
3
+ size 15891442
byte_128k/tokenizers/tokmix_ar_tokenizer_em_1.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ffa14e3772b0852cfbc42e43143e9f3e06203061a854ed2ff00af04b565f470e
3
+ size 13700748
byte_128k/tokenizers/tokmix_ar_tokenizer_em_2.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66f0cec8552fc44e56441a6670cbee1472895dd510d3d8da0de29ae383f4219b
3
+ size 11539496
byte_128k/tokenizers/tokmix_cy_tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/tokmix_de_tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/tokmix_de_tokenizer_em_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01baf2e3769dbb5c884b30a9af6738a326f9f3ca2352b8d6f48d3e7a05ec0c65
3
+ size 10484717
byte_128k/tokenizers/tokmix_de_tokenizer_em_1.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8fbe009c2e69045f0049cb155f91b8d059bf9e238be4fe045b036a758dbe9458
3
+ size 9693914
byte_128k/tokenizers/tokmix_de_tokenizer_em_2.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e7c52d57eefde485f11191118ee68247cd44cdffc8bc7c1068cd8f6c979b6f0
3
+ size 8901702
byte_128k/tokenizers/tokmix_el_tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/tokmix_en_tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/tokmix_es_tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/tokmix_ga_tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/tokmix_gn_tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/tokmix_hi_tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
byte_128k/tokenizers/tokmix_id_tokenizer.json ADDED
The diff for this file is too large to render. See raw diff