Samuael commited on
Commit
d0961c3
·
verified ·
1 Parent(s): 5b8f014

Upload tokenizer

Browse files
sentencepiece.bpe.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a801c63cf0822cc3a880177fd5895196337d7e3813edde88c428061c263354a4
3
- size 240461
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34727325a68a6a1e8580bf1e33934313f5f232bc92448d5ff77a5a850f25dbaa
3
+ size 240462
special_tokens_map.json CHANGED
@@ -1,71 +1,12 @@
1
  {
2
  "additional_special_tokens": [
3
- "ar_AR",
4
- "cs_CZ",
5
- "de_DE",
6
- "en_XX",
7
- "es_XX",
8
- "et_EE",
9
- "fi_FI",
10
- "fr_XX",
11
- "gu_IN",
12
- "hi_IN",
13
- "it_IT",
14
- "ja_XX",
15
- "kk_KZ",
16
- "ko_KR",
17
- "lt_LT",
18
- "lv_LV",
19
- "my_MM",
20
- "ne_NP",
21
- "nl_XX",
22
- "ro_RO",
23
- "ru_RU",
24
- "si_LK",
25
- "tr_TR",
26
- "vi_VN",
27
- "zh_CN"
28
  ],
29
- "bos_token": {
30
- "content": "<s>",
31
- "lstrip": false,
32
- "normalized": false,
33
- "rstrip": false,
34
- "single_word": false
35
- },
36
- "cls_token": {
37
- "content": "<s>",
38
- "lstrip": false,
39
- "normalized": false,
40
- "rstrip": false,
41
- "single_word": false
42
- },
43
- "eos_token": {
44
- "content": "</s>",
45
- "lstrip": false,
46
- "normalized": false,
47
- "rstrip": false,
48
- "single_word": false
49
- },
50
- "pad_token": {
51
- "content": "<pad>",
52
- "lstrip": false,
53
- "normalized": false,
54
- "rstrip": false,
55
- "single_word": false
56
- },
57
- "sep_token": {
58
- "content": "</s>",
59
- "lstrip": false,
60
- "normalized": false,
61
- "rstrip": false,
62
- "single_word": false
63
- },
64
- "unk_token": {
65
- "content": "<unk>",
66
- "lstrip": false,
67
- "normalized": false,
68
- "rstrip": false,
69
- "single_word": false
70
- }
71
  }
 
1
  {
2
  "additional_special_tokens": [
3
+ "",
4
+ "ar_AR"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  ],
6
+ "bos_token": "<s>",
7
+ "cls_token": "<s>",
8
+ "eos_token": "</s>",
9
+ "pad_token": "<pad>",
10
+ "sep_token": "</s>",
11
+ "unk_token": "<unk>"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
12
  }
tokenizer_config.json CHANGED
@@ -234,31 +234,8 @@
234
  }
235
  },
236
  "additional_special_tokens": [
237
- "ar_AR",
238
- "cs_CZ",
239
- "de_DE",
240
- "en_XX",
241
- "es_XX",
242
- "et_EE",
243
- "fi_FI",
244
- "fr_XX",
245
- "gu_IN",
246
- "hi_IN",
247
- "it_IT",
248
- "ja_XX",
249
- "kk_KZ",
250
- "ko_KR",
251
- "lt_LT",
252
- "lv_LV",
253
- "my_MM",
254
- "ne_NP",
255
- "nl_XX",
256
- "ro_RO",
257
- "ru_RU",
258
- "si_LK",
259
- "tr_TR",
260
- "vi_VN",
261
- "zh_CN"
262
  ],
263
  "bos_token": "<s>",
264
  "clean_up_tokenization_spaces": true,
@@ -272,5 +249,6 @@
272
  "src_lang": "ar_AR",
273
  "tgt_lang": "cs_CZ",
274
  "tokenizer_class": "MBartTokenizer",
 
275
  "unk_token": "<unk>"
276
  }
 
234
  }
235
  },
236
  "additional_special_tokens": [
237
+ "",
238
+ "ar_AR"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
239
  ],
240
  "bos_token": "<s>",
241
  "clean_up_tokenization_spaces": true,
 
249
  "src_lang": "ar_AR",
250
  "tgt_lang": "cs_CZ",
251
  "tokenizer_class": "MBartTokenizer",
252
+ "tokenizer_file": null,
253
  "unk_token": "<unk>"
254
  }