alina0195 commited on
Commit
1fcb968
·
verified ·
1 Parent(s): 234e54c

Upload tokenizer with proper BPE files (vocab.json + merges.txt)

Browse files
Files changed (4) hide show
  1. merges.txt +0 -0
  2. special_tokens_map.json +1 -37
  3. tokenizer_config.json +3 -13
  4. vocab.json +0 -0
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json CHANGED
@@ -1,37 +1 @@
1
- {
2
- "cls_token": {
3
- "content": "[CLS]",
4
- "lstrip": false,
5
- "normalized": false,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "mask_token": {
10
- "content": "[MASK]",
11
- "lstrip": false,
12
- "normalized": false,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
- "pad_token": {
17
- "content": "[PAD]",
18
- "lstrip": false,
19
- "normalized": false,
20
- "rstrip": false,
21
- "single_word": false
22
- },
23
- "sep_token": {
24
- "content": "[SEP]",
25
- "lstrip": false,
26
- "normalized": false,
27
- "rstrip": false,
28
- "single_word": false
29
- },
30
- "unk_token": {
31
- "content": "[UNK]",
32
- "lstrip": false,
33
- "normalized": false,
34
- "rstrip": false,
35
- "single_word": false
36
- }
37
- }
 
1
+ {}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
tokenizer_config.json CHANGED
@@ -1,5 +1,4 @@
1
  {
2
- "add_prefix_space": false,
3
  "added_tokens_decoder": {
4
  "0": {
5
  "content": "[MASK]",
@@ -226,17 +225,8 @@
226
  "special": true
227
  }
228
  },
229
- "clean_up_tokenization_spaces": true,
230
- "cls_token": "[CLS]",
231
  "extra_special_tokens": {},
232
- "mask_token": "[MASK]",
233
- "model_input_names": [
234
- "input_ids",
235
- "attention_mask"
236
- ],
237
- "model_max_length": 128000,
238
- "pad_token": "[PAD]",
239
- "sep_token": "[SEP]",
240
- "tokenizer_class": "PreTrainedTokenizerFast",
241
- "unk_token": "[UNK]"
242
  }
 
1
  {
 
2
  "added_tokens_decoder": {
3
  "0": {
4
  "content": "[MASK]",
 
225
  "special": true
226
  }
227
  },
228
+ "clean_up_tokenization_spaces": false,
 
229
  "extra_special_tokens": {},
230
+ "model_max_length": 1000000000000000019884624838656,
231
+ "tokenizer_class": "PreTrainedTokenizerFast"
 
 
 
 
 
 
 
 
232
  }
vocab.json ADDED
The diff for this file is too large to render. See raw diff