mirajnair commited on
Commit
396c102
·
verified ·
1 Parent(s): 1f58f52

Upload tokenizer

Browse files
Files changed (3) hide show
  1. special_tokens_map.json +1 -1
  2. tokenizer.json +1 -9
  3. tokenizer_config.json +24 -24
special_tokens_map.json CHANGED
@@ -1 +1 @@
1
- {}
 
1
+ {}
tokenizer.json CHANGED
@@ -42,15 +42,7 @@
42
  "<|pad|>": 1,
43
  "+": 2,
44
  "0": 3,
45
- "1": 4,
46
- "2": 5,
47
- "3": 6,
48
- "4": 7,
49
- "5": 8,
50
- "6": 9,
51
- "7": 10,
52
- "8": 11,
53
- "9": 12
54
  },
55
  "merges": []
56
  }
 
42
  "<|pad|>": 1,
43
  "+": 2,
44
  "0": 3,
45
+ "1": 4
 
 
 
 
 
 
 
 
46
  },
47
  "merges": []
48
  }
tokenizer_config.json CHANGED
@@ -1,24 +1,24 @@
1
- {
2
- "added_tokens_decoder": {
3
- "0": {
4
- "content": "<|UNK|>",
5
- "lstrip": false,
6
- "normalized": false,
7
- "rstrip": false,
8
- "single_word": false,
9
- "special": true
10
- },
11
- "1": {
12
- "content": "<|pad|>",
13
- "lstrip": false,
14
- "normalized": false,
15
- "rstrip": false,
16
- "single_word": false,
17
- "special": true
18
- }
19
- },
20
- "clean_up_tokenization_spaces": false,
21
- "extra_special_tokens": {},
22
- "model_max_length": 1000000000000000019884624838656,
23
- "tokenizer_class": "PreTrainedTokenizer"
24
- }
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<|UNK|>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<|pad|>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ }
19
+ },
20
+ "clean_up_tokenization_spaces": false,
21
+ "extra_special_tokens": {},
22
+ "model_max_length": 1000000000000000019884624838656,
23
+ "tokenizer_class": "PreTrainedTokenizer"
24
+ }