samleeasus commited on
Commit
f6cf8fa
·
1 Parent(s): 79d3212

Upload tokenizer

Browse files
special_tokens_map.json CHANGED
@@ -13,13 +13,7 @@
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
- "pad_token": {
17
- "content": "<|pad|>",
18
- "lstrip": false,
19
- "normalized": true,
20
- "rstrip": false,
21
- "single_word": false
22
- },
23
  "unk_token": {
24
  "content": "<unk>",
25
  "lstrip": false,
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
+ "pad_token": "<|pad|>",
 
 
 
 
 
 
17
  "unk_token": {
18
  "content": "<unk>",
19
  "lstrip": false,
tokenizer.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7374e92f490738d7fd473f6062aab4f6a14c998104037ef393eafdef571bccd0
3
- size 827700
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb3607f62751f5595f4f135617e930cf34a11213803db34898a7c9c942ec18c0
3
+ size 827711
tokenizer_config.json CHANGED
@@ -19,14 +19,7 @@
19
  "single_word": false
20
  },
21
  "model_max_length": 1000000000000000019884624838656,
22
- "pad_token": {
23
- "__type": "AddedToken",
24
- "content": "<|pad|>",
25
- "lstrip": false,
26
- "normalized": true,
27
- "rstrip": false,
28
- "single_word": false
29
- },
30
  "sp_model_kwargs": {},
31
  "tokenizer_class": "LlamaTokenizer",
32
  "unk_token": {
 
19
  "single_word": false
20
  },
21
  "model_max_length": 1000000000000000019884624838656,
22
+ "pad_token": null,
 
 
 
 
 
 
 
23
  "sp_model_kwargs": {},
24
  "tokenizer_class": "LlamaTokenizer",
25
  "unk_token": {