0.4825663803100586
Browse files- tokenizer.json +0 -0
- tokenizer_config.json +4 -4
tokenizer.json
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
tokenizer_config.json
CHANGED
|
@@ -40,7 +40,7 @@
|
|
| 40 |
"single_word": false,
|
| 41 |
"special": true
|
| 42 |
},
|
| 43 |
-
"
|
| 44 |
"content": "<|user|>",
|
| 45 |
"lstrip": false,
|
| 46 |
"normalized": false,
|
|
@@ -48,7 +48,7 @@
|
|
| 48 |
"single_word": false,
|
| 49 |
"special": true
|
| 50 |
},
|
| 51 |
-
"
|
| 52 |
"content": "<|bot|>",
|
| 53 |
"lstrip": false,
|
| 54 |
"normalized": false,
|
|
@@ -56,7 +56,7 @@
|
|
| 56 |
"single_word": false,
|
| 57 |
"special": true
|
| 58 |
},
|
| 59 |
-
"
|
| 60 |
"content": "<|end|>",
|
| 61 |
"lstrip": false,
|
| 62 |
"normalized": false,
|
|
@@ -75,7 +75,7 @@
|
|
| 75 |
"clean_up_tokenization_spaces": true,
|
| 76 |
"eos_token": "</s>",
|
| 77 |
"mask_token": "<mask>",
|
| 78 |
-
"max_length":
|
| 79 |
"model_max_length": 1000000000000000019884624838656,
|
| 80 |
"pad_token": "<pad>",
|
| 81 |
"stride": 0,
|
|
|
|
| 40 |
"single_word": false,
|
| 41 |
"special": true
|
| 42 |
},
|
| 43 |
+
"2805": {
|
| 44 |
"content": "<|user|>",
|
| 45 |
"lstrip": false,
|
| 46 |
"normalized": false,
|
|
|
|
| 48 |
"single_word": false,
|
| 49 |
"special": true
|
| 50 |
},
|
| 51 |
+
"2806": {
|
| 52 |
"content": "<|bot|>",
|
| 53 |
"lstrip": false,
|
| 54 |
"normalized": false,
|
|
|
|
| 56 |
"single_word": false,
|
| 57 |
"special": true
|
| 58 |
},
|
| 59 |
+
"2807": {
|
| 60 |
"content": "<|end|>",
|
| 61 |
"lstrip": false,
|
| 62 |
"normalized": false,
|
|
|
|
| 75 |
"clean_up_tokenization_spaces": true,
|
| 76 |
"eos_token": "</s>",
|
| 77 |
"mask_token": "<mask>",
|
| 78 |
+
"max_length": 16,
|
| 79 |
"model_max_length": 1000000000000000019884624838656,
|
| 80 |
"pad_token": "<pad>",
|
| 81 |
"stride": 0,
|