Upload tokenizer
Browse files- tokenizer.json +0 -0
- tokenizer_config.json +0 -4
tokenizer.json
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
tokenizer_config.json
CHANGED
|
@@ -932,13 +932,9 @@
|
|
| 932 |
"eos_token": "</s>",
|
| 933 |
"extra_ids": 100,
|
| 934 |
"legacy": true,
|
| 935 |
-
"max_length": 512,
|
| 936 |
"model_max_length": 1000000000000000019884624838656,
|
| 937 |
"pad_token": "<pad>",
|
| 938 |
"sp_model_kwargs": {},
|
| 939 |
-
"stride": 0,
|
| 940 |
"tokenizer_class": "T5Tokenizer",
|
| 941 |
-
"truncation_side": "right",
|
| 942 |
-
"truncation_strategy": "longest_first",
|
| 943 |
"unk_token": "<unk>"
|
| 944 |
}
|
|
|
|
| 932 |
"eos_token": "</s>",
|
| 933 |
"extra_ids": 100,
|
| 934 |
"legacy": true,
|
|
|
|
| 935 |
"model_max_length": 1000000000000000019884624838656,
|
| 936 |
"pad_token": "<pad>",
|
| 937 |
"sp_model_kwargs": {},
|
|
|
|
| 938 |
"tokenizer_class": "T5Tokenizer",
|
|
|
|
|
|
|
| 939 |
"unk_token": "<unk>"
|
| 940 |
}
|