Update tokenizer_script.py
Browse files- tokenizer_script.py +3 -3
tokenizer_script.py
CHANGED
|
@@ -82,13 +82,13 @@ class CharacterTokenizer(PreTrainedTokenizer):
|
|
| 82 |
if max_length is not None:
|
| 83 |
encoded_texts = [ids[:max_length] for ids in encoded_texts]
|
| 84 |
if add_special_tokens:
|
| 85 |
-
bos_token_id = self.convert_tokens_to_ids(
|
| 86 |
-
eos_token_id = self.convert_tokens_to_ids(
|
| 87 |
encoded_texts = [[bos_token_id] + ids + [eos_token_id] for ids in encoded_texts]
|
| 88 |
# Handle padding
|
| 89 |
if padding:
|
| 90 |
# properly handle padding side
|
| 91 |
-
pad_id = self.vocab.get(
|
| 92 |
max_len = max(len(ids) for ids in encoded_texts) if max_length is None else max_length
|
| 93 |
if tokenizer.padding_side == "right":
|
| 94 |
encoded_texts = [ids + [pad_id] * (max_len - len(ids)) for ids in encoded_texts]
|
|
|
|
| 82 |
if max_length is not None:
|
| 83 |
encoded_texts = [ids[:max_length] for ids in encoded_texts]
|
| 84 |
if add_special_tokens:
|
| 85 |
+
bos_token_id = self.convert_tokens_to_ids(self.bos_token)
|
| 86 |
+
eos_token_id = self.convert_tokens_to_ids(self.eos_token)
|
| 87 |
encoded_texts = [[bos_token_id] + ids + [eos_token_id] for ids in encoded_texts]
|
| 88 |
# Handle padding
|
| 89 |
if padding:
|
| 90 |
# properly handle padding side
|
| 91 |
+
pad_id = self.vocab.get(self.pad_token, 0)
|
| 92 |
max_len = max(len(ids) for ids in encoded_texts) if max_length is None else max_length
|
| 93 |
if tokenizer.padding_side == "right":
|
| 94 |
encoded_texts = [ids + [pad_id] * (max_len - len(ids)) for ids in encoded_texts]
|