fix-tokenizer
#35
by
xinsu - opened
- processing_phi3_v.py +1 -1
processing_phi3_v.py
CHANGED
|
@@ -150,7 +150,7 @@ class Phi3VProcessor(ProcessorMixin):
|
|
| 150 |
return BatchFeature(data={**model_inputs})
|
| 151 |
|
| 152 |
pattern = r"<\|image_\d+\|>"
|
| 153 |
-
prompt_chunks = [self.tokenizer(chunk).input_ids for chunk in re.split(pattern, texts)]
|
| 154 |
|
| 155 |
if 'num_img_tokens' in images:
|
| 156 |
num_img_tokens = images['num_img_tokens']
|
|
|
|
| 150 |
return BatchFeature(data={**model_inputs})
|
| 151 |
|
| 152 |
pattern = r"<\|image_\d+\|>"
|
| 153 |
+
prompt_chunks = [self.tokenizer(chunk, return_tensors=return_tensors, padding=padding, truncation=truncation, max_length=max_length).input_ids for chunk in re.split(pattern, texts)]
|
| 154 |
|
| 155 |
if 'num_img_tokens' in images:
|
| 156 |
num_img_tokens = images['num_img_tokens']
|