|
|
|
|
|
|
|
|
from PIL import Image |
|
|
from transformers import TrOCRProcessor |
|
|
from optimum.onnxruntime import ORTModelForVision2Seq |
|
|
from transformers import VisionEncoderDecoderModel |
|
|
|
|
|
def test_tokenizer_consistency(processor, test_strings=None): |
|
|
""" |
|
|
测试Tokenizer的编码和解码是否一致 |
|
|
|
|
|
Args: |
|
|
processor: TrOCRProcessor实例 |
|
|
test_strings (list): 要测试的字符串列表 |
|
|
""" |
|
|
if test_strings is None: |
|
|
test_strings = [ |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
"\\hat { N } _ { 3 } = \\sum \\sp f _ { j = 1 } a _ { j } \\sp { \\dagger } a _ { j } .", |
|
|
] |
|
|
|
|
|
print("\n" + "="*50) |
|
|
print("Testing Tokenizer Consistency") |
|
|
print("="*50) |
|
|
|
|
|
all_passed = True |
|
|
for text in test_strings: |
|
|
|
|
|
encoded = processor.tokenizer.encode_plus(text, return_tensors="pt") |
|
|
outs = processor.tokenizer( |
|
|
[text], |
|
|
padding="max_length", |
|
|
truncation=True, |
|
|
max_length=512, |
|
|
)["input_ids"] |
|
|
input_ids = encoded["input_ids"][0] |
|
|
breakpoint() |
|
|
|
|
|
|
|
|
decoded = processor.tokenizer.decode(input_ids, skip_special_tokens=True) |
|
|
|
|
|
|
|
|
is_match = (text == decoded) |
|
|
if not is_match: |
|
|
all_passed = False |
|
|
|
|
|
print(f"\nOriginal: {repr(text)}") |
|
|
print(f"Encoded: {input_ids.tolist()}") |
|
|
print(f"Decoded: {repr(decoded)}") |
|
|
print(f"Match: {is_match}") |
|
|
|
|
|
print("\n" + "="*50) |
|
|
if all_passed: |
|
|
print("✅ All tests passed! Tokenizer encoding and decoding are consistent.") |
|
|
else: |
|
|
print("❌ Some tests failed. Tokenizer encoding and decoding are not consistent.") |
|
|
print("="*50 + "\n") |
|
|
|
|
|
model = 'breezedeus/pix2text-mfr' |
|
|
|
|
|
processor = TrOCRProcessor.from_pretrained(model) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
model = 'models/checkpoint-683356' |
|
|
model = VisionEncoderDecoderModel.from_pretrained(model) |
|
|
|
|
|
image_fps = [ |
|
|
|
|
|
'docs/examples/formula.jpg', |
|
|
|
|
|
|
|
|
] |
|
|
images = [Image.open(fp).convert('RGB') for fp in image_fps] |
|
|
pixel_values = processor(images=images, return_tensors="pt").pixel_values |
|
|
generated_ids = model.generate(pixel_values) |
|
|
generated_text = processor.batch_decode(generated_ids, skip_special_tokens=True) |
|
|
print(f'generated_ids: {generated_ids}, \ngenerated text: {generated_text}') |
|
|
|