| |
| |
| import os |
| import gradio as gr |
|
|
| os.environ['FLAGS_allocator_strategy'] = 'auto_growth' |
| import cv2 |
| import numpy as np |
| import json |
| import time |
| from PIL import Image |
| from tools.infer_e2e import OpenOCR, check_and_download_font, draw_ocr_box_txt |
|
|
|
|
| def initialize_ocr(model_type, drop_score): |
| return OpenOCR(mode=model_type, drop_score=drop_score) |
|
|
|
|
| |
| model_type = 'mobile' |
| drop_score = 0.4 |
| text_sys = initialize_ocr(model_type, drop_score) |
|
|
| |
| if True: |
| img = np.random.uniform(0, 255, [640, 640, 3]).astype(np.uint8) |
| for i in range(5): |
| res = text_sys(img_numpy=img) |
|
|
| font_path = './simfang.ttf' |
| font_path = check_and_download_font(font_path) |
|
|
|
|
| def main(input_image, |
| model_type_select, |
| det_input_size_textbox=960, |
| rec_drop_score=0.4, |
| mask_thresh=0.3, |
| box_thresh=0.6, |
| unclip_ratio=1.5, |
| det_score_mode='slow'): |
| global text_sys, model_type |
|
|
| |
| if model_type_select != model_type: |
| model_type = model_type_select |
| text_sys = initialize_ocr(model_type, rec_drop_score) |
|
|
| img = input_image[:, :, ::-1] |
| starttime = time.time() |
| results, time_dict, mask = text_sys( |
| img_numpy=img, |
| return_mask=True, |
| det_input_size=int(det_input_size_textbox), |
| thresh=mask_thresh, |
| box_thresh=box_thresh, |
| unclip_ratio=unclip_ratio, |
| score_mode=det_score_mode) |
| elapse = time.time() - starttime |
| save_pred = json.dumps(results[0], ensure_ascii=False) |
| image = Image.fromarray(cv2.cvtColor(img, cv2.COLOR_BGR2RGB)) |
| boxes = [res['points'] for res in results[0]] |
| txts = [res['transcription'] for res in results[0]] |
| scores = [res['score'] for res in results[0]] |
| draw_img = draw_ocr_box_txt( |
| image, |
| boxes, |
| txts, |
| scores, |
| drop_score=rec_drop_score, |
| font_path=font_path, |
| ) |
| mask = mask[0, 0, :, :] > mask_thresh |
| return save_pred, elapse, draw_img, mask.astype('uint8') * 255 |
|
|
|
|
| def get_all_file_names_including_subdirs(dir_path): |
| all_file_names = [] |
|
|
| for root, dirs, files in os.walk(dir_path): |
| for file_name in files: |
| all_file_names.append(os.path.join(root, file_name)) |
|
|
| file_names_only = [os.path.basename(file) for file in all_file_names] |
| return file_names_only |
|
|
|
|
| def list_image_paths(directory): |
| image_extensions = ('.png', '.jpg', '.jpeg', '.gif', '.bmp', '.tiff') |
|
|
| image_paths = [] |
|
|
| for root, dirs, files in os.walk(directory): |
| for file in files: |
| if file.lower().endswith(image_extensions): |
| relative_path = os.path.relpath(os.path.join(root, file), |
| directory) |
| full_path = os.path.join(directory, relative_path) |
| image_paths.append(full_path) |
| image_paths = sorted(image_paths) |
| return image_paths |
|
|
|
|
| def find_file_in_current_dir_and_subdirs(file_name): |
| for root, dirs, files in os.walk('.'): |
| if file_name in files: |
| relative_path = os.path.join(root, file_name) |
| return relative_path |
|
|
|
|
| e2e_img_example = list_image_paths('./OCR_e2e_img') |
|
|
| if __name__ == '__main__': |
| css = '.image-container img { width: 100%; max-height: 320px;}' |
|
|
| with gr.Blocks(css=css) as demo: |
| gr.HTML(""" |
| <h1 style='text-align: center;'><a href="https://github.com/Topdu/OpenOCR">OpenOCR</a></h1> |
| <p style='text-align: center;'>准确高效的通用 OCR 系统 (由<a href="https://fvl.fudan.edu.cn">FVL实验室</a> <a href="https://github.com/Topdu/OpenOCR">OCR Team</a> 创建) <a href="https://github.com/Topdu/OpenOCR/tree/main?tab=readme-ov-file#quick-start">[本地快速部署]</a></p>""" |
| ) |
| with gr.Row(): |
| with gr.Column(scale=1): |
| input_image = gr.Image(label='Input image', |
| elem_classes=['image-container']) |
|
|
| examples = gr.Examples(examples=e2e_img_example, |
| inputs=input_image, |
| label='Examples') |
| downstream = gr.Button('Run') |
|
|
| |
| with gr.Column(): |
| with gr.Row(): |
| det_input_size_textbox = gr.Number( |
| label='Detection Input Size', |
| value=960, |
| info='检测网络输入尺寸的最长边,默认为960。') |
| det_score_mode_dropdown = gr.Dropdown( |
| ['slow', 'fast'], |
| value='slow', |
| label='Detection Score Mode', |
| info='文本框的置信度计算模式,默认为 slow。slow 模式计算速度较慢,但准确度较高。fast 模式计算速度较快,但准确度较低。' |
| ) |
| with gr.Row(): |
| rec_drop_score_slider = gr.Slider( |
| 0.0, |
| 1.0, |
| value=0.4, |
| step=0.01, |
| label='Recognition Drop Score', |
| info='识别置信度阈值,默认值为0.4。低于该阈值的识别结果和对应的文本框被丢弃。') |
| mask_thresh_slider = gr.Slider( |
| 0.0, |
| 1.0, |
| value=0.3, |
| step=0.01, |
| label='Mask Threshold', |
| info='Mask 阈值,用于二值化 mask,默认值为0.3。如果存在文本截断时,请调低该值。') |
| with gr.Row(): |
| box_thresh_slider = gr.Slider( |
| 0.0, |
| 1.0, |
| value=0.6, |
| step=0.01, |
| label='Box Threshold', |
| info='文本框置信度阈值,默认值为0.6。如果存在文本被漏检时,请调低该值。') |
| unclip_ratio_slider = gr.Slider( |
| 1.5, |
| 2.0, |
| value=1.5, |
| step=0.05, |
| label='Unclip Ratio', |
| info='文本框解析时的膨胀系数,默认值为1.5。值越大文本框越大。') |
|
|
| |
| model_type_dropdown = gr.Dropdown( |
| ['mobile', 'server'], |
| value='mobile', |
| label='Model Type', |
| info='选择 OCR 模型类型:高效率模型mobile,高精度模型server。') |
|
|
| with gr.Column(scale=1): |
| img_mask = gr.Image(label='mask', |
| interactive=False, |
| elem_classes=['image-container']) |
| img_output = gr.Image(label=' ', |
| interactive=False, |
| elem_classes=['image-container']) |
|
|
| output = gr.Textbox(label='Result') |
| confidence = gr.Textbox(label='Latency') |
|
|
| downstream.click(fn=main, |
| inputs=[ |
| input_image, model_type_dropdown, |
| det_input_size_textbox, rec_drop_score_slider, |
| mask_thresh_slider, box_thresh_slider, |
| unclip_ratio_slider, det_score_mode_dropdown |
| ], |
| outputs=[ |
| output, |
| confidence, |
| img_output, |
| img_mask, |
| ]) |
|
|
| demo.launch(share=True) |
|
|