| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| """ |
| Convert document images to text/tables/formulas using PaddleOCR-VL-1.5 with transformers. |
| |
| PaddleOCR-VL-1.5 is a compact 0.9B OCR model that achieves 94.5% SOTA accuracy on |
| OmniDocBench v1.5. It supports multiple task modes including OCR, table recognition, |
| formula extraction, chart analysis, text spotting, and seal recognition. |
| |
| NOTE: This script uses transformers batch inference (not vLLM) because PaddleOCR-VL |
| only supports vLLM in server mode, which doesn't fit the UV script pattern. |
| Transformers batch inference via apply_chat_template(padding=True) provides |
| efficient batching while keeping the single-command interface. |
| |
| Features: |
| - 🎯 SOTA Performance: 94.5% on OmniDocBench v1.5 |
| - 🧩 Ultra-compact: Only 0.9B parameters |
| - 📝 OCR mode: General text extraction to markdown |
| - 📊 Table mode: HTML table recognition |
| - 📐 Formula mode: LaTeX mathematical notation |
| - 📈 Chart mode: Structured chart analysis |
| - 🔍 Spotting mode: Text spotting with localization |
| - 🔖 Seal mode: Seal/stamp recognition |
| - 🌍 Multilingual support |
| - ⚡ Fast inference with batch processing |
| |
| Model: PaddlePaddle/PaddleOCR-VL-1.5 |
| Backend: Transformers (batch inference) |
| Performance: 94.5% SOTA on OmniDocBench v1.5 |
| """ |
|
|
| import argparse |
| import io |
| import json |
| import logging |
| import os |
| import sys |
| import time |
| from datetime import datetime |
| from typing import Any, Dict, List, Union |
|
|
| import torch |
| from datasets import load_dataset |
| from huggingface_hub import DatasetCard, login |
| from PIL import Image |
| from tqdm.auto import tqdm |
|
|
| logging.basicConfig(level=logging.INFO) |
| logger = logging.getLogger(__name__) |
|
|
|
|
| |
| MODEL_ID = "PaddlePaddle/PaddleOCR-VL-1.5" |
|
|
| |
| TASK_MODES = { |
| "ocr": "OCR:", |
| "table": "Table Recognition:", |
| "formula": "Formula Recognition:", |
| "chart": "Chart Recognition:", |
| "spotting": "Spotting:", |
| "seal": "Seal Recognition:", |
| } |
|
|
| |
| TASK_DESCRIPTIONS = { |
| "ocr": "General text extraction to markdown format", |
| "table": "Table extraction to HTML format", |
| "formula": "Mathematical formula recognition to LaTeX", |
| "chart": "Chart and diagram analysis", |
| "spotting": "Text spotting with localization", |
| "seal": "Seal and stamp recognition", |
| } |
|
|
| |
| MAX_PIXELS = { |
| "spotting": 2048 * 28 * 28, |
| "default": 1280 * 28 * 28, |
| } |
|
|
|
|
| def check_cuda_availability(): |
| """Check if CUDA is available and exit if not.""" |
| if not torch.cuda.is_available(): |
| logger.error("CUDA is not available. This script requires a GPU.") |
| logger.error("Please run on a machine with a CUDA-capable GPU.") |
| sys.exit(1) |
| else: |
| logger.info(f"CUDA is available. GPU: {torch.cuda.get_device_name(0)}") |
|
|
|
|
| def prepare_image( |
| image: Union[Image.Image, Dict[str, Any], str], |
| ) -> Image.Image: |
| """ |
| Prepare image for PaddleOCR-VL-1.5 processing. |
| |
| Note: Image resizing is handled by the processor via images_kwargs. |
| |
| Args: |
| image: PIL Image, dict with bytes, or file path |
| |
| Returns: |
| Processed PIL Image in RGB format |
| """ |
| |
| if isinstance(image, Image.Image): |
| pil_img = image |
| elif isinstance(image, dict) and "bytes" in image: |
| pil_img = Image.open(io.BytesIO(image["bytes"])) |
| elif isinstance(image, str): |
| pil_img = Image.open(image) |
| else: |
| raise ValueError(f"Unsupported image type: {type(image)}") |
|
|
| |
| pil_img = pil_img.convert("RGB") |
|
|
| return pil_img |
|
|
|
|
| def create_message(image: Image.Image, task_mode: str) -> List[Dict]: |
| """ |
| Create chat message for PaddleOCR-VL-1.5 processing. |
| |
| Args: |
| image: Prepared PIL Image |
| task_mode: Task mode (ocr, table, formula, chart, spotting, seal) |
| |
| Returns: |
| Message in chat format |
| """ |
| return [ |
| { |
| "role": "user", |
| "content": [ |
| {"type": "image", "image": image}, |
| {"type": "text", "text": TASK_MODES[task_mode]}, |
| ], |
| } |
| ] |
|
|
|
|
| def create_dataset_card( |
| source_dataset: str, |
| model: str, |
| task_mode: str, |
| num_samples: int, |
| processing_time: str, |
| max_tokens: int, |
| image_column: str = "image", |
| split: str = "train", |
| ) -> str: |
| """Create a dataset card documenting the OCR process.""" |
| task_description = TASK_DESCRIPTIONS[task_mode] |
|
|
| return f"""--- |
| tags: |
| - ocr |
| - document-processing |
| - paddleocr-vl-1.5 |
| - {task_mode} |
| - uv-script |
| - generated |
| --- |
| |
| # Document Processing using PaddleOCR-VL-1.5 ({task_mode.upper()} mode) |
| |
| This dataset contains {task_mode.upper()} results from images in [{source_dataset}](https://huggingface.co/datasets/{source_dataset}) using PaddleOCR-VL-1.5, an ultra-compact 0.9B SOTA OCR model. |
| |
| ## Processing Details |
| |
| - **Source Dataset**: [{source_dataset}](https://huggingface.co/datasets/{source_dataset}) |
| - **Model**: [{model}](https://huggingface.co/{model}) |
| - **Task Mode**: `{task_mode}` - {task_description} |
| - **Number of Samples**: {num_samples:,} |
| - **Processing Time**: {processing_time} |
| - **Processing Date**: {datetime.now().strftime("%Y-%m-%d %H:%M UTC")} |
| |
| ### Configuration |
| |
| - **Image Column**: `{image_column}` |
| - **Output Column**: `paddleocr_1.5_{task_mode}` |
| - **Dataset Split**: `{split}` |
| - **Max Output Tokens**: {max_tokens:,} |
| - **Backend**: Transformers (single image processing) |
| |
| ## Model Information |
| |
| PaddleOCR-VL-1.5 is a state-of-the-art, resource-efficient model for document parsing: |
| - 🎯 **SOTA Performance** - 94.5% on OmniDocBench v1.5 |
| - 🧩 **Ultra-compact** - Only 0.9B parameters |
| - 📝 **OCR mode** - General text extraction |
| - 📊 **Table mode** - HTML table recognition |
| - 📐 **Formula mode** - LaTeX mathematical notation |
| - 📈 **Chart mode** - Structured chart analysis |
| - 🔍 **Spotting mode** - Text spotting with localization |
| - 🔖 **Seal mode** - Seal and stamp recognition |
| - 🌍 **Multilingual** - Support for multiple languages |
| - ⚡ **Fast** - Efficient batch inference |
| |
| ### Task Modes |
| |
| - **OCR**: Extract text content to markdown format |
| - **Table Recognition**: Extract tables to HTML format |
| - **Formula Recognition**: Extract mathematical formulas to LaTeX |
| - **Chart Recognition**: Analyze and describe charts/diagrams |
| - **Spotting**: Text spotting with location information |
| - **Seal Recognition**: Extract text from seals and stamps |
| |
| ## Dataset Structure |
| |
| The dataset contains all original columns plus: |
| - `paddleocr_1.5_{task_mode}`: The extracted content based on task mode |
| - `inference_info`: JSON list tracking all OCR models applied to this dataset |
| |
| ## Usage |
| |
| ```python |
| from datasets import load_dataset |
| import json |
| |
| # Load the dataset |
| dataset = load_dataset("{{output_dataset_id}}", split="{split}") |
| |
| # Access the extracted content |
| for example in dataset: |
| print(example["paddleocr_1.5_{task_mode}"]) |
| break |
| |
| # View all OCR models applied to this dataset |
| inference_info = json.loads(dataset[0]["inference_info"]) |
| for info in inference_info: |
| print(f"Task: {{info['task_mode']}} - Model: {{info['model_id']}}") |
| ``` |
| |
| ## Reproduction |
| |
| This dataset was generated using the [uv-scripts/ocr](https://huggingface.co/datasets/uv-scripts/ocr) PaddleOCR-VL-1.5 script: |
| |
| ```bash |
| uv run https://huggingface.co/datasets/uv-scripts/ocr/raw/main/paddleocr-vl-1.5.py \\ |
| {source_dataset} \\ |
| <output-dataset> \\ |
| --task-mode {task_mode} \\ |
| --image-column {image_column} |
| ``` |
| |
| ## Performance |
| |
| - **Model Size**: 0.9B parameters |
| - **Benchmark Score**: 94.5% SOTA on OmniDocBench v1.5 |
| - **Processing Speed**: ~{num_samples / (float(processing_time.split()[0]) * 60):.2f} images/second |
| - **Backend**: Transformers (single image processing) |
| |
| Generated with 🤖 [UV Scripts](https://huggingface.co/uv-scripts) |
| """ |
|
|
|
|
| def main( |
| input_dataset: str, |
| output_dataset: str, |
| image_column: str = "image", |
| task_mode: str = "ocr", |
| max_tokens: int = 512, |
| hf_token: str = None, |
| split: str = "train", |
| max_samples: int = None, |
| private: bool = False, |
| shuffle: bool = False, |
| seed: int = 42, |
| output_column: str = "markdown", |
| config: str = None, |
| create_pr: bool = False, |
| verbose: bool = False, |
| ): |
| """Process images from HF dataset through PaddleOCR-VL-1.5 model.""" |
|
|
| |
| check_cuda_availability() |
|
|
| |
| start_time = datetime.now() |
|
|
| |
| HF_TOKEN = hf_token or os.environ.get("HF_TOKEN") |
| if HF_TOKEN: |
| login(token=HF_TOKEN) |
|
|
| |
| if task_mode not in TASK_MODES: |
| raise ValueError( |
| f"Invalid task_mode '{task_mode}'. Choose from: {list(TASK_MODES.keys())}" |
| ) |
|
|
| logger.info(f"Using task mode: {task_mode} - {TASK_DESCRIPTIONS[task_mode]}") |
| logger.info(f"Output will be written to column: {output_column}") |
|
|
| |
| logger.info(f"Loading dataset: {input_dataset}") |
| dataset = load_dataset(input_dataset, split=split) |
|
|
| |
| if image_column not in dataset.column_names: |
| raise ValueError( |
| f"Column '{image_column}' not found. Available: {dataset.column_names}" |
| ) |
|
|
| |
| if shuffle: |
| logger.info(f"Shuffling dataset with seed {seed}") |
| dataset = dataset.shuffle(seed=seed) |
|
|
| |
| if max_samples: |
| dataset = dataset.select(range(min(max_samples, len(dataset)))) |
| logger.info(f"Limited to {len(dataset)} samples") |
|
|
| |
| logger.info(f"Loading model: {MODEL_ID}") |
| logger.info("This may take a minute on first run...") |
|
|
| from transformers import AutoModelForImageTextToText, AutoProcessor |
|
|
| |
| processor = AutoProcessor.from_pretrained(MODEL_ID) |
| |
| processor.tokenizer.padding_side = "left" |
|
|
| model = ( |
| AutoModelForImageTextToText.from_pretrained( |
| MODEL_ID, |
| torch_dtype=torch.bfloat16, |
| ) |
| .to("cuda") |
| .eval() |
| ) |
|
|
| logger.info(f"Model loaded on {next(model.parameters()).device}") |
| max_pixels = MAX_PIXELS.get(task_mode, MAX_PIXELS["default"]) |
| logger.info(f"Processing {len(dataset)} images (one at a time for stability)") |
| logger.info(f"Image resizing: max_pixels={max_pixels:,} (handled by processor)") |
|
|
| |
| |
| |
| all_outputs = [] |
|
|
| for i in tqdm(range(len(dataset)), desc=f"PaddleOCR-VL-1.5 {task_mode.upper()}"): |
| try: |
| |
| image = dataset[i][image_column] |
| pil_image = prepare_image(image) |
| messages = create_message(pil_image, task_mode) |
|
|
| |
| inputs = processor.apply_chat_template( |
| messages, |
| add_generation_prompt=True, |
| tokenize=True, |
| return_dict=True, |
| return_tensors="pt", |
| images_kwargs={ |
| "size": { |
| "shortest_edge": processor.image_processor.min_pixels, |
| "longest_edge": max_pixels, |
| } |
| }, |
| ).to(model.device) |
|
|
| |
| with torch.no_grad(): |
| outputs = model.generate( |
| **inputs, |
| max_new_tokens=max_tokens, |
| do_sample=False, |
| ) |
|
|
| |
| input_len = inputs["input_ids"].shape[1] |
| generated_ids = outputs[0, input_len:] |
| result = processor.decode(generated_ids, skip_special_tokens=True) |
| all_outputs.append(result.strip()) |
|
|
| except Exception as e: |
| logger.error(f"Error processing image {i}: {e}") |
| all_outputs.append(f"[{task_mode.upper()} ERROR: {str(e)[:100]}]") |
|
|
| |
| processing_duration = datetime.now() - start_time |
| processing_time_str = f"{processing_duration.total_seconds() / 60:.1f} min" |
|
|
| |
| logger.info(f"Adding '{output_column}' column to dataset") |
| dataset = dataset.add_column(output_column, all_outputs) |
|
|
| |
| inference_entry = { |
| "model_id": MODEL_ID, |
| "model_name": "PaddleOCR-VL-1.5", |
| "model_size": "0.9B", |
| "task_mode": task_mode, |
| "column_name": output_column, |
| "timestamp": datetime.now().isoformat(), |
| "max_tokens": max_tokens, |
| "backend": "transformers", |
| } |
|
|
| if "inference_info" in dataset.column_names: |
| |
| logger.info("Updating existing inference_info column") |
|
|
| def update_inference_info(example): |
| try: |
| existing_info = ( |
| json.loads(example["inference_info"]) |
| if example["inference_info"] |
| else [] |
| ) |
| except (json.JSONDecodeError, TypeError): |
| existing_info = [] |
|
|
| existing_info.append(inference_entry) |
| return {"inference_info": json.dumps(existing_info)} |
|
|
| dataset = dataset.map(update_inference_info) |
| else: |
| |
| logger.info("Creating new inference_info column") |
| inference_list = [json.dumps([inference_entry])] * len(dataset) |
| dataset = dataset.add_column("inference_info", inference_list) |
|
|
| |
| logger.info(f"Pushing to {output_dataset}") |
| max_retries = 3 |
| for attempt in range(1, max_retries + 1): |
| try: |
| if attempt > 1: |
| logger.warning("Disabling XET (fallback to HTTP upload)") |
| os.environ["HF_HUB_DISABLE_XET"] = "1" |
| dataset.push_to_hub( |
| output_dataset, |
| private=private, |
| token=HF_TOKEN, |
| max_shard_size="500MB", |
| **({"config_name": config} if config else {}), |
| create_pr=create_pr, |
| commit_message=f"Add {MODEL_ID} OCR results ({len(dataset)} samples)" |
| + (f" [{config}]" if config else ""), |
| ) |
| break |
| except Exception as e: |
| logger.error(f"Upload attempt {attempt}/{max_retries} failed: {e}") |
| if attempt < max_retries: |
| delay = 30 * (2 ** (attempt - 1)) |
| logger.info(f"Retrying in {delay}s...") |
| time.sleep(delay) |
| else: |
| logger.error("All upload attempts failed. OCR results are lost.") |
| sys.exit(1) |
|
|
| |
| logger.info("Creating dataset card") |
| card_content = create_dataset_card( |
| source_dataset=input_dataset, |
| model=MODEL_ID, |
| task_mode=task_mode, |
| num_samples=len(dataset), |
| processing_time=processing_time_str, |
| max_tokens=max_tokens, |
| image_column=image_column, |
| split=split, |
| ) |
|
|
| card = DatasetCard(card_content) |
| card.push_to_hub(output_dataset, token=HF_TOKEN) |
|
|
| logger.info("PaddleOCR-VL-1.5 processing complete!") |
| logger.info( |
| f"Dataset available at: https://huggingface.co/datasets/{output_dataset}" |
| ) |
| logger.info(f"Processing time: {processing_time_str}") |
| logger.info( |
| f"Processing speed: {len(dataset) / processing_duration.total_seconds():.2f} images/sec" |
| ) |
| logger.info(f"Task mode: {task_mode} - {TASK_DESCRIPTIONS[task_mode]}") |
|
|
| if verbose: |
| import importlib.metadata |
|
|
| logger.info("--- Resolved package versions ---") |
| for pkg in ["vllm", "transformers", "torch", "datasets", "pyarrow", "pillow"]: |
| try: |
| logger.info(f" {pkg}=={importlib.metadata.version(pkg)}") |
| except importlib.metadata.PackageNotFoundError: |
| logger.info(f" {pkg}: not installed") |
| logger.info("--- End versions ---") |
|
|
|
|
| if __name__ == "__main__": |
| |
| if len(sys.argv) == 1: |
| print("=" * 80) |
| print("PaddleOCR-VL-1.5 Document Processing") |
| print("=" * 80) |
| print("\nSOTA 0.9B OCR model (94.5% on OmniDocBench v1.5)") |
| print("\nFeatures:") |
| print("- 🎯 SOTA Performance - 94.5% on OmniDocBench v1.5") |
| print("- 🧩 Ultra-compact - Only 0.9B parameters") |
| print("- 📝 OCR mode - General text extraction") |
| print("- 📊 Table mode - HTML table recognition") |
| print("- 📐 Formula mode - LaTeX mathematical notation") |
| print("- 📈 Chart mode - Structured chart analysis") |
| print("- 🔍 Spotting mode - Text spotting with localization") |
| print("- 🔖 Seal mode - Seal and stamp recognition") |
| print("- 🌍 Multilingual support") |
| print("- ⚡ Fast batch inference with transformers") |
| print("\nTask Modes:") |
| for mode, description in TASK_DESCRIPTIONS.items(): |
| print(f" {mode:10} - {description}") |
| print("\nExample usage:") |
| print("\n1. Basic OCR (default mode):") |
| print(" uv run paddleocr-vl-1.5.py input-dataset output-dataset") |
| print("\n2. Table extraction:") |
| print(" uv run paddleocr-vl-1.5.py docs tables-extracted --task-mode table") |
| print("\n3. Formula recognition:") |
| print(" uv run paddleocr-vl-1.5.py papers formulas --task-mode formula") |
| print("\n4. Text spotting (higher resolution):") |
| print(" uv run paddleocr-vl-1.5.py images spotted --task-mode spotting") |
| print("\n5. Seal recognition:") |
| print(" uv run paddleocr-vl-1.5.py seals recognized --task-mode seal") |
| print("\n6. Test with small sample:") |
| print(" uv run paddleocr-vl-1.5.py dataset test --max-samples 10 --shuffle") |
| print("\n7. Running on HF Jobs:") |
| print(" hf jobs uv run --flavor l4x1 \\") |
| print(" -s HF_TOKEN \\") |
| print( |
| " https://huggingface.co/datasets/uv-scripts/ocr/raw/main/paddleocr-vl-1.5.py \\" |
| ) |
| print(" input-dataset output-dataset --task-mode ocr") |
| print("\n" + "=" * 80) |
| print("\nBackend: Transformers (batch inference)") |
| print( |
| "Note: Uses transformers instead of vLLM for single-command UV script compatibility" |
| ) |
| print("\nFor full help, run: uv run paddleocr-vl-1.5.py --help") |
| sys.exit(0) |
|
|
| parser = argparse.ArgumentParser( |
| description="Document processing using PaddleOCR-VL-1.5 (0.9B SOTA model, transformers backend)", |
| formatter_class=argparse.RawDescriptionHelpFormatter, |
| epilog=""" |
| Task Modes: |
| ocr General text extraction to markdown (default) |
| table Table extraction to HTML format |
| formula Mathematical formula recognition to LaTeX |
| chart Chart and diagram analysis |
| spotting Text spotting with localization (higher resolution) |
| seal Seal and stamp recognition |
| |
| Examples: |
| # Basic text OCR |
| uv run paddleocr-vl-1.5.py my-docs analyzed-docs |
| |
| # Extract tables from documents |
| uv run paddleocr-vl-1.5.py papers tables --task-mode table |
| |
| # Recognize mathematical formulas |
| uv run paddleocr-vl-1.5.py textbooks formulas --task-mode formula |
| |
| # Text spotting (uses higher resolution) |
| uv run paddleocr-vl-1.5.py images spotted --task-mode spotting |
| |
| # Seal recognition |
| uv run paddleocr-vl-1.5.py documents seals --task-mode seal |
| |
| # Test with random sampling |
| uv run paddleocr-vl-1.5.py large-dataset test --max-samples 50 --shuffle --task-mode ocr |
| |
| # Disable smart resize for original resolution |
| uv run paddleocr-vl-1.5.py images output --no-smart-resize |
| |
| Backend: Transformers batch inference (not vLLM) |
| """, |
| ) |
|
|
| parser.add_argument("input_dataset", help="Input dataset ID from Hugging Face Hub") |
| parser.add_argument("output_dataset", help="Output dataset ID for Hugging Face Hub") |
| parser.add_argument( |
| "--image-column", |
| default="image", |
| help="Column containing images (default: image)", |
| ) |
| parser.add_argument( |
| "--task-mode", |
| choices=list(TASK_MODES.keys()), |
| default="ocr", |
| help="Task type: ocr (default), table, formula, chart, spotting, or seal", |
| ) |
| parser.add_argument( |
| "--max-tokens", |
| type=int, |
| default=512, |
| help="Maximum tokens to generate (default: 512, per model card element-level example; increase for full pages)", |
| ) |
| parser.add_argument("--hf-token", help="Hugging Face API token") |
| parser.add_argument( |
| "--split", default="train", help="Dataset split to use (default: train)" |
| ) |
| parser.add_argument( |
| "--max-samples", |
| type=int, |
| help="Maximum number of samples to process (for testing)", |
| ) |
| parser.add_argument( |
| "--private", action="store_true", help="Make output dataset private" |
| ) |
| parser.add_argument( |
| "--shuffle", action="store_true", help="Shuffle dataset before processing" |
| ) |
| parser.add_argument( |
| "--seed", |
| type=int, |
| default=42, |
| help="Random seed for shuffling (default: 42)", |
| ) |
| parser.add_argument( |
| "--output-column", |
| default="markdown", |
| help="Column name for output text (default: markdown)", |
| ) |
| parser.add_argument( |
| "--config", |
| help="Config/subset name when pushing to Hub (for benchmarking multiple models in one repo)", |
| ) |
| parser.add_argument( |
| "--create-pr", |
| action="store_true", |
| help="Create a pull request instead of pushing directly (for parallel benchmarking)", |
| ) |
| parser.add_argument( |
| "--verbose", |
| action="store_true", |
| help="Log resolved package versions after processing (useful for pinning deps)", |
| ) |
|
|
| args = parser.parse_args() |
|
|
| main( |
| input_dataset=args.input_dataset, |
| output_dataset=args.output_dataset, |
| image_column=args.image_column, |
| task_mode=args.task_mode, |
| max_tokens=args.max_tokens, |
| hf_token=args.hf_token, |
| split=args.split, |
| max_samples=args.max_samples, |
| private=args.private, |
| shuffle=args.shuffle, |
| seed=args.seed, |
| output_column=args.output_column, |
| config=args.config, |
| create_pr=args.create_pr, |
| verbose=args.verbose, |
| ) |
|
|