davanstrien's picture
davanstrien HF Staff
Add deepseek-ai/DeepSeek-OCR OCR results (50 samples) [deepseek-ocr]
fc01577 verified
|
raw
history blame
3.18 kB
metadata
tags:
  - ocr
  - document-processing
  - dots-ocr
  - multilingual
  - markdown
  - uv-script
  - generated
configs:
  - config_name: deepseek-ocr
    data_files:
      - split: train
        path: deepseek-ocr/train-*
dataset_info:
  config_name: deepseek-ocr
  features:
    - name: image
      dtype: image
    - name: drawer_id
      dtype: string
    - name: card_number
      dtype: int64
    - name: filename
      dtype: string
    - name: text
      dtype: string
    - name: has_ocr
      dtype: bool
    - name: source
      dtype: string
    - name: source_url
      dtype: string
    - name: ia_collection
      dtype: string
    - name: markdown
      dtype: string
    - name: inference_info
      dtype: string
  splits:
    - name: train
      num_bytes: 14636668
      num_examples: 50
  download_size: 14453748
  dataset_size: 14636668

Document OCR using dots.ocr

This dataset contains OCR results from images in biglam/rubenstein-manuscript-catalog using DoTS.ocr, a compact 1.7B multilingual model.

Processing Details

Configuration

  • Image Column: image
  • Output Column: markdown
  • Dataset Split: train
  • Batch Size: 16
  • Prompt Mode: ocr
  • Max Model Length: 8,192 tokens
  • Max Output Tokens: 8,192
  • GPU Memory Utilization: 80.0%

Model Information

DoTS.ocr is a compact multilingual document parsing model that excels at:

  • 🌍 100+ Languages - Multilingual document support
  • 📊 Table extraction - Structured data recognition
  • 📐 Formulas - Mathematical notation preservation
  • 📝 Layout-aware - Reading order and structure preservation
  • 🎯 Compact - Only 1.7B parameters

Dataset Structure

The dataset contains all original columns plus:

  • markdown: The extracted text in markdown format
  • inference_info: JSON list tracking all OCR models applied to this dataset

Usage

from datasets import load_dataset
import json

# Load the dataset
dataset = load_dataset("{output_dataset_id}", split="train")

# Access the markdown text
for example in dataset:
    print(example["markdown"])
    break

# View all OCR models applied to this dataset
inference_info = json.loads(dataset[0]["inference_info"])
for info in inference_info:
    print(f"Column: {info['column_name']} - Model: {info['model_id']}")

Reproduction

This dataset was generated using the uv-scripts/ocr DoTS OCR script:

uv run https://huggingface.co/datasets/uv-scripts/ocr/raw/main/dots-ocr.py \
    biglam/rubenstein-manuscript-catalog \
    <output-dataset> \
    --image-column image \
    --batch-size 16 \
    --prompt-mode ocr \
    --max-model-len 8192 \
    --max-tokens 8192 \
    --gpu-memory-utilization 0.8

Generated with 🤖 UV Scripts