Dataset Viewer
Auto-converted to Parquet Duplicate
Search is not available for this dataset
image
image
End of preview. Expand in Data Studio

YAML Metadata Warning:empty or missing yaml metadata in repo card

Check out the documentation for more information.

Wild-OmniDocBench

A Real-World Captured Document Parsing Benchmark for Robustness Evaluation

中文版PaperGitHubHuggingFace

Overview

Wild-OmniDocBench is a benchmark for evaluating document parsing robustness under real-world captured conditions. It is derived from OmniDocBench by converting scanned/digital documents into naturally captured images through controlled physical simulation, including printing, deformation, and photography under diverse lighting conditions.

Unlike standard benchmarks that rely on clean scanned or digital-born pages, Wild-OmniDocBench introduces realistic artifacts such as:

  • Geometric distortions (perspective shifts, bends, wrinkles)
  • Illumination variations (directional, uneven, low-light)
  • Screen capture artifacts (moire patterns, reflections)
  • Environmental interference (background overlays, shadows)

Note: The current release of Wild-OmniDocBench corresponds to OmniDocBench v1.5. We are currently processing the extended portions for v1.6 and will release them in a future update.

Wild-OmniDocBench Construction

Benchmark Statistics

Item Details
Total Images 1,350
Source Real-world captured variant of OmniDocBench
Document Types Books, Textbooks, Papers, PPTs, Newspapers, Notes, Exams, Magazines, Financial Reports, etc.
Capture Methods (i) Print + physical deformation + photography; (ii) Screen display + re-capture
Annotations Inherited from OmniDocBench (full structural and reading-order annotations)

Data Format

Directory Structure

Wild_OmniDocBench/
├── README.md                   # English README
├── README_ZH.md                # Chinese README
├── wild_omnidocbench.zip       # Benchmark images (1,350 JPGs)
└── assets/
    └── overview.png            # Overview figure

Images

After unzipping wild_omnidocbench.zip, images are named following the OmniDocBench convention:

{doc_type}_{language}_{source}_{page}.jpg

For example: book_en_A.Concise.Introduction.to.Linear.Algebra_page_065.jpg

Evaluation

Wild-OmniDocBench uses the same annotation format and evaluation protocol as OmniDocBench. To evaluate on Wild-OmniDocBench:

  1. Obtain annotations and evaluation scripts from the official OmniDocBench repository:

    https://github.com/opendatalab/OmniDocBench
    
  2. Replace the image source with Wild-OmniDocBench images (from wild_omnidocbench.zip).

  3. Run evaluation following the OmniDocBench protocol. Metrics include:

    • Overall Score (↑)
    • Text Edit Distance (↓)
    • Formula CDM (↑)
    • Table TEDS (↑)
    • Reading Order Edit Distance (↓)

Key Results

Performance degradation from OmniDocBench to Wild-OmniDocBench (from the DocHumming paper):

Model Type Overall (Origin) Overall (Wild) Degradation
DocHumming (1B) End2End 93.75 87.03 −6.72
dots.ocr (3B) End2End 88.41 78.01 −10.40
Qwen3-VL (235B) General 89.15 79.69 −9.46
MinerU2.5 (1.2B) Modular 90.67 70.91 −19.76
PaddleOCR-VL (0.9B) Modular 91.93 72.19 −19.74

End-to-end models exhibit significantly less degradation than modular cascaded pipelines under real-world capture conditions.

Citation

@misc{li2026towardsrealworlddocument,
      title={Towards Real-World Document Parsing via Realistic Scene Synthesis and Document-Aware Training},
      author={Gengluo Li and Pengyuan Lyu and Chengquan Zhang and Huawen Shen and Liang Wu and Xingyu Wan and Gangyan Zeng and Han Hu and Can Ma and Yu Zhou},
      year={2026},
      journal={arXiv preprint arXiv:2603.23885},
      url={https://arxiv.org/abs/2603.23885},
}

Acknowledgements

Wild-OmniDocBench is built upon OmniDocBench. We thank the OmniDocBench team for providing the original annotations and evaluation framework.

License

This benchmark is released for research purposes only.

Downloads last month
-

Paper for VirtualLUO/Wild_OmniDocBench