# Real5-OmniDocBench
**Real5-OmniDocBench** is a brand-new benchmark oriented toward real-world scenarios, which we constructed based on the OmniDocBench v1.5 dataset. The dataset comprises five distinct scenarios: Scanning, Warping, Screen-Photography, Illumination, and Skew. Apart from the Scanning category, all images were manually acquired via handheld mobile devices to closely simulate real-world conditions. Each subset maintains a one-to-one correspondence with the original OmniDocBench, strictly adhering to its ground-truth annotations and evaluation protocols. Given its empirical and realistic nature, this dataset serves as a rigorous benchmark for assessing the robustness of document parsing models in practical applications.
---
## Key Features
### 1. Real-world Scenarios
Real5-OmniDocBench contains five challenging and representative photographic scenarios:
* **Scanning**: Images captured by scanning devices, simulating flat and clean document scans.
* **Warping**: Photos of documents with visible page curvatures, mimicking distortions caused by book spines or folding.
* **Skew**: Documents photographed at an angle, resulting in perspective distortion.
* **Screen-Photography**: Photographs of screens displaying documents, introducing moiré patterns and reflections.
* **Illumination**: Images taken under varied lighting conditions, including shadows and glare.
### 2. Comprehensive Coverage
For each scenario, the dataset contains **1,355 images**, corresponding exactly to the original pages in the OmniDocBench evaluation set. This enables direct, controlled comparison of model performance across different real-world conditions.
### 3. High-Quality Realism
All images are captured manually under real-world conditions. Care was taken to cover typical distortions and artifacts encountered in mobile or camera-based document digitization.
### 4. Evaluation Protocol Compatibility
The evaluation scripts and metrics from OmniDocBench (e.g., Normalized Edit Distance, BLEU, METEOR, TEDS, COCODet) are fully compatible and directly applicable to Real5-OmniDocBench.
---
## Dataset Structure
The dataset follows the directory structure shown below:
```text
Real5-OmniDocBench/
├── Real5-OmniDocBench-Warping/
│ ├── book_en_[搬书匠#20][HTML5 Canvas].2011.英文版_page_208.png
│ └── ...
├── Real5-OmniDocBench-Illumination/
│ └── ...
├── Real5-OmniDocBench-Scanning/
│ └── ...
├── Real5-OmniDocBench-Screen-Photography/
│ └── ...
└── Real5-OmniDocBench-Skew/
└── ...
```
## Usage & Evaluation
* **Input:** For each scenario, models should take the corresponding set of 1,355 images as input.
* **Output:** Model predictions should follow the same format as OmniDocBench, e.g., Markdown for end-to-end parsing.
* **Evaluation:** Use the official OmniDocBench evaluation scripts and metrics for assessment. Direct comparison across scenarios is encouraged to measure model robustness.
---
## Benchmark Results
To illustrate the impact of real-world visual distortions on document parsing, we evaluated several mainstream models across all five scenarios in Real5-OmniDocBench. The same metrics as OmniDocBench are used: **Overall↑**, **TextEdit↓**, **FormulaCDM↑**, **TableTEDS↑**, and **Reading OrderEdit↓**.
### 1. Scanning
| Model Type |
Methods |
Parameters |
Overall↑ |
TextEdit↓ |
FormulaCDM↑ |
TableTEDS↑ |
Reading OrderEdit↓ |
| Pipeline Tools |
Maker-1.8.2 |
- |
70.27 |
0.223 |
77.03 |
56.05 |
0.238 |
| PP-StructureV3 |
- |
84.68 |
0.094 |
84.34 |
79.06 |
0.092 |
| General VLMs |
GPT-5.2 |
- |
84.43 |
0.142 |
85.68 |
81.78 |
0.109 |
| Qwen2.5-VL-72B |
72B |
86.19 |
0.110 |
86.14 |
83.41 |
0.114 |
| Gemini-2.5 Pro |
- |
89.25 |
0.073 |
87.44 |
87.62 |
0.098 |
| Qwen3-VL-235B-A22B-Instruct |
235B |
89.43 |
0.059 |
89.01 |
85.19 |
0.066 |
| Gemini-3 Pro |
- |
89.47 |
0.071 |
88.16 |
87.37 |
0.078 |
| Specialized VLMs |
Dolphin |
322M |
72.16 |
0.154 |
64.58 |
67.27 |
0.130 |
| Dolphin-1.5 |
0.3B |
83.39 |
0.097 |
76.25 |
83.65 |
0.090 |
| MinerU2-VLM |
0.9B |
83.60 |
0.094 |
79.76 |
80.44 |
0.091 |
| MonkeyOCR-pro-1.2B |
1.9B |
84.64 |
0.123 |
84.17 |
82.13 |
0.145 |
| MonkeyOCR-3B |
3.7B |
84.65 |
0.100 |
84.16 |
79.81 |
0.143 |
| Nanonets-OCR-s |
3B |
85.52 |
0.106 |
88.09 |
79.11 |
0.106 |
| Deepseek-OCR |
3B |
86.17 |
0.078 |
83.59 |
82.69 |
0.085 |
| dots.ocr |
3B |
86.87 |
0.083 |
83.27 |
85.68 |
0.081 |
| MonkeyOCR-pro-3B |
3.7B |
86.94 |
0.103 |
86.29 |
84.86 |
0.141 |
| MinerU2.5 |
1.2B |
90.06 |
0.052 |
88.22 |
87.16 |
0.050 |
| PaddleOCR-VL |
0.9B |
92.11 |
0.039 |
90.35 |
89.90 |
0.048 |
| PaddleOCR-VL-1.5 |
0.9B |
93.43 |
0.037 |
93.04 |
90.97 |
0.045 |
*In the scanning scenario, the models’ overall performance remains high, with PaddleOCR-VL-1.5 typically achieving the best results.*
---
### 2. Warping
| Model Type |
Methods |
Parameters |
Overall↑ |
TextEdit↓ |
FormulaCDM↑ |
TableTEDS↑ |
Reading OrderEdit↓ |
| Pipeline Tools |
Maker-1.8.2 |
- |
58.98 |
0.349 |
72.71 |
39.08 |
0.390 |
| PP-StructureV3 |
- |
59.34 |
0.376 |
68.22 |
47.40 |
0.261 |
| General VLMs |
GPT-5.2 |
- |
76.26 |
0.239 |
80.90 |
71.80 |
0.165 |
| Gemini-2.5 Pro |
- |
87.63 |
0.092 |
86.50 |
85.59 |
0.109 |
| Qwen2.5-VL-72B |
72B |
87.77 |
0.086 |
88.85 |
83.06 |
0.102 |
| Gemini-3 Pro |
- |
88.90 |
0.086 |
88.10 |
87.20 |
0.087 |
| Qwen3-VL-235B-A22B-Instruct |
235B |
89.99 |
0.051 |
89.06 |
85.95 |
0.064 |
| Specialized VLMs |
Dolphin-1.5 |
0.3B |
50.50 |
0.383 |
47.24 |
42.52 |
0.309 |
| Dolphin |
322M |
60.35 |
0.316 |
61.06 |
51.58 |
0.247 |
| Deepseek-OCR |
3B |
67.20 |
0.328 |
73.59 |
60.80 |
0.226 |
| MinerU2-VLM |
0.9B |
73.73 |
0.202 |
77.72 |
63.65 |
0.173 |
| MonkeyOCR-pro-1.2B |
1.9B |
76.59 |
0.196 |
78.85 |
70.52 |
0.221 |
| MonkeyOCR-3B |
3.7B |
77.27 |
0.164 |
79.08 |
69.18 |
0.211 |
| MonkeyOCR-pro-3B |
3.7B |
78.90 |
0.168 |
79.55 |
73.94 |
0.212 |
| Nanonets-OCR-s |
3B |
83.56 |
0.121 |
86.24 |
76.57 |
0.124 |
| MinerU2.5 |
1.2B |
83.76 |
0.154 |
85.92 |
80.71 |
0.104 |
| PaddleOCR-VL |
0.9B |
85.97 |
0.093 |
85.45 |
81.77 |
0.092 |
| dots.ocr |
3B |
86.01 |
0.087 |
85.03 |
81.74 |
0.093 |
| PaddleOCR-VL-1.5 |
0.9B |
91.25 |
0.053 |
90.94 |
88.10 |
0.063 |
*Warping documents introduce moderate challenges, causing a slight performance drop, but specialized VLMs maintain competitive accuracy.*
---
### 3. Screen-Photography
| Model Type |
Methods |
Parameters |
Overall↑ |
TextEdit↓ |
FormulaCDM↑ |
TableTEDS↑ |
Reading OrderEdit↓ |
| Pipeline Tools |
Maker-1.8.2 |
- |
63.65 |
0.290 |
72.73 |
47.21 |
0.325 |
| PP-StructureV3 |
- |
66.89 |
0.204 |
73.26 |
47.82 |
0.165 |
| General VLMs |
GPT-5.2 |
- |
76.75 |
0.208 |
79.27 |
71.73 |
0.148 |
| Qwen2.5-VL-72B |
72B |
86.48 |
0.100 |
87.46 |
82.00 |
0.102 |
| Gemini-2.5 Pro |
- |
87.11 |
0.103 |
85.30 |
86.31 |
0.117 |
| Gemini-3 Pro |
- |
88.86 |
0.084 |
87.33 |
87.65 |
0.087 |
| Qwen3-VL-235B-A22B-Instruct |
235B |
89.27 |
0.068 |
88.72 |
85.85 |
0.071 |
| Specialized VLMs |
Dolphin |
322M |
64.29 |
0.232 |
58.66 |
57.38 |
0.195 |
| Dolphin-1.5 |
0.3B |
69.76 |
0.205 |
61.80 |
68.00 |
0.177 |
| Deepseek-OCR |
3B |
75.31 |
0.220 |
77.68 |
70.26 |
0.169 |
| MinerU2-VLM |
0.9B |
78.77 |
0.139 |
79.02 |
71.17 |
0.123 |
| MonkeyOCR-pro-1.2B |
1.9B |
80.24 |
0.148 |
80.78 |
74.74 |
0.179 |
| MonkeyOCR-3B |
3.7B |
80.71 |
0.122 |
81.33 |
73.04 |
0.177 |
| MonkeyOCR-pro-3B |
3.7B |
82.44 |
0.124 |
81.55 |
78.13 |
0.177 |
| PaddleOCR-VL |
0.9B |
82.54 |
0.103 |
83.58 |
74.36 |
0.107 |
| Nanonets-OCR-s |
3B |
84.86 |
0.112 |
86.65 |
79.09 |
0.117 |
| dots.ocr |
3B |
87.18 |
0.081 |
85.34 |
84.26 |
0.079 |
| MinerU2.5 |
1.2B |
89.41 |
0.062 |
87.55 |
86.83 |
0.053 |
| PaddleOCR-VL-1.5 |
0.9B |
91.76 |
0.050 |
90.88 |
89.38 |
0.059 |
### 4. Illumination
| Model Type |
Methods |
Parameters |
Overall↑ |
TextEdit↓ |
FormulaCDM↑ |
TableTEDS↑ |
Reading OrderEdit↓ |
| Pipeline Tools |
Maker-1.8.2 |
- |
66.31 |
0.259 |
74.80 |
50.03 |
0.337 |
| PP-StructureV3 |
- |
73.38 |
0.158 |
77.75 |
58.19 |
0.126 |
| General VLMs |
GPT-5.2 |
- |
80.88 |
0.191 |
84.41 |
77.37 |
0.134 |
| Qwen2.5-VL-72B |
72B |
87.25 |
0.087 |
86.44 |
84.03 |
0.097 |
| Gemini-2.5 Pro |
- |
87.97 |
0.083 |
86.13 |
86.11 |
0.103 |
| Qwen3-VL-235B-A22B-Instruct |
235B |
89.27 |
0.060 |
87.81 |
86.05 |
0.070 |
| Gemini-3 Pro |
- |
89.53 |
0.073 |
87.78 |
88.14 |
0.080 |
| Specialized VLMs |
Dolphin |
322M |
67.29 |
0.197 |
61.42 |
60.10 |
0.173 |
| Dolphin-1.5 |
0.3B |
75.61 |
0.159 |
70.04 |
72.69 |
0.133 |
| Deepseek-OCR |
3B |
78.10 |
0.192 |
81.71 |
71.81 |
0.156 |
| MinerU2-VLM |
0.9B |
80.51 |
0.135 |
80.72 |
74.29 |
0.123 |
| MonkeyOCR-pro-1.2B |
1.9B |
82.11 |
0.144 |
82.07 |
78.67 |
0.172 |
| MonkeyOCR-3B |
3.7B |
83.16 |
0.118 |
83.63 |
77.62 |
0.168 |
| MonkeyOCR-pro-3B |
3.7B |
84.71 |
0.120 |
84.13 |
82.02 |
0.171 |
| Nanonets-OCR-s |
3B |
85.01 |
0.099 |
87.94 |
76.96 |
0.112 |
| dots.ocr |
3B |
87.57 |
0.068 |
85.07 |
84.44 |
0.076 |
| MinerU2.5 |
1.2B |
89.57 |
0.065 |
88.36 |
86.87 |
0.062 |
| PaddleOCR-VL |
0.9B |
89.61 |
0.049 |
86.66 |
87.02 |
0.055 |
| PaddleOCR-VL-1.5 |
0.9B |
92.16 |
0.046 |
91.80 |
89.33 |
0.051 |
*Severe illumination variations present significant challenges, resulting in a noticeable performance drop for most models.*
---
### 5. Skew
| Model Type |
Methods |
Parameters |
Overall↑ |
TextEdit↓ |
FormulaCDM↑ |
TableTEDS↑ |
Reading OrderEdit↓ |
| Pipeline Tools |
PP-StructureV3 |
- |
37.98 |
0.557 |
44.37 |
25.27 |
0.417 |
| Maker-1.8.2 |
- |
41.27 |
0.536 |
60.16 |
17.23 |
0.543 |
| General VLMs |
GPT-5.2 |
- |
75.00 |
0.257 |
80.27 |
70.47 |
0.167 |
| Qwen3-VL-235B-A22B-Instruct |
235B |
86.56 |
0.077 |
83.96 |
83.41 |
0.091 |
| Qwen2.5-VL-72B |
72B |
86.90 |
0.077 |
87.26 |
81.14 |
0.091 |
| Gemini-2.5 Pro |
- |
89.07 |
0.077 |
87.89 |
86.99 |
0.104 |
| Gemini-3 Pro |
- |
89.45 |
0.080 |
88.33 |
88.06 |
0.092 |
| Specialized VLMs |
Dolphin-1.5 |
0.3B |
28.16 |
0.553 |
25.60 |
14.18 |
0.419 |
| Dolphin |
322M |
44.83 |
0.500 |
51.34 |
33.22 |
0.321 |
| MonkeyOCR-pro-1.2B |
1.9B |
62.18 |
0.292 |
66.25 |
49.46 |
0.317 |
| Deepseek-OCR |
3B |
63.01 |
0.327 |
73.27 |
48.48 |
0.231 |
| MonkeyOCR-pro-3B |
3.7B |
64.47 |
0.251 |
69.06 |
49.42 |
0.301 |
| MonkeyOCR-3B |
3.7B |
65.67 |
0.248 |
69.23 |
52.59 |
0.300 |
| MinerU2-VLM |
0.9B |
68.16 |
0.230 |
74.45 |
53.07 |
0.191 |
| MinerU2.5 |
1.2B |
75.24 |
0.305 |
81.78 |
74.39 |
0.151 |
| PaddleOCR-VL |
0.9B |
77.47 |
0.192 |
78.81 |
72.83 |
0.193 |
| Nanonets-OCR-s |
3B |
81.98 |
0.121 |
85.78 |
72.22 |
0.133 |
| dots.ocr |
3B |
84.27 |
0.087 |
85.73 |
75.74 |
0.094 |
| PaddleOCR-VL-1.5 |
0.9B |
91.66 |
0.047 |
91.00 |
88.69 |
0.061 |
*Skewed documents also impact accuracy, but top-performing models still provide reasonable results.*
---
## Acknowledgements
Real5-OmniDocBench is based on the original OmniDocBench dataset and adopts its evaluation protocols. We thank the authors of [OmniDocBench](https://github.com/opendatalab/OmniDocBench) for their foundational work.
---
## Citation
If you use Real5-OmniDocBench in your research, please cite our dataset and refer also to the original OmniDocBench paper.