The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
Error code: DatasetGenerationError
Exception: FileNotFoundError
Message: PaddlePaddle/Real5-OmniDocBench@0403ef7775a7f17209dd49a8bf508afc56345110/Real5-OmniDocBench-Illumination/PPT_1001115_eng_page_003.png (repository not found)
Traceback: Traceback (most recent call last):
File "/usr/local/lib/python3.12/site-packages/huggingface_hub/utils/_http.py", line 409, in hf_raise_for_status
response.raise_for_status()
File "/usr/local/lib/python3.12/site-packages/requests/models.py", line 1026, in raise_for_status
raise HTTPError(http_error_msg, response=self)
requests.exceptions.HTTPError: 401 Client Error: Unauthorized for url: https://huggingface.co/api/datasets/PaddlePaddle/Real5-OmniDocBench/revision/0403ef7775a7f17209dd49a8bf508afc56345110
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/local/lib/python3.12/site-packages/huggingface_hub/hf_file_system.py", line 125, in _repo_and_revision_exist
self._api.repo_info(
File "/usr/local/lib/python3.12/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn
return fn(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/huggingface_hub/hf_api.py", line 2816, in repo_info
return method(
^^^^^^^
File "/usr/local/lib/python3.12/site-packages/huggingface_hub/utils/_validators.py", line 114, in _inner_fn
return fn(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/huggingface_hub/hf_api.py", line 2674, in dataset_info
hf_raise_for_status(r)
File "/usr/local/lib/python3.12/site-packages/huggingface_hub/utils/_http.py", line 459, in hf_raise_for_status
raise _format(RepositoryNotFoundError, message, response) from e
huggingface_hub.errors.RepositoryNotFoundError: 401 Client Error. (Request ID: Root=1-697ab3af-338d2ec162feb7312a1be589;366d4c80-e92c-4291-bd7f-be339376b25a)
Repository Not Found for url: https://huggingface.co/api/datasets/PaddlePaddle/Real5-OmniDocBench/revision/0403ef7775a7f17209dd49a8bf508afc56345110.
Please make sure you specified the correct `repo_id` and `repo_type`.
If you are trying to access a private or gated repo, make sure you are authenticated. For more details, see https://huggingface.co/docs/huggingface_hub/authentication
Invalid username or password.
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 1608, in _prepare_split_single
num_examples, num_bytes = writer.finalize()
^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/arrow_writer.py", line 684, in finalize
self.write_examples_on_file()
File "/usr/local/lib/python3.12/site-packages/datasets/arrow_writer.py", line 571, in write_examples_on_file
self.write_batch(batch_examples=batch_examples)
File "/usr/local/lib/python3.12/site-packages/datasets/arrow_writer.py", line 661, in write_batch
self.write_table(pa_table, writer_batch_size)
File "/usr/local/lib/python3.12/site-packages/datasets/arrow_writer.py", line 676, in write_table
pa_table = embed_table_storage(pa_table)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 2249, in embed_table_storage
embed_array_storage(table[name], feature, token_per_repo_id=token_per_repo_id)
File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 1795, in wrapper
return pa.chunked_array([func(chunk, *args, **kwargs) for chunk in array.chunks])
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/table.py", line 2124, in embed_array_storage
return feature.embed_storage(array, token_per_repo_id=token_per_repo_id)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/features/image.py", line 303, in embed_storage
(path_to_bytes(x["path"]) if x["bytes"] is None else x["bytes"]) if x is not None else None
^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/utils/py_utils.py", line 310, in wrapper
return func(value) if value is not None else None
^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/features/image.py", line 298, in path_to_bytes
with xopen(path, "rb", download_config=download_config) as f:
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/utils/file_utils.py", line 977, in xopen
file_obj = fsspec.open(file, mode=mode, *args, **kwargs).open()
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/fsspec/core.py", line 135, in open
return self.__enter__()
^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/fsspec/core.py", line 103, in __enter__
f = self.fs.open(self.path, mode=mode)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "<string>", line 3, in open
File "/usr/local/lib/python3.12/unittest/mock.py", line 1139, in __call__
return self._mock_call(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/unittest/mock.py", line 1143, in _mock_call
return self._execute_mock_call(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/unittest/mock.py", line 1204, in _execute_mock_call
result = effect(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^
File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 770, in wrapped
f = fs_open(self, urlpath, mode, *args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/fsspec/spec.py", line 1293, in open
f = self._open(
^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/huggingface_hub/hf_file_system.py", line 275, in _open
return HfFileSystemFile(self, path, mode=mode, revision=revision, block_size=block_size, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/huggingface_hub/hf_file_system.py", line 938, in __init__
self.resolved_path = fs.resolve_path(path, revision=revision)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/huggingface_hub/hf_file_system.py", line 200, in resolve_path
_raise_file_not_found(path, err)
File "/usr/local/lib/python3.12/site-packages/huggingface_hub/hf_file_system.py", line 1138, in _raise_file_not_found
raise FileNotFoundError(msg) from err
FileNotFoundError: PaddlePaddle/Real5-OmniDocBench@0403ef7775a7f17209dd49a8bf508afc56345110/Real5-OmniDocBench-Illumination/PPT_1001115_eng_page_003.png (repository not found)
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 1342, in compute_config_parquet_and_info_response
parquet_operations, partial, estimated_dataset_info = stream_convert_to_parquet(
^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/src/services/worker/src/worker/job_runners/config/parquet_and_info.py", line 907, in stream_convert_to_parquet
builder._prepare_split(split_generator=splits_generators[split], file_format="parquet")
File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 1438, in _prepare_split
for job_id, done, content in self._prepare_split_single(
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/builder.py", line 1617, in _prepare_split_single
raise DatasetGenerationError("An error occurred while generating the dataset") from e
datasets.exceptions.DatasetGenerationError: An error occurred while generating the datasetNeed help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
image
image | label
class label |
|---|---|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
|
0Real5-OmniDocBench-Illumination
|
Real5-OmniDocBench
Real5-OmniDocBench is a brand-new benchmark oriented toward real-world scenarios, which we constructed based on the OmniDocBench v1.5 dataset. The dataset comprises five distinct scenarios: Scanning, Warping, Screen-Photography, Illumination, and Skew. Apart from the Scanning category, all images were manually acquired via handheld mobile devices to closely simulate real-world conditions. Each subset maintains a one-to-one correspondence with the original OmniDocBench, strictly adhering to its ground-truth annotations and evaluation protocols. Given its empirical and realistic nature, this dataset serves as a rigorous benchmark for assessing the robustness of document parsing models in practical applications.
Key Features
1. Real-world Scenarios
Real5-OmniDocBench contains five challenging and representative photographic scenarios:
- Scanning: Images captured by scanning devices, simulating flat and clean document scans.
- Warping: Photos of documents with visible page curvatures, mimicking distortions caused by book spines or folding.
- Skew: Documents photographed at an angle, resulting in perspective distortion.
- Screen-Photography: Photographs of screens displaying documents, introducing moiré patterns and reflections.
- Illumination: Images taken under varied lighting conditions, including shadows and glare.
2. Comprehensive Coverage
For each scenario, the dataset contains 1,355 images, corresponding exactly to the original pages in the OmniDocBench evaluation set. This enables direct, controlled comparison of model performance across different real-world conditions.
3. High-Quality Realism
All images are captured manually under real-world conditions. Care was taken to cover typical distortions and artifacts encountered in mobile or camera-based document digitization.
4. Evaluation Protocol Compatibility
The evaluation scripts and metrics from OmniDocBench (e.g., Normalized Edit Distance, BLEU, METEOR, TEDS, COCODet) are fully compatible and directly applicable to Real5-OmniDocBench.
Dataset Structure
The dataset follows the directory structure shown below:
Real5-OmniDocBench/
├── Real5-OmniDocBench-Warping/
│ ├── book_en_[搬书匠#20][HTML5 Canvas].2011.英文版_page_208.png
│ └── ...
├── Real5-OmniDocBench-Illumination/
│ └── ...
├── Real5-OmniDocBench-Scanning/
│ └── ...
├── Real5-OmniDocBench-Screen-Photography/
│ └── ...
└── Real5-OmniDocBench-Skew/
└── ...
Usage & Evaluation
- Input: For each scenario, models should take the corresponding set of 1,355 images as input.
- Output: Model predictions should follow the same format as OmniDocBench, e.g., Markdown for end-to-end parsing.
- Evaluation: Use the official OmniDocBench evaluation scripts and metrics for assessment. Direct comparison across scenarios is encouraged to measure model robustness.
Benchmark Results
To illustrate the impact of real-world visual distortions on document parsing, we evaluated several mainstream models across all five scenarios in Real5-OmniDocBench. The same metrics as OmniDocBench are used: Overall↑, TextEdit↓, FormulaCDM↑, TableTEDS↑, and Reading OrderEdit↓.
1. Scanning
| Model Type | Methods | Parameters | Overall↑ | TextEdit↓ | FormulaCDM↑ | TableTEDS↑ | Reading OrderEdit↓ |
|---|---|---|---|---|---|---|---|
| Pipeline Tools | Maker-1.8.2 | - | 70.27 | 0.223 | 77.03 | 56.05 | 0.238 |
| PP-StructureV3 | - | 84.68 | 0.094 | 84.34 | 79.06 | 0.092 | |
| General VLMs | GPT-5.2 | - | 84.43 | 0.142 | 85.68 | 81.78 | 0.109 |
| Qwen2.5-VL-72B | 72B | 86.19 | 0.110 | 86.14 | 83.41 | 0.114 | |
| Gemini-2.5 Pro | - | 89.25 | 0.073 | 87.44 | 87.62 | 0.098 | |
| Qwen3-VL-235B-A22B-Instruct | 235B | 89.43 | 0.059 | 89.01 | 85.19 | 0.066 | |
| Gemini-3 Pro | - | 89.47 | 0.071 | 88.16 | 87.37 | 0.078 | |
| Specialized VLMs | Dolphin | 322M | 72.16 | 0.154 | 64.58 | 67.27 | 0.130 |
| Dolphin-1.5 | 0.3B | 83.39 | 0.097 | 76.25 | 83.65 | 0.090 | |
| MinerU2-VLM | 0.9B | 83.60 | 0.094 | 79.76 | 80.44 | 0.091 | |
| MonkeyOCR-pro-1.2B | 1.9B | 84.64 | 0.123 | 84.17 | 82.13 | 0.145 | |
| MonkeyOCR-3B | 3.7B | 84.65 | 0.100 | 84.16 | 79.81 | 0.143 | |
| Nanonets-OCR-s | 3B | 85.52 | 0.106 | 88.09 | 79.11 | 0.106 | |
| Deepseek-OCR | 3B | 86.17 | 0.078 | 83.59 | 82.69 | 0.085 | |
| dots.ocr | 3B | 86.87 | 0.083 | 83.27 | 85.68 | 0.081 | |
| MonkeyOCR-pro-3B | 3.7B | 86.94 | 0.103 | 86.29 | 84.86 | 0.141 | |
| MinerU2.5 | 1.2B | 90.06 | 0.052 | 88.22 | 87.16 | 0.050 | |
| PaddleOCR-VL | 0.9B | 92.11 | 0.039 | 90.35 | 89.90 | 0.048 | |
| PaddleOCR-VL-1.5 | 0.9B | 93.43 | 0.037 | 93.04 | 90.97 | 0.045 |
In the scanning scenario, the models’ overall performance remains high, with PaddleOCR-VL-1.5 typically achieving the best results.
2. Warping
| Model Type | Methods | Parameters | Overall↑ | TextEdit↓ | FormulaCDM↑ | TableTEDS↑ | Reading OrderEdit↓ |
|---|---|---|---|---|---|---|---|
| Pipeline Tools | Maker-1.8.2 | - | 58.98 | 0.349 | 72.71 | 39.08 | 0.390 |
| PP-StructureV3 | - | 59.34 | 0.376 | 68.22 | 47.40 | 0.261 | |
| General VLMs | GPT-5.2 | - | 76.26 | 0.239 | 80.90 | 71.80 | 0.165 |
| Gemini-2.5 Pro | - | 87.63 | 0.092 | 86.50 | 85.59 | 0.109 | |
| Qwen2.5-VL-72B | 72B | 87.77 | 0.086 | 88.85 | 83.06 | 0.102 | |
| Gemini-3 Pro | - | 88.90 | 0.086 | 88.10 | 87.20 | 0.087 | |
| Qwen3-VL-235B-A22B-Instruct | 235B | 89.99 | 0.051 | 89.06 | 85.95 | 0.064 | |
| Specialized VLMs | Dolphin-1.5 | 0.3B | 50.50 | 0.383 | 47.24 | 42.52 | 0.309 |
| Dolphin | 322M | 60.35 | 0.316 | 61.06 | 51.58 | 0.247 | |
| Deepseek-OCR | 3B | 67.20 | 0.328 | 73.59 | 60.80 | 0.226 | |
| MinerU2-VLM | 0.9B | 73.73 | 0.202 | 77.72 | 63.65 | 0.173 | |
| MonkeyOCR-pro-1.2B | 1.9B | 76.59 | 0.196 | 78.85 | 70.52 | 0.221 | |
| MonkeyOCR-3B | 3.7B | 77.27 | 0.164 | 79.08 | 69.18 | 0.211 | |
| MonkeyOCR-pro-3B | 3.7B | 78.90 | 0.168 | 79.55 | 73.94 | 0.212 | |
| Nanonets-OCR-s | 3B | 83.56 | 0.121 | 86.24 | 76.57 | 0.124 | |
| MinerU2.5 | 1.2B | 83.76 | 0.154 | 85.92 | 80.71 | 0.104 | |
| PaddleOCR-VL | 0.9B | 85.97 | 0.093 | 85.45 | 81.77 | 0.092 | |
| dots.ocr | 3B | 86.01 | 0.087 | 85.03 | 81.74 | 0.093 | |
| PaddleOCR-VL-1.5 | 0.9B | 91.25 | 0.053 | 90.94 | 88.10 | 0.063 |
Warping documents introduce moderate challenges, causing a slight performance drop, but specialized VLMs maintain competitive accuracy.
3. Screen-Photography
| Model Type | Methods | Parameters | Overall↑ | TextEdit↓ | FormulaCDM↑ | TableTEDS↑ | Reading OrderEdit↓ |
|---|---|---|---|---|---|---|---|
| Pipeline Tools | Maker-1.8.2 | - | 63.65 | 0.290 | 72.73 | 47.21 | 0.325 |
| PP-StructureV3 | - | 66.89 | 0.204 | 73.26 | 47.82 | 0.165 | |
| General VLMs | GPT-5.2 | - | 76.75 | 0.208 | 79.27 | 71.73 | 0.148 |
| Qwen2.5-VL-72B | 72B | 86.48 | 0.100 | 87.46 | 82.00 | 0.102 | |
| Gemini-2.5 Pro | - | 87.11 | 0.103 | 85.30 | 86.31 | 0.117 | |
| Gemini-3 Pro | - | 88.86 | 0.084 | 87.33 | 87.65 | 0.087 | |
| Qwen3-VL-235B-A22B-Instruct | 235B | 89.27 | 0.068 | 88.72 | 85.85 | 0.071 | |
| Specialized VLMs | Dolphin | 322M | 64.29 | 0.232 | 58.66 | 57.38 | 0.195 |
| Dolphin-1.5 | 0.3B | 69.76 | 0.205 | 61.80 | 68.00 | 0.177 | |
| Deepseek-OCR | 3B | 75.31 | 0.220 | 77.68 | 70.26 | 0.169 | |
| MinerU2-VLM | 0.9B | 78.77 | 0.139 | 79.02 | 71.17 | 0.123 | |
| MonkeyOCR-pro-1.2B | 1.9B | 80.24 | 0.148 | 80.78 | 74.74 | 0.179 | |
| MonkeyOCR-3B | 3.7B | 80.71 | 0.122 | 81.33 | 73.04 | 0.177 | |
| MonkeyOCR-pro-3B | 3.7B | 82.44 | 0.124 | 81.55 | 78.13 | 0.177 | |
| PaddleOCR-VL | 0.9B | 82.54 | 0.103 | 83.58 | 74.36 | 0.107 | |
| Nanonets-OCR-s | 3B | 84.86 | 0.112 | 86.65 | 79.09 | 0.117 | |
| dots.ocr | 3B | 87.18 | 0.081 | 85.34 | 84.26 | 0.079 | |
| MinerU2.5 | 1.2B | 89.41 | 0.062 | 87.55 | 86.83 | 0.053 | |
| PaddleOCR-VL-1.5 | 0.9B | 91.76 | 0.050 | 90.88 | 89.38 | 0.059 |
4. Illumination
| Model Type | Methods | Parameters | Overall↑ | TextEdit↓ | FormulaCDM↑ | TableTEDS↑ | Reading OrderEdit↓ |
|---|---|---|---|---|---|---|---|
| Pipeline Tools | Maker-1.8.2 | - | 66.31 | 0.259 | 74.80 | 50.03 | 0.337 |
| PP-StructureV3 | - | 73.38 | 0.158 | 77.75 | 58.19 | 0.126 | |
| General VLMs | GPT-5.2 | - | 80.88 | 0.191 | 84.41 | 77.37 | 0.134 |
| Qwen2.5-VL-72B | 72B | 87.25 | 0.087 | 86.44 | 84.03 | 0.097 | |
| Gemini-2.5 Pro | - | 87.97 | 0.083 | 86.13 | 86.11 | 0.103 | |
| Qwen3-VL-235B-A22B-Instruct | 235B | 89.27 | 0.060 | 87.81 | 86.05 | 0.070 | |
| Gemini-3 Pro | - | 89.53 | 0.073 | 87.78 | 88.14 | 0.080 | |
| Specialized VLMs | Dolphin | 322M | 67.29 | 0.197 | 61.42 | 60.10 | 0.173 |
| Dolphin-1.5 | 0.3B | 75.61 | 0.159 | 70.04 | 72.69 | 0.133 | |
| Deepseek-OCR | 3B | 78.10 | 0.192 | 81.71 | 71.81 | 0.156 | |
| MinerU2-VLM | 0.9B | 80.51 | 0.135 | 80.72 | 74.29 | 0.123 | |
| MonkeyOCR-pro-1.2B | 1.9B | 82.11 | 0.144 | 82.07 | 78.67 | 0.172 | |
| MonkeyOCR-3B | 3.7B | 83.16 | 0.118 | 83.63 | 77.62 | 0.168 | |
| MonkeyOCR-pro-3B | 3.7B | 84.71 | 0.120 | 84.13 | 82.02 | 0.171 | |
| Nanonets-OCR-s | 3B | 85.01 | 0.099 | 87.94 | 76.96 | 0.112 | |
| dots.ocr | 3B | 87.57 | 0.068 | 85.07 | 84.44 | 0.076 | |
| MinerU2.5 | 1.2B | 89.57 | 0.065 | 88.36 | 86.87 | 0.062 | |
| PaddleOCR-VL | 0.9B | 89.61 | 0.049 | 86.66 | 87.02 | 0.055 | |
| PaddleOCR-VL-1.5 | 0.9B | 92.16 | 0.046 | 91.80 | 89.33 | 0.051 |
Severe illumination variations present significant challenges, resulting in a noticeable performance drop for most models.
5. Skew
| Model Type | Methods | Parameters | Overall↑ | TextEdit↓ | FormulaCDM↑ | TableTEDS↑ | Reading OrderEdit↓ |
|---|---|---|---|---|---|---|---|
| Pipeline Tools | PP-StructureV3 | - | 37.98 | 0.557 | 44.37 | 25.27 | 0.417 |
| Maker-1.8.2 | - | 41.27 | 0.536 | 60.16 | 17.23 | 0.543 | |
| General VLMs | GPT-5.2 | - | 75.00 | 0.257 | 80.27 | 70.47 | 0.167 |
| Qwen3-VL-235B-A22B-Instruct | 235B | 86.56 | 0.077 | 83.96 | 83.41 | 0.091 | |
| Qwen2.5-VL-72B | 72B | 86.90 | 0.077 | 87.26 | 81.14 | 0.091 | |
| Gemini-2.5 Pro | - | 89.07 | 0.077 | 87.89 | 86.99 | 0.104 | |
| Gemini-3 Pro | - | 89.45 | 0.080 | 88.33 | 88.06 | 0.092 | |
| Specialized VLMs | Dolphin-1.5 | 0.3B | 28.16 | 0.553 | 25.60 | 14.18 | 0.419 |
| Dolphin | 322M | 44.83 | 0.500 | 51.34 | 33.22 | 0.321 | |
| MonkeyOCR-pro-1.2B | 1.9B | 62.18 | 0.292 | 66.25 | 49.46 | 0.317 | |
| Deepseek-OCR | 3B | 63.01 | 0.327 | 73.27 | 48.48 | 0.231 | |
| MonkeyOCR-pro-3B | 3.7B | 64.47 | 0.251 | 69.06 | 49.42 | 0.301 | |
| MonkeyOCR-3B | 3.7B | 65.67 | 0.248 | 69.23 | 52.59 | 0.300 | |
| MinerU2-VLM | 0.9B | 68.16 | 0.230 | 74.45 | 53.07 | 0.191 | |
| MinerU2.5 | 1.2B | 75.24 | 0.305 | 81.78 | 74.39 | 0.151 | |
| PaddleOCR-VL | 0.9B | 77.47 | 0.192 | 78.81 | 72.83 | 0.193 | |
| Nanonets-OCR-s | 3B | 81.98 | 0.121 | 85.78 | 72.22 | 0.133 | |
| dots.ocr | 3B | 84.27 | 0.087 | 85.73 | 75.74 | 0.094 | |
| PaddleOCR-VL-1.5 | 0.9B | 91.66 | 0.047 | 91.00 | 88.69 | 0.061 |
Skewed documents also impact accuracy, but top-performing models still provide reasonable results.
Acknowledgements
Real5-OmniDocBench is based on the original OmniDocBench dataset and adopts its evaluation protocols. We thank the authors of OmniDocBench for their foundational work.
Citation
If you use Real5-OmniDocBench in your research, please cite our dataset and refer also to the original OmniDocBench paper.
- Downloads last month
- -