Datasets:
File size: 4,752 Bytes
1f1b32f 40b9671 1f1b32f 40b9671 1f1b32f 40b9671 1f1b32f 40b9671 1f1b32f | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 | ---
license: odc-by
tags:
- ocr
- benchmark
- pdf
- document-understanding
language:
- en
pretty_name: olmOCR-bench Pre-Rendered
size_categories:
- 1K<n<10K
configs:
- config_name: arxiv_math
data_files:
- split: test
path: images/arxiv_math/**
- config_name: headers_footers
data_files:
- split: test
path: images/headers_footers/**
- config_name: long_tiny_text
data_files:
- split: test
path: images/long_tiny_text/**
- config_name: multi_column
data_files:
- split: test
path: images/multi_column/**
- config_name: old_scans
data_files:
- split: test
path: images/old_scans/**
- config_name: old_scans_math
data_files:
- split: test
path: images/old_scans_math/**
- config_name: tables
data_files:
- split: test
path: images/tables/**
---
# olmOCR-bench Pre-Rendered
Pre-rendered PNG images of the [olmOCR-bench](https://huggingface.co/datasets/allenai/olmOCR-bench) benchmark dataset, ready for zero-setup evaluation of any OCR / vision model.
## What This Is
The official [olmOCR benchmark](https://github.com/allenai/olmocr) requires downloading 1,403 PDFs locally and rendering each page to a PNG image before sending it to a model. **Every benchmark runner in the official repo does this same rendering step internally** — see [`olmocr/data/renderpdf.py::render_pdf_to_base64png()`](https://github.com/allenai/olmocr/blob/main/olmocr/data/renderpdf.py).
This dataset eliminates that setup entirely by hosting the pre-rendered images directly. The PNGs are rendered at **`target_longest_image_dim=2048`** — the same default resolution used by the official olmOCR `render_pdf_to_base64png()` function and by the GPT-4o, Claude, and Gemini benchmark runners.
All files are accessible via direct URL, so you can evaluate any model by just pointing at these URLs — no local downloads, no PDF rendering tools, no dataset cloning.
## Dataset Structure
The dataset has **7 subsets** (one per benchmark category), each with a **test** split:
| Subset | PDFs | Tests | Test Types |
|--------|------|-------|------------|
| `arxiv_math` | 522 | 2,927 | math |
| `headers_footers` | 266 | 753 | absent |
| `long_tiny_text` | 62 | 442 | present |
| `multi_column` | 231 | 884 | order |
| `old_scans` | 98 | 526 | present, absent, order |
| `old_scans_math` | 36 | 458 | math |
| `tables` | 188 | 1,020 | table |
### Loading with `datasets`
```python
from datasets import load_dataset
ds = load_dataset("shhdwi/olmocr-pre-rendered", "arxiv_math", split="test")
print(ds[0]) # {'image': <PIL.Image>, 'pdf_stem': '...', 'category': '...', ...}
```
## Contents
| Directory | Contents | Count |
|-----------|----------|-------|
| `images/` | Pre-rendered PNG images (page 1, 2048px longest dim) with `metadata.jsonl` per category | 1,403 images |
| `ground_truth/` | JSONL test case files (from allenai/olmOCR-bench) | 7,010 tests |
| `predictions/` | Published model prediction caches | 1,403 .md files |
## Rendering Details
Each PDF page is rendered to PNG matching the official olmOCR benchmark process:
- **Resolution**: `target_longest_image_dim = 2048` (longest side scaled to 2048px, aspect ratio preserved)
- **Renderer**: PyMuPDF (same pixel output as pdftoppm used in the official repo)
- **Pages**: Page 1 only (the benchmark tests only page 1 of each PDF)
- **Naming**: `{pdf_stem}_pg1.png`
This matches what the official benchmark runners do internally:
- `run_chatgpt.py`: `render_pdf_to_base64png(pdf_path, target_longest_image_dim=2048)`
- `run_claude.py`: `render_pdf_to_base64png(pdf_path, target_longest_image_dim=2048)`
- `run_gemini.py`: `render_pdf_to_base64png(pdf_path, target_longest_image_dim=2048)`
- `run_server.py`: `render_pdf_to_base64png(pdf_path, target_longest_image_dim=1024)` (for smaller models)
## Quick Start
Evaluate any model with zero setup:
```bash
pip install litellm httpx
# Run on any litellm-supported model
python run_bench.py --model gpt-4o
python run_bench.py --model claude-sonnet-4-20250514
python run_bench.py --model gemini/gemini-2.0-flash
# Run specific categories only
python run_bench.py --model gpt-4o --categories arxiv_math headers_footers
# Evaluate published predictions (no API key needed)
python run_bench.py --evaluate nanonets-optimal-v4
```
## Direct File Access
Every file is accessible via URL:
```
https://huggingface.co/datasets/shhdwi/olmocr-pre-rendered/resolve/main/images/arxiv_math/2503.05390_pg14_pg1.png
https://huggingface.co/datasets/shhdwi/olmocr-pre-rendered/resolve/main/ground_truth/arxiv_math.jsonl
```
## Attribution
Based on [olmOCR-bench](https://huggingface.co/datasets/allenai/olmOCR-bench) by Allen AI ([paper](https://huggingface.co/papers/2502.18443)). Licensed ODC-BY-1.0.
|