File size: 5,878 Bytes
f037451
d6a115b
f037451
 
 
 
 
 
 
 
d6a115b
 
 
 
 
640da6e
d6a115b
 
8890d72
d6a115b
8890d72
 
d6a115b
 
 
 
 
f037451
9c9ec3d
 
 
 
 
 
 
 
 
 
7c82cc8
9c9ec3d
 
 
 
 
62c3733
 
9c9ec3d
 
 
 
 
7c82cc8
9c9ec3d
 
 
62c3733
 
9c9ec3d
 
 
 
 
 
 
 
62c3733
 
 
9c9ec3d
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
62c3733
9c9ec3d
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
62c3733
9c9ec3d
62c3733
 
 
 
 
 
9c9ec3d
62c3733
 
 
9c9ec3d
 
 
 
 
 
 
 
 
 
 
 
fd4e492
 
 
 
 
 
 
 
9c9ec3d
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
---
pretty_name: KITAB PDF to Markdown (Reviewed)
language:
- ar
license: apache-2.0
tags:
- ocr
- arabic
- document-understanding
- pdf-to-markdown
dataset_info:
  features:
  - name: markdown
    dtype: string
  - name: image
    dtype: image
  splits:
  - name: train
    num_bytes: 68643604.0
    num_examples: 62
  download_size: 68467976
  dataset_size: 68643604.0
configs:
- config_name: default
  data_files:
  - split: train
    path: data/train-*
---
# KITAB_pdf_to_markdown_reviewed (Corrected KITAB-Bench PDF→Markdown)

**Short description.** A carefully reviewed and corrected version of the KITAB-Bench PDF-to-Markdown subset for **Arabic document OCR** evaluation. We fixed ground-truth errors (hallucinated text, missing page numbers, omissions of small-font text) and standardized formatting to provide a **reliable benchmark** for model comparison.

**TL;DR**
- ✅ Human-verified ground truth for Arabic PDF→Markdown
- ✅ Removes hallucinations and fills missing/omitted content
- ✅ Keeps the original task and schema for drop-in evaluation
- 🔗 [Based on KITAB-Bench](https://github.com/mbzuai-oryx/KITAB-Bench)

---

## Motivation & Background

Evaluating Arabic OCR and document understanding models requires robust, accurate benchmarks. During an assessment of the original **KITAB-Bench** PDF-to-Markdown subset[^kitab], we found problems that compromise fair evaluation:

- **Hallucinated ground truth:** some reference markdown contained phrases not present in the source page (likely tool-generated).
  *Example:* one entry included the English sentence:
  > “**You're right - let me write it exactly as it appears in the image, maintaining the right-to-left direction:**- **Missing page numbers** in references.
- **Omission of small-font text** that is clearly visible in the source image.

To address this, we manually reviewed and corrected the ground truth, producing this dataset.
---

## What’s in this dataset?

- **Split:** `train`
- **Records:** currently ~60+ page-level samples (may grow in future versions).
- **Fields:**
  - `image` — the page image.
  - `markdown` — human-verified, structure-preserving Markdown for the page.

---

## How we corrected the data

1. **Removed hallucinated phrases** that do not appear in the image.
2. **Restored omitted content**, including **small-font text**.
3. **Added/verified page markers** when appropriate.
4. **Normalized minor formatting** to keep the task consistent across samples.

Our goal was **minimal, faithful correction**: keep the original task and layout intent, while ensuring the ground truth actually matches the page.

---

## Usage

```python
from datasets import load_dataset

ds = load_dataset("Misraj/KITAB_pdf_to_markdown_reviewed", split="train")
row = ds[0]

# image preview
row["image"].show()

# markdown preview
print(row["markdown"][:800])
```

---

## Evaluation protocol (suggested)

Commonly reported metrics for this task include:

* **WER / CER** — word/character error rate (↓ better)
* **BLEU / ChrF** — text similarity (↑ better)
* **TEDS** — structural fidelity of tree/HTML/Markdown (↑ better)
* **MARS** — combined structure + text score (↑ better)

> Evaluate text metrics on normalized text; compute TEDS/MARS on rendered trees/blocks to reflect layout/structure preservation.

---

## Example results (on the corrected KITAB-Bench PDF→Markdown)

> Snapshot from our experiments using only open-source models for fairness; best in **bold**, second-best <u>underlined</u>.

| Model             |       WER ↓ |       CER ↓ |      BLEU ↑ |      CHRF ↑ |     TEDS ↑ |      MARS ↑ |
| ----------------- | ----------: | ----------: | -----------: | -----------: | --------: | -----------: |
| Dots.ocr          |    **0.39** |    **0.28** |    **59.28** |    **83.16** |         43 | <u>63.08</u> |
| **Baseer (ours)** |        0.61 | <u>0.40</u> | <u>55.78</u> | <u>80.26</u> |     **56** |    **68.13** |
| Nanonets          | <u>0.51</u> | <u>0.40</u> |        51.37 |        77.45 |         33 |       55.225 |
| Qari              |        0.65 |        0.48 |        44.61 |        71.45 |         43 |       57.225 |
| Qwen2.5-VL-3B     |        0.70 |        0.57 |        40.44 |        66.78 |         31 |        48.89 |
| Qwen2.5-VL-7B     |        0.76 |        0.63 |        36.76 |        62.45 |         24 |       43.225 |
| Gemma3-12B        |        0.85 |        0.69 |        27.56 |        52.09 | <u>55</u> |       53.545 |
| Gemma3-4B         |        0.95 |        0.82 |        12.94 |        31.72 |         27 |        29.36 |
| Aya-vision        |        1.27 |        0.96 |         5.58 |        16.19 |         26 |       21.095 |
| AIN               |        1.18 |        1.08 |         2.61 |         3.99 |         24 |       13.995 |

**Reading the snapshot.** Dots.ocr leads most text-centric metrics, while **Baseer** achieves the **best structural** score (TEDS) and **best overall MARS**, reflecting stronger layout understanding. The KITAB-Bench subset is small (tens of pages), so each misprediction impacts the score noticeably. On our larger and more diverse **Misraj-DocOCR** benchmark (400 expert-verified pages), Baseer’s advantage is more pronounced.

---

## How to cite

If you use this dataset, please cite **both** this corrected release and the original KITAB-Bench:

**This dataset (recommended):**

```bibtex
@misc{hennara2025baseervisionlanguagemodelarabic,
      title={Baseer: A Vision-Language Model for Arabic Document-to-Markdown OCR}, 
      author={Khalil Hennara and Muhammad Hreden and Mohamed Motasim Hamed and Ahmad Bastati and Zeina Aldallal and Sara Chrouf and Safwan AlModhayan},
      year={2025},
      eprint={2509.18174},
      archivePrefix={arXiv},
      primaryClass={cs.CV},
      url={https://arxiv.org/abs/2509.18174}, 
}
```