metadata
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: validation
path: data/validation-*
- split: test
path: data/test-*
dataset_info:
features:
- name: id
dtype: string
- name: image
dtype: image
- name: output
dtype: string
splits:
- name: train
num_bytes: 2440808481
num_examples: 22435
- name: validation
num_bytes: 329337240
num_examples: 2804
- name: test
num_bytes: 328649745
num_examples: 2805
download_size: 3127169673
dataset_size: 3098795466
Arabic-Image2Html Dataset
A dataset of 28K image-HTML pairs for training OCR models to transform Arabic documents into structured and semantic HTML.
Dataset Description
This dataset was created to address the lack of available open-source Arabic OCR datasets with image-to-semantic HTML pairs. It contains diverse Arabic document images paired with clean, semantic HTML output.
Dataset Composition
The dataset consists of two main components:
1. Web-Scraped Wikipedia Content (~13K samples, 46%)
- Extracted from Arabic Wikipedia articles
- Post-processed HTML with only semantic tags preserved
- Screenshots captured with real styling using Playwright
- Cleaned structure with proper semantic elements (section, header, main, etc.)
2. Synthetically Generated Documents (~15K samples, 54%)
- HTML documents rendered into images using CSS styling
- Mimics various real-world document types:
- Historical manuscripts
- Newspaper articles
- Scientific papers
- Invoices
- Recipes
- And more (~13 formats total)
- Diverse layouts, styles, noise levels, fonts, and text flows
- Filled with plain Arabic text from open datasets
- Multiple semantic tag combinations (footer, table, section, etc.)
Features
- Total samples: 28,000 image-HTML pairs
- Language: Arabic
- Output format: Semantic HTML (clean tags without id, class attributes)
- Document diversity: Multiple formats and layouts
Usage
Loading the Dataset
from datasets import load_dataset
dataset = load_dataset("OussamaBenSlama/arabic-image2html")
Limitations
- Limited examples with diacritical marks (tashkeel), which may affect model performance on texts with extensive diacritics
- Wikipedia samples share similar design patterns
- Synthetic generation may not capture all real-world document variations
Related Resources
- Model: Alef-OCR-Image2Html
- Training Notebooks: Github Repository
Citation
@misc{arabic_image2html_2025,
title={Arabic-Image2Html: A Dataset for Arabic OCR to Semantic HTML},
author={Oussama Ben Slama},
year={2025},
howpublished={Hugging Face Datasets},
url={https://huggingface.co/datasets/OussamaBenSlama/arabic-image2html}
}
License
Apache2.0
Acknowledgments
This work builds upon the excellent research by the NAMAA community and their state-of-the-art Qari-OCR model.