ocr / CLAUDE.md
davanstrien's picture
davanstrien HF Staff
Update nanonets-ocr.py HF Jobs syntax and add smoke test notes
0bc9b0a
# OCR Scripts - Development Notes
## Active Scripts
### DeepSeek-OCR v1 (`deepseek-ocr-vllm.py`)
**Production Ready**
- Fully supported by vLLM
- Fast batch processing
- Tested and working on HF Jobs
### LightOnOCR-2-1B (`lighton-ocr2.py`)
**Production Ready** (Fixed 2026-01-29)
**Status:** Working with vLLM nightly
**What was fixed:**
- Root cause was NOT vLLM - it was the deprecated `HF_HUB_ENABLE_HF_TRANSFER=1` env var
- The script was setting this env var but `hf_transfer` package no longer exists
- This caused download failures that manifested as "Can't load image processor" errors
- Fix: Removed the `HF_HUB_ENABLE_HF_TRANSFER=1` setting from the script
**Test results (2026-01-29):**
- 10/10 samples processed successfully
- Clean markdown output with proper headers and paragraphs
- Output dataset: `davanstrien/lighton-ocr2-test-v4`
**Example usage:**
```bash
hf jobs uv run --flavor a100-large \
-s HF_TOKEN \
https://huggingface.co/datasets/uv-scripts/ocr/raw/main/lighton-ocr2.py \
davanstrien/ufo-ColPali output-dataset \
--max-samples 10 --shuffle --seed 42
```
**Model Info:**
- Model: `lightonai/LightOnOCR-2-1B`
- Architecture: Pixtral ViT encoder + Qwen3 LLM
- Training: RLVR (Reinforcement Learning with Verifiable Rewards)
- Performance: 83.2% on OlmOCR-Bench, 42.8 pages/sec on H100
### PaddleOCR-VL-1.5 (`paddleocr-vl-1.5.py`)
**Production Ready** (Added 2026-01-30)
**Status:** Working with transformers
**Note:** Uses transformers backend (not vLLM) because PaddleOCR-VL only supports vLLM in server mode, which doesn't fit the single-command UV script pattern. Images are processed one at a time for stability.
**Test results (2026-01-30):**
- 10/10 samples processed successfully
- Processing time: ~50s per image on L4 GPU
- Output dataset: `davanstrien/paddleocr-vl15-final-test`
**Example usage:**
```bash
hf jobs uv run --flavor l4x1 \
-s HF_TOKEN \
https://huggingface.co/datasets/uv-scripts/ocr/raw/main/paddleocr-vl-1.5.py \
davanstrien/ufo-ColPali output-dataset \
--max-samples 10 --shuffle --seed 42
```
**Task modes:**
- `ocr` (default): General text extraction to markdown
- `table`: Table extraction to HTML format
- `formula`: Mathematical formula recognition to LaTeX
- `chart`: Chart and diagram analysis
- `spotting`: Text spotting with localization (uses higher resolution)
- `seal`: Seal and stamp recognition
**Model Info:**
- Model: `PaddlePaddle/PaddleOCR-VL-1.5`
- Size: 0.9B parameters (ultra-compact)
- Performance: 94.5% SOTA on OmniDocBench v1.5
- Backend: Transformers (single image processing)
- Requires: `transformers>=5.0.0`
## Pending Development
### DeepSeek-OCR-2 (Visual Causal Flow Architecture)
**Status:** ⏳ Waiting for vLLM upstream support
**Context:**
DeepSeek-OCR-2 is the next generation OCR model (3B parameters) with Visual Causal Flow architecture offering improved quality. We attempted to create a UV script (`deepseek-ocr2-vllm.py`) but encountered a blocker.
**Blocker:**
vLLM does not yet support `DeepseekOCR2ForCausalLM` architecture in the official release.
**PR to Watch:**
🔗 https://github.com/vllm-project/vllm/pull/33165
This PR adds DeepSeek-OCR-2 support but is currently:
- ⚠️ **Open** (not merged)
- Has unresolved review comments
- Pre-commit checks failing
- Issues: hardcoded parameters, device mismatch bugs, missing error handling
**What's Needed:**
1. PR #33165 needs to be reviewed, fixed, and merged
2. vLLM needs to release a version including the merge
3. Then we can add these dependencies to our script:
```python
# dependencies = [
# "datasets>=4.0.0",
# "huggingface-hub",
# "pillow",
# "vllm",
# "tqdm",
# "toolz",
# "torch",
# "addict",
# "matplotlib",
# ]
```
**Implementation Progress:**
- ✅ Created `deepseek-ocr2-vllm.py` script
- ✅ Fixed dependency issues (pyarrow, datasets>=4.0.0)
- ✅ Tested script structure on HF Jobs
- ❌ Blocked: vLLM doesn't recognize architecture
**Partial Implementation:**
The file `deepseek-ocr2-vllm.py` exists in this repo but is **not functional** until vLLM support lands. Consider it a draft.
**Testing Evidence:**
When we ran on HF Jobs, we got:
```
ValidationError: Model architectures ['DeepseekOCR2ForCausalLM'] are not supported for now.
Supported architectures: [...'DeepseekOCRForCausalLM'...]
```
**Next Steps (when PR merges):**
1. Update `deepseek-ocr2-vllm.py` dependencies to include `addict` and `matplotlib`
2. Test on HF Jobs with small dataset (10 samples)
3. Verify output quality
4. Update README.md with DeepSeek-OCR-2 section
5. Document v1 vs v2 differences
**Alternative Approaches (if urgent):**
- Create transformers-based script (slower, no vLLM batching)
- Use DeepSeek's official repo setup (complex, not UV-script compatible)
**Model Information:**
- Model ID: `deepseek-ai/DeepSeek-OCR-2`
- Model Card: https://huggingface.co/deepseek-ai/DeepSeek-OCR-2
- GitHub: https://github.com/deepseek-ai/DeepSeek-OCR-2
- Parameters: 3B
- Resolution: (0-6)×768×768 + 1×1024×1024 patches
- Key improvement: Visual Causal Flow architecture
**Resolution Modes (for v2):**
```python
RESOLUTION_MODES = {
"tiny": {"base_size": 512, "image_size": 512, "crop_mode": False},
"small": {"base_size": 640, "image_size": 640, "crop_mode": False},
"base": {"base_size": 1024, "image_size": 768, "crop_mode": False}, # v2 optimized
"large": {"base_size": 1280, "image_size": 1024, "crop_mode": False},
"gundam": {"base_size": 1024, "image_size": 768, "crop_mode": True}, # v2 optimized
}
```
## Other OCR Scripts
### Nanonets OCR (`nanonets-ocr.py`, `nanonets-ocr2.py`)
✅ Both versions working
### PaddleOCR-VL (`paddleocr-vl.py`)
✅ Working
---
## Future: OCR Smoke Test Dataset
**Status:** Idea (noted 2026-02-12)
Build a small curated dataset (`uv-scripts/ocr-smoke-test`?) with ~2-5 samples from diverse sources. Purpose: fast CI-style verification that scripts still work after dep updates, without downloading full datasets.
**Design goals:**
- Tiny (~20-30 images total) so download is seconds not minutes
- Covers the axes that break things: document type, image quality, language, layout complexity
- Has ground truth text where possible for quality regression checks
- All permissively licensed (CC0/CC-BY preferred)
**Candidate sources:**
| Source | What it covers | Why |
|--------|---------------|-----|
| `NationalLibraryOfScotland/medical-history-of-british-india` | Historical English, degraded scans | Has hand-corrected `text` column for comparison. CC0. Already tested with GLM-OCR. |
| `davanstrien/ufo-ColPali` | Mixed modern documents | Already used as our go-to test set. Varied layouts. |
| Something with **tables** | Structured data extraction | Tests `--task table` modes. Maybe a financial report or census page. |
| Something with **formulas/LaTeX** | Math notation | Tests `--task formula`. arXiv pages or textbook scans. |
| Something **multilingual** (CJK, Arabic, etc.) | Non-Latin scripts | GLM-OCR claims zh/ja/ko support. Good to verify. |
| Something **handwritten** | Handwriting recognition | Edge case that reveals model limits. |
**How it would work:**
```bash
# Quick smoke test for any script
uv run glm-ocr.py uv-scripts/ocr-smoke-test smoke-out --max-samples 5
# Or a dedicated test runner that checks all scripts against it
```
**Open questions:**
- Build as a proper HF dataset, or just a folder of images in the repo?
- Should we include expected output for regression testing (fragile if models change)?
- Could we add a `--smoke-test` flag to each script that auto-uses this dataset?
- Worth adding to HF Jobs scheduled runs for ongoing monitoring?
---
**Last Updated:** 2026-02-12
**Watch PRs:**
- DeepSeek-OCR-2: https://github.com/vllm-project/vllm/pull/33165