File size: 2,703 Bytes
3f00040
 
 
 
d2c1469
 
3f00040
 
 
 
 
d2c1469
3f00040
d2c1469
3f00040
 
 
 
d2c1469
3f00040
d2c1469
 
3f00040
 
 
 
 
 
d2c1469
3f00040
8851443
3f00040
 
 
 
d2c1469
 
 
 
 
 
 
3f00040
 
 
 
d2c1469
3f00040
 
ed451ad
 
 
 
 
 
 
d2c1469
ed451ad
 
 
 
 
 
 
 
 
d2c1469
ed451ad
 
3f00040
 
d2c1469
ed451ad
3f00040
d2c1469
 
 
 
3f00040
 
d2c1469
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
---
tags:
- ocr
- document-processing
- deepseek
- deepseek-ocr
- markdown
- uv-script
- generated
---

# Document OCR using DeepSeek-OCR

This dataset contains markdown-formatted OCR results from images in [NationalLibraryOfScotland/medical-history-of-british-india](https://huggingface.co/datasets/NationalLibraryOfScotland/medical-history-of-british-india) using DeepSeek-OCR.

## Processing Details

- **Source Dataset**: [NationalLibraryOfScotland/medical-history-of-british-india](https://huggingface.co/datasets/NationalLibraryOfScotland/medical-history-of-british-india)
- **Model**: [deepseek-ai/DeepSeek-OCR](https://huggingface.co/deepseek-ai/DeepSeek-OCR)
- **Number of Samples**: 50
- **Processing Time**: 13.6 min
- **Processing Date**: 2026-02-14 19:17 UTC

### Configuration

- **Image Column**: `image`
- **Output Column**: `markdown`
- **Dataset Split**: `train`
- **Batch Size**: 8
- **Max Model Length**: 8,192 tokens
- **Max Output Tokens**: 8,192
- **GPU Memory Utilization**: 80.0%

## Model Information

DeepSeek-OCR is a state-of-the-art document OCR model that excels at:
- LaTeX equations - Mathematical formulas preserved in LaTeX format
- Tables - Extracted and formatted as HTML/markdown
- Document structure - Headers, lists, and formatting maintained
- Image grounding - Spatial layout and bounding box information
- Complex layouts - Multi-column and hierarchical structures
- Multilingual - Supports multiple languages

## Dataset Structure

The dataset contains all original columns plus:
- `markdown`: The extracted text in markdown format with preserved structure
- `inference_info`: JSON list tracking all OCR models applied to this dataset

## Usage

```python
from datasets import load_dataset
import json

# Load the dataset
dataset = load_dataset("{{output_dataset_id}}", split="train")

# Access the markdown text
for example in dataset:
    print(example["markdown"])
    break

# View all OCR models applied to this dataset
inference_info = json.loads(dataset[0]["inference_info"])
for info in inference_info:
    print(f"Column: {{info['column_name']}} - Model: {{info['model_id']}}")
```

## Reproduction

This dataset was generated using the [uv-scripts/ocr](https://huggingface.co/datasets/uv-scripts/ocr) DeepSeek OCR vLLM script:

```bash
uv run https://huggingface.co/datasets/uv-scripts/ocr/raw/main/deepseek-ocr-vllm.py \\
    NationalLibraryOfScotland/medical-history-of-british-india \\
    <output-dataset> \\
    --image-column image
```

## Performance

- **Processing Speed**: ~0.1 images/second
- **Processing Method**: Batch processing with vLLM (2-3x speedup over sequential)

Generated with [UV Scripts](https://huggingface.co/uv-scripts)