File size: 1,186 Bytes
522276a |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 |
---
license: mit
language:
- en
- tr
---
# PaddleOCR Mobile Quantized Models (ONNX)
## Overview
This repo hosts four **ONNX** models converted from PaddleOCR mobile checkpoints
| File | Task | Language scope | Input shape |
|------|------|----------------|-------------|
| `Multilingual_PP-OCRv3_det_infer.onnx` | Text-detection | 80+ scripts | **NCHW • 1×3×H×W** |
| `PP-OCRv3_mobile_det_infer.onnx` | Text-detection | Latin only | 1×3×H×W |
| `ch_ppocr_mobile_v2.0_cls_infer.onnx` | Angle classifier | Chinese/Latin | 1×3×H×W |
| `latin_PP-OCRv3_mobile_rec_infer.onnx` | Text-recognition | Latin | 1×3×H×W |
All models were:
* exported with **paddle2onnx 1.2.3** (`opset 11`)
* simplified via **onnx-simplifier 0.4+**
## Quick Start
```python
import onnxruntime as ort, numpy as np
img = np.random.rand(1, 3, 224, 224).astype("float32")
det = ort.InferenceSession("Multilingual_PP-OCRv3_det_infer.onnx")
cls = ort.InferenceSession("ch_ppocr_mobile_v2.0_cls_infer.onnx")
rec = ort.InferenceSession("latin_PP-OCRv3_mobile_rec_infer.onnx")
det_out = det.run(None, {det.get_inputs()[0].name: img})[0]
# add your post-processing / cropping / decoding here …
|