glm-ocr-onnx / README.md
aoiandroid's picture
Upload README.md with huggingface_hub
bb792b9 verified
---
license: mit
pipeline_tag: image-to-text
tags:
- onnx
- ocr
- glm-ocr
base_model: zai-org/GLM-OCR
---
# GLM-OCR ONNX (Decoder)
ONNX export of the **decoder** of [zai-org/GLM-OCR](https://huggingface.co/zai-org/GLM-OCR). Exported with `scripts/export_glm_ocr_onnx.py` (Transformers 5.1.0, custom torch.onnx path).
## Contents
- `glm_ocr_decoder.onnx` / `glm_ocr_decoder.onnx.data` – Decoder ONNX (inputs: `decoder_input_ids`, `encoder_hidden_states`; output: `logits`).
- `tokenizer.json`, `tokenizer_config.json` – Tokenizer from zai-org/GLM-OCR.
## Note
The **vision encoder** was not exported (model forward requires either `input_ids` or `inputs_embeds` when called with image inputs only). To run full OCR you need encoder hidden states from another source or the original PyTorch model for the vision part.
## Usage
Load with ONNX Runtime; feed `encoder_hidden_states` (from your vision encoder or zai-org/GLM-OCR in PyTorch) and `decoder_input_ids`; get `logits` and decode with the included tokenizer.
## Source
- Base model: [zai-org/GLM-OCR](https://huggingface.co/zai-org/GLM-OCR)
- Export spec: See [Docs/GLM_OCR_ONNX_Export.md](https://github.com/...) in the TranslateBlue repo.