Instructions to use echo840/MonkeyOCR with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- MonkeyOCR
How to use echo840/MonkeyOCR with MonkeyOCR:
# No code snippets available yet for this library. # To use this model, check the repository files and the library's documentation. # Want to help? PRs adding snippets are welcome at: # https://github.com/huggingface/huggingface.js
- Notebooks
- Google Colab
- Kaggle
Add pipeline tag and library name
#1
by nielsr HF Staff - opened
README.md
CHANGED
|
@@ -1,6 +1,9 @@
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
|
|
|
|
|
|
| 3 |
---
|
|
|
|
| 4 |
<div align="center" xmlns="http://www.w3.org/1999/html">
|
| 5 |
<h1 align="center">
|
| 6 |
MonkeyOCR: Document Parsing with a Structure-Recognition-Relation Triplet Paradigm
|
|
@@ -579,6 +582,7 @@ Here are the evaluation results of our model on OmniDocBench. MonkeyOCR-3B uses
|
|
| 579 |
|
| 580 |
## Visualization Demo
|
| 581 |
|
|
|
|
| 582 |
> Our demo is simple and easy to use:
|
| 583 |
>
|
| 584 |
> 1. Upload a PDF or image.
|
|
@@ -587,6 +591,12 @@ Here are the evaluation results of our model on OmniDocBench. MonkeyOCR-3B uses
|
|
| 587 |
|
| 588 |
|
| 589 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 590 |
## Citing MonkeyOCR
|
| 591 |
|
| 592 |
If you wish to refer to the baseline results published here, please use the following BibTeX entries:
|
|
@@ -610,4 +620,4 @@ We would like to thank [MinerU](https://github.com/opendatalab/MinerU), [DocLayo
|
|
| 610 |
|
| 611 |
|
| 612 |
## Copyright
|
| 613 |
-
MonkeyDoc dataset was collected from public datasets, crawled from the internet, and obtained through our own photography.
|
|
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
| 3 |
+
pipeline_tag: visual-document-retrieval
|
| 4 |
+
library_name: transformers
|
| 5 |
---
|
| 6 |
+
|
| 7 |
<div align="center" xmlns="http://www.w3.org/1999/html">
|
| 8 |
<h1 align="center">
|
| 9 |
MonkeyOCR: Document Parsing with a Structure-Recognition-Relation Triplet Paradigm
|
|
|
|
| 582 |
|
| 583 |
## Visualization Demo
|
| 584 |
|
| 585 |
+
Demo Link: http://vlrlabmonkey.xyz:7685
|
| 586 |
> Our demo is simple and easy to use:
|
| 587 |
>
|
| 588 |
> 1. Upload a PDF or image.
|
|
|
|
| 591 |
|
| 592 |
|
| 593 |
|
| 594 |
+
### Support diverse Chinese and English PDF types
|
| 595 |
+
|
| 596 |
+
<p align="center">
|
| 597 |
+
<img src="asserts/Visualization.GIF?raw=true" width="600"/>
|
| 598 |
+
</p>
|
| 599 |
+
|
| 600 |
## Citing MonkeyOCR
|
| 601 |
|
| 602 |
If you wish to refer to the baseline results published here, please use the following BibTeX entries:
|
|
|
|
| 620 |
|
| 621 |
|
| 622 |
## Copyright
|
| 623 |
+
MonkeyDoc dataset was collected from public datasets, crawled from the internet, and obtained through our own photography. The current technical report only presents the results of the 3B model. If you are interested in larger one, please contact Prof. Yuliang Liu at ylliu@hust.edu.cn.
|