Update README.md
Browse files
README.md
CHANGED
|
@@ -6,7 +6,7 @@
|
|
| 6 |
|
| 7 |
# Overview
|
| 8 |
|
| 9 |
-
Infinity-Doc-55K is a high-quality diverse full-text parsing dataset, comprising 55,
|
| 10 |
|
| 11 |

|
| 12 |
|
|
@@ -14,6 +14,8 @@ Infinity-Doc-55K is a high-quality diverse full-text parsing dataset, comprising
|
|
| 14 |
|
| 15 |
To construct a comprehensive dataset for document parsing, we integrate both real-world and synthetic data generation pipelines. Our real-world data pipeline collects diverse scanned documents from various practical domains (such as financial reports, medical records, and academic papers), employing a multi-expert strategy with cross-validation to generate reliable pseudo-ground-truth annotations for structural elements like text, tables, and formulas. Complementing this, our synthetic data pipeline programmatically creates a wide array of documents by injecting content from sources like Wikipedia into predefined HTML layouts, rendering them into scanned formats, and extracting precise ground-truth annotations directly from the original HTML. This dual approach yields a rich, diverse, and cost-effective dataset with accurate and well-aligned supervision, effectively overcoming common issues of imprecise or inconsistent labeling found in other datasets and enabling robust training for end-to-end document parsing models.
|
| 16 |
|
|
|
|
|
|
|
| 17 |

|
| 18 |
|
| 19 |
# Data Statistics
|
|
|
|
| 6 |
|
| 7 |
# Overview
|
| 8 |
|
| 9 |
+
Infinity-Doc-55K is a high-quality diverse full-text parsing dataset, comprising 55,059 real-world and synthetic scanned documents. The dataset features rich layout variations and comprehensive structural annotations, enabling robust training of document parsing models. Additionally, this dataset encompasses a broad spectrum of document types, including financial reports, medical reports, academic reports, books, magazines, web pages, and synthetic documents.
|
| 10 |
|
| 11 |

|
| 12 |
|
|
|
|
| 14 |
|
| 15 |
To construct a comprehensive dataset for document parsing, we integrate both real-world and synthetic data generation pipelines. Our real-world data pipeline collects diverse scanned documents from various practical domains (such as financial reports, medical records, and academic papers), employing a multi-expert strategy with cross-validation to generate reliable pseudo-ground-truth annotations for structural elements like text, tables, and formulas. Complementing this, our synthetic data pipeline programmatically creates a wide array of documents by injecting content from sources like Wikipedia into predefined HTML layouts, rendering them into scanned formats, and extracting precise ground-truth annotations directly from the original HTML. This dual approach yields a rich, diverse, and cost-effective dataset with accurate and well-aligned supervision, effectively overcoming common issues of imprecise or inconsistent labeling found in other datasets and enabling robust training for end-to-end document parsing models.
|
| 16 |
|
| 17 |
+
Note that the elements in 'form' are sorted in reading order.
|
| 18 |
+
|
| 19 |

|
| 20 |
|
| 21 |
# Data Statistics
|