--- license: mit task_categories: - table-question-answering - image-to-text tags: - table-extraction - benchmark - fintabnet - document-ai - docld pretty_name: DocLD FinTabNet Benchmark size_categories: - n<1K --- # DocLD FinTabNet Benchmark Results Benchmark results for [DocLD](https://docld.com) table extraction on the [FinTabNet](https://paperswithcode.com/dataset/fintabnet) dataset. ## Results Summary | Metric | Value | |--------|-------| | **Mean Accuracy** | 82.1% | | **Median** | 83.2% | | **P25 / P75** | 73.3% / 97.4% | | **Min / Max** | 22.7% / 100.0% | | **Scored Samples** | 500 | | **Total Samples** | 500 | ## Methodology - **Dataset**: [FinTabNet_OTSL](https://huggingface.co/datasets/docling-project/FinTabNet_OTSL) — 500 samples from the test split - **Extraction**: DocLD vision-based table extraction - **Scoring**: Needleman-Wunsch hierarchical alignment (same as [RD-TableBench](https://github.com/reductoai/rd-tablebench)) - **Output**: HTML tables with rowspan/colspan for merged cells ## Comparison | Provider | FinTabNet Accuracy | |----------|-------------------| | **DocLD** | **82.1%** | | GTE (IBM) | ~78% | | TATR (Microsoft) | ~65% | ## Files - `results.json` — Full benchmark results with per-sample scores - `predictions/` — HTML predictions for each sample - `charts/` — Visualization PNGs ## Links - [DocLD](https://docld.com) - [Blog Post](https://docld.com/blog/docld-fintabnet) - [Benchmark Code](https://github.com/Doc-LD/fintabnet-bench) - [RD-TableBench Results](https://docld.com/blog/docld-tablebench)