Spaces:
Running
Running
Update README.md
Browse files
README.md
CHANGED
|
@@ -6,63 +6,68 @@ colorTo: red
|
|
| 6 |
sdk: static
|
| 7 |
pinned: true
|
| 8 |
---
|
| 9 |
-
# ColPali: Efficient Document Retrieval with Vision Language Models 👀
|
| 10 |
|
| 11 |
-
|
| 12 |
|
| 13 |
-
<
|
| 14 |
|
| 15 |
-
|
| 16 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 17 |
|
| 18 |
-
|
| 19 |
|
| 20 |
-
Documents are visually rich structures that convey information through text, as well as tables, figures, page layouts, or fonts.
|
| 21 |
-
While modern document retrieval systems exhibit strong performance on query-to-text matching, they struggle to exploit visual cues efficiently, hindering their performance on practical document retrieval applications such as Retrieval Augmented Generation.
|
| 22 |
-
To benchmark current systems on visually rich document retrieval, we introduce the Visual Document Retrieval Benchmark *ViDoRe*, composed of various page-level retrieving tasks spanning multiple domains, languages, and settings.
|
| 23 |
-
The inherent shortcomings of modern systems motivate the introduction of a new retrieval model architecture, *ColPali*, which leverages the document understanding capabilities of recent Vision Language Models to produce high-quality contextualized embeddings solely from images of document pages.
|
| 24 |
-
Combined with a late interaction matching mechanism, *ColPali* largely outperforms modern document retrieval pipelines while being drastically faster and end-to-end trainable.
|
| 25 |
|
| 26 |
-
|
|
|
|
| 27 |
|
| 28 |
-
-
|
| 29 |
-
|
| 30 |
-
- [
|
| 31 |
-
|
| 32 |
-
- [*BiSigLIP*](https://huggingface.co/vidore/bisiglip): Finetuned version of original [SigLIP](https://huggingface.co/google/siglip-so400m-patch14-384), a strong vision-language bi-encoder model.
|
| 33 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 34 |
|
| 35 |
-
|
| 36 |
|
| 37 |
-
|
| 38 |
|
| 39 |
-
|
| 40 |
|
| 41 |
-
|
| 42 |
|
| 43 |
-
|
| 44 |
-
datasets ([ArXiVQA](https://huggingface.co/datasets/vidore/arxivqa_test_subsampled), [DocVQA](https://huggingface.co/datasets/vidore/docvqa_test_subsampled),
|
| 45 |
-
[InfoVQA](https://huggingface.co/datasets/vidore/infovqa_test_subsampled), [TATDQA](https://huggingface.co/datasets/vidore/tatdqa_test), [TabFQuAD](https://huggingface.co/datasets/vidore/tabfquad_test_subsampled)) and from datasets synthetically generated spanning various themes and industrial applications:
|
| 46 |
-
([Artificial Intelligence](https://huggingface.co/datasets/vidore/syntheticDocQA_artificial_intelligence_test), [Government Reports](https://huggingface.co/datasets/vidore/syntheticDocQA_government_reports_test), [Healthcare Industry](https://huggingface.co/datasets/vidore/syntheticDocQA_healthcare_industry_test), [Energy](https://huggingface.co/datasets/vidore/syntheticDocQA_energy_test) and [Shift Project](https://huggingface.co/datasets/vidore/shiftproject_test)).
|
| 47 |
-
Further details can be found on the corresponding dataset cards.
|
| 48 |
-
- [*OCR Baseline*](https://huggingface.co/collections/vidore/vidore-chunk-ocr-baseline-666acce88c294ef415548a56): Datasets in this collection are the same as in ViDoRe but preprocessed for textual retrieving. The original ViDoRe benchmark was passed to Unstructured to partition each page into chunks. Visual chunks are OCRized with Tesseract.
|
| 49 |
-
- [*Captioning Baseline*](https://huggingface.co/collections/vidore/vidore-captioning-baseline-6658a2a62d857c7a345195fd): Datasets in this collection are the same as in ViDoRe but preprocessed for textual retrieving. The original ViDoRe benchmark was passed to Unstructured to partition each page into chunks. Visual chunks are captioned using Claude Sonnet.
|
| 50 |
|
| 51 |
-
|
| 52 |
|
| 53 |
-
|
| 54 |
-
- [*ViDoRe Benchmark*](https://github.com/illuin-tech/vidore-benchmark): A Python package/CLI tool to evaluate document retrieval systems on the ViDoRe benchmark.
|
| 55 |
|
| 56 |
-
|
|
|
|
| 57 |
|
| 58 |
-
|
| 59 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 60 |
|
| 61 |
## Contact
|
| 62 |
|
| 63 |
-
-
|
| 64 |
-
|
| 65 |
-
|
| 66 |
|
| 67 |
## Citation
|
| 68 |
|
|
@@ -90,8 +95,9 @@ If you use any datasets or models from this organization in your research, pleas
|
|
| 90 |
}
|
| 91 |
```
|
| 92 |
|
|
|
|
|
|
|
| 93 |
## Acknowledgments
|
| 94 |
|
| 95 |
-
This work is partially supported by [ILLUIN Technology](https://www.illuin.tech/)
|
| 96 |
-
This work was performed using HPC resources from the CINES ADASTRA through Grant 2024-AD011015443.
|
| 97 |
|
|
|
|
| 6 |
sdk: static
|
| 7 |
pinned: true
|
| 8 |
---
|
|
|
|
| 9 |
|
| 10 |
+
<img src="https://cdn-uploads.huggingface.co/production/uploads/66e16a677c2eb2da5109fb5c/x99xqw__fl2UaPbiIdC_f.png" width="180" style="display: block; margin-left: auto; margin-right: auto;" />
|
| 11 |
|
| 12 |
+
<h1 align="center">ViDoRe: Visual Document Retrieval 👀</h1>
|
| 13 |
|
| 14 |
+
> [!IMPORTANT]
|
| 15 |
+
> **ViDoRe V3 is our new benchmark release!**
|
| 16 |
+
>
|
| 17 |
+
> Built in collaboration with NVIDIA, it is the most diverse visual document retrieval benchmark to date for **real-world applications**.
|
| 18 |
+
> It includes **10 datasets**, **+26,000 pages**, **+3000 queries** in **6 languages**.
|
| 19 |
+
> Check it out here! [⚠ link to the blogpost]
|
| 20 |
|
| 21 |
+
This organization contains models, datasets, benchmarks and code released with the ViDoRe project by Illuin Technology.
|
| 22 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 23 |
|
| 24 |
+
- **Leaderboard**:
|
| 25 |
+
- [ViDoRe Leaderboard](https://huggingface.co/spaces/vidore/vidore-leaderboard)
|
| 26 |
|
| 27 |
+
- **Benchmarks:**
|
| 28 |
+
- [ViDoRe V1](https://huggingface.co/collections/vidore/vidore-benchmark)
|
| 29 |
+
- [ViDoRe V2](https://huggingface.co/collections/vidore/vidore-benchmark-v2)
|
| 30 |
+
- ViDoRe V3 (⚠ LINK)
|
|
|
|
| 31 |
|
| 32 |
+
- **Models:**
|
| 33 |
+
- ColPali ([latest: v1.3](https://huggingface.co/vidore/colpali-v1.3))
|
| 34 |
+
- ColQwen2 ([latest: v1.0](https://huggingface.co/vidore/colqwen2-v1.0))
|
| 35 |
+
- ColQwen2.5 ([latest: v0.2](https://huggingface.co/vidore/colqwen2.5-v0.2))
|
| 36 |
+
- ColSmol ([256M](https://huggingface.co/vidore/colSmol-256M) & [500M](https://huggingface.co/vidore/colSmol-500M))
|
| 37 |
+
- ModernVBERT ([latest: v1.0](https://huggingface.co/ModernVBERT))
|
| 38 |
|
| 39 |
+
---
|
| 40 |
|
| 41 |
+
# 👷♂️ ViDoRe V3: A comprehensive evaluation of Retrieval in real-world use cases
|
| 42 |
|
| 43 |
+
<img src="https://cdn-uploads.huggingface.co/production/uploads/66e16a677c2eb2da5109fb5c/-zqFfhdtsC1VzQH-rLkLa.png" width="1300" style="display: block; margin-left: auto; margin-right: auto;" />
|
| 44 |
|
| 45 |
+
ILLUIN Technology is proud to release the **ViDoRe V3 benchmark**, designed and developed with contributions from NVIDIA. ViDoRe V3 is our latest benchmark, engineered to set a new industry gold standard for multi-modal, enterprise document retrieval evaluation. It addresses a critical challenge in production RAG systems: retrieving accurate information from complex, visually-rich documents.
|
| 46 |
|
| 47 |
+
ViDoRe V3 improves on existing RAG benchmarks by prioritizing enterprise relevance and rigorous data quality. Instead of relying on clean academic texts, the benchmark draws from 10 challenging, real-world datasets spanning diverse industrial domains, with 8 publicly released and 2 kept private. In addition, while previous benchmarks often rely on synthetically generated data, ViDoRe V3 features human-created and human-verified annotations.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 48 |
|
| 49 |
+
This benchmark contains 26,000 pages and 3,099 queries translated into 6 languages. Each query is linked to retrieval ground truth data created and verified by human annotators: relevant pages, precise bounding box annotations for key elements, and a comprehensive reference answer.
|
| 50 |
|
| 51 |
+
---
|
|
|
|
| 52 |
|
| 53 |
+
# 👀 ColPali: Efficient Document Retrieval with Vision Language Models
|
| 54 |
+
[](https://arxiv.org/abs/2407.01449)
|
| 55 |
|
| 56 |
+
<img src="https://cdn-uploads.huggingface.co/production/uploads/60f2e021adf471cbdf8bb660/T3z7_Biq3oW6b8I9ZwpIa.png" width="800" style="display: block; margin-left: auto; margin-right: auto;" />
|
| 57 |
+
|
| 58 |
+
Documents are visually rich structures that convey information through text, as well as tables, figures, page layouts, or fonts.
|
| 59 |
+
While modern document retrieval systems exhibit strong performance on query-to-text matching, they struggle to exploit visual cues efficiently, hindering their performance on practical document retrieval applications such as Retrieval Augmented Generation.
|
| 60 |
+
To benchmark current systems on visually rich document retrieval, we introduce the Visual Document Retrieval Benchmark *ViDoRe*, composed of various page-level retrieving tasks spanning multiple domains, languages, and settings.
|
| 61 |
+
The inherent shortcomings of modern systems motivate the introduction of a new retrieval model architecture, *ColPali*, which leverages the document understanding capabilities of recent Vision Language Models to produce high-quality contextualized embeddings solely from images of document pages.
|
| 62 |
+
Combined with a late interaction matching mechanism, *ColPali* largely outperforms modern document retrieval pipelines while being drastically faster and end-to-end trainable.
|
| 63 |
+
|
| 64 |
+
---
|
| 65 |
|
| 66 |
## Contact
|
| 67 |
|
| 68 |
+
- Quentin Macé: `quentin.mace@illuin.tech`
|
| 69 |
+
|
| 70 |
+
---
|
| 71 |
|
| 72 |
## Citation
|
| 73 |
|
|
|
|
| 95 |
}
|
| 96 |
```
|
| 97 |
|
| 98 |
+
---
|
| 99 |
+
|
| 100 |
## Acknowledgments
|
| 101 |
|
| 102 |
+
This work is partially supported by [ILLUIN Technology](https://www.illuin.tech/).
|
|
|
|
| 103 |
|