Instructions to use tsystems/colqwen2.5-3b-base with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use tsystems/colqwen2.5-3b-base with Transformers:
# Load model directly from transformers import AutoProcessor, ColQwen2_5 processor = AutoProcessor.from_pretrained("tsystems/colqwen2.5-3b-base") model = ColQwen2_5.from_pretrained("tsystems/colqwen2.5-3b-base") - Notebooks
- Google Colab
- Kaggle
ColQwen2.5-3b: Visual Retriever based on Qwen2.5-VL-3B-Instruct with ColBERT strategy
ColQwen is a model based on a novel model architecture and training strategy based on Vision Language Models (VLMs) to efficiently index documents from their visual features. It is a Qwen2.5-VL-3B extension that generates ColBERT- style multi-vector representations of text and images. It was introduced in the paper ColPali: Efficient Document Retrieval with Vision Language Models and first released in this repository
This version is the untrained base version to guarantee deterministic projection layer initialization.
Usage
This version should not be used: it is solely the base version useful for deterministic LoRA initialization.
Citation
If you use any datasets or models from this organization in your research, please cite the original dataset as follows:
@misc{faysse2024colpaliefficientdocumentretrieval,
title={ColPali: Efficient Document Retrieval with Vision Language Models},
author={Manuel Faysse and Hugues Sibille and Tony Wu and Bilel Omrani and Gautier Viaud and Céline Hudelot and Pierre Colombo},
year={2024},
eprint={2407.01449},
archivePrefix={arXiv},
primaryClass={cs.IR},
url={https://arxiv.org/abs/2407.01449},
}
- Developed by: T-Systems International
- Downloads last month
- 1,769
Model tree for tsystems/colqwen2.5-3b-base
Base model
Qwen/Qwen2.5-VL-3B-Instruct