|
|
--- |
|
|
base_model: |
|
|
- Qwen/Qwen2.5-VL-7B-Instruct |
|
|
pipeline_tag: image-text-to-text |
|
|
library_name: adapter-transformers |
|
|
--- |
|
|
|
|
|
This is the **vSearcher** model introduced in paper ["InSight-o3: Empowering Multimodal Foundation Models with Generalized Visual Search"](https://arxiv.org/abs/2512.18745). |
|
|
The model is finetuned from `Qwen2.5-VL-7B-Instruct` via RL as a subagent under **vReasoner** `GPT-5-mini`. |
|
|
For more information on how to use this model, see our [GitHub page](https://github.com/m-Just/InSight-o3). |
|
|
|
|
|
``` |
|
|
@inproceedings{li2026insight_o3, |
|
|
title={InSight-o3: Empowering Multimodal Foundation Models with Generalized Visual Search}, |
|
|
author={Kaican Li and Lewei Yao and Jiannan Wu and Tiezheng Yu and Jierun Chen and Haoli Bai and Lu Hou and Lanqing Hong and Wei Zhang and Nevin L. Zhang}, |
|
|
booktitle={The Fourteenth International Conference on Learning Representations}, |
|
|
year={2026} |
|
|
} |
|
|
``` |