Enhance OpenVLThinker-7B model card: Add paper info, abstract, and descriptive tags
#3
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -1,11 +1,25 @@
|
|
| 1 |
---
|
| 2 |
base_model:
|
| 3 |
- Qwen/Qwen2.5-VL-7B-Instruct
|
| 4 |
-
license: apache-2.0
|
| 5 |
library_name: transformers
|
|
|
|
| 6 |
pipeline_tag: image-text-to-text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 7 |
---
|
| 8 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 9 |
## Overview
|
| 10 |
OpenVLThinker-7B is a vision-language reasoning model designed to handle multimodal tasks. It is especially tuned for visual mathematical problem-solving.
|
| 11 |
|
|
@@ -85,14 +99,14 @@ print(generated_text)
|
|
| 85 |
```
|
| 86 |
|
| 87 |
### Citation
|
| 88 |
-
```
|
| 89 |
@misc{deng2025openvlthinker,
|
| 90 |
-
title={OpenVLThinker:
|
| 91 |
author={Yihe Deng and Hritik Bansal and Fan Yin and Nanyun Peng and Wei Wang and Kai-Wei Chang},
|
| 92 |
year={2025},
|
| 93 |
eprint={2503.17352},
|
| 94 |
archivePrefix={arXiv},
|
| 95 |
primaryClass={cs.CV},
|
| 96 |
-
url={https://
|
| 97 |
}
|
| 98 |
```
|
|
|
|
| 1 |
---
|
| 2 |
base_model:
|
| 3 |
- Qwen/Qwen2.5-VL-7B-Instruct
|
|
|
|
| 4 |
library_name: transformers
|
| 5 |
+
license: apache-2.0
|
| 6 |
pipeline_tag: image-text-to-text
|
| 7 |
+
tags:
|
| 8 |
+
- reasoning
|
| 9 |
+
- multimodal
|
| 10 |
+
- vlm
|
| 11 |
+
- math
|
| 12 |
+
- visual-question-answering
|
| 13 |
---
|
| 14 |
|
| 15 |
+
# OpenVLThinker: Complex Vision-Language Reasoning via Iterative SFT-RL Cycles
|
| 16 |
+
|
| 17 |
+
This model was presented in the paper [OpenVLThinker: Complex Vision-Language Reasoning via Iterative SFT-RL Cycles](https://huggingface.co/papers/2503.17352).
|
| 18 |
+
|
| 19 |
+
## Abstract
|
| 20 |
+
|
| 21 |
+
We introduce OpenVLThinker, one of the first open-source large vision-language models (LVLMs) to exhibit sophisticated chain-of-thought reasoning, achieving notable performance gains on challenging visual reasoning tasks. While text-based reasoning models (e.g., Deepseek R1) show promising results in text-only tasks, distilling their reasoning into LVLMs via supervised fine-tuning (SFT) often results in performance degradation due to imprecise visual grounding. Conversely, purely reinforcement learning (RL)-based methods face a large search space, hindering the emergence of reflective behaviors in smaller models (e.g., 7B LVLMs). Surprisingly, alternating between SFT and RL ultimately results in significant performance improvements after a few iterations. Our analysis reveals that the base model rarely exhibits reasoning behaviors initially, but SFT effectively surfaces these latent actions and narrows the RL search space, accelerating the development of reasoning capabilities. Each subsequent RL stage further refines the model's reasoning skills, producing higher-quality SFT data for continued self-improvement. OpenVLThinker-7B consistently advances performance across six benchmarks demanding mathematical and general reasoning, notably improving MathVista by 3.8%, EMMA by 2.4%, and HallusionBench by 1.6%. Beyond demonstrating the synergy between SFT and RL for complex reasoning tasks, our findings provide early evidence towards achieving R1-style reasoning in multimodal contexts. The code, model and data are held at this https URL .
|
| 22 |
+
|
| 23 |
## Overview
|
| 24 |
OpenVLThinker-7B is a vision-language reasoning model designed to handle multimodal tasks. It is especially tuned for visual mathematical problem-solving.
|
| 25 |
|
|
|
|
| 99 |
```
|
| 100 |
|
| 101 |
### Citation
|
| 102 |
+
```bibtex
|
| 103 |
@misc{deng2025openvlthinker,
|
| 104 |
+
title={OpenVLThinker: Complex Vision-Language Reasoning via Iterative SFT-RL Cycles},
|
| 105 |
author={Yihe Deng and Hritik Bansal and Fan Yin and Nanyun Peng and Wei Wang and Kai-Wei Chang},
|
| 106 |
year={2025},
|
| 107 |
eprint={2503.17352},
|
| 108 |
archivePrefix={arXiv},
|
| 109 |
primaryClass={cs.CV},
|
| 110 |
+
url={https://huggingface.co/papers/2503.17352},
|
| 111 |
}
|
| 112 |
```
|