Create README.md
Browse files
README.md
ADDED
|
@@ -0,0 +1,32 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: mit
|
| 3 |
+
datasets:
|
| 4 |
+
- KwaiVGI/VideoGen-RewardBench
|
| 5 |
+
- TIGER-Lab/GenAI-Bench
|
| 6 |
+
base_model:
|
| 7 |
+
- Qwen/Qwen2.5-VL-7B-Instruct
|
| 8 |
+
---
|
| 9 |
+
|
| 10 |
+
|
| 11 |
+
## Model Summary
|
| 12 |
+
|
| 13 |
+
VR-Thinker is the first Multimodal Reward Model utilizing Thinking-with-Image framework.
|
| 14 |
+
|
| 15 |
+
For further details, please refer to the following:
|
| 16 |
+
- 📰 Paper: https://arxiv.org/pdf/2510.10518
|
| 17 |
+
- 📚 Github: https://github.com/qunzhongwang/vr-thinker
|
| 18 |
+
- 👋 Contact: [Qunzhong Wang](http://qunzhongwang.github.io/)
|
| 19 |
+
|
| 20 |
+
|
| 21 |
+
## Citation
|
| 22 |
+
```
|
| 23 |
+
@misc{wang2025vrthinkerboostingvideoreward,
|
| 24 |
+
title={VR-Thinker: Boosting Video Reward Models through Thinking-with-Image Reasoning},
|
| 25 |
+
author={Qunzhong Wang and Jie Liu and Jiajun Liang and Yilei Jiang and Yuanxing Zhang and Jinyuan Chen and Yaozhi Zheng and Xintao Wang and Pengfei Wan and Xiangyu Yue and Jiaheng Liu},
|
| 26 |
+
year={2025},
|
| 27 |
+
eprint={2510.10518},
|
| 28 |
+
archivePrefix={arXiv},
|
| 29 |
+
primaryClass={cs.CV},
|
| 30 |
+
url={https://arxiv.org/abs/2510.10518},
|
| 31 |
+
}
|
| 32 |
+
```
|