--- title: Video-Reason emoji: "\U0001F3AC" colorFrom: blue colorTo: purple sdk: static pinned: false tags: - video-reasoning - benchmark - video-generation - reasoning ---
### We bet on a future where video reasoning is the next fundamental intelligence paradigm. [๐ŸŒ Website](https://video-reason.com) ยท [๐Ÿ’ป GitHub](https://github.com/VBVR-DataFactory)
## About Us **Video-Reason** is a research initiative dedicated to advancing **video reasoning** as the next foundational intelligence paradigm โ€” where spatiotemporal, embodied world experiences are more naturally captured than through text alone. We build large-scale datasets, benchmarks, and models to systematically study and scale video reasoning capabilities. ## VBVR: A Very Big Video Reasoning Suite Our flagship project, **VBVR (Very Big Video Reasoning)**, introduces an unprecedentedly large-scale resource for video reasoning research: - **200** curated reasoning tasks across 5 domains: **Perception**, **Abstraction**, **Spatiality**, **Transformation**, and **Knowledge** - **1,000,000+** video clips โ€” approximately three orders of magnitude larger than existing datasets - **Verifiable evaluation** via rule-based, human-aligned scorers (no model-based judging) - **Early signs of emergent generalization** to unseen reasoning tasks through large-scale scaling studies ## Releases | Resource | Description | Link | |---|---|---| | VBVR-Wan2.2 | Strong baseline model fine-tuned from Wan2.2-I2V-A14B on the VBVR Dataset | [Model](https://huggingface.co/Video-Reason/VBVR-Wan2.2) | | VBVR-Dataset | 1M video reasoning training samples across 100 curated task generators (~370 GB) | [Dataset](https://huggingface.co/datasets/Video-Reason/VBVR-Dataset) | | VBVR-Bench-Data | Official benchmark test set (500 samples across in-domain and out-of-domain splits) | [Dataset](https://huggingface.co/datasets/Video-Reason/VBVR-Bench-Data) | | VBVR-Bench-Leaderboard | Public leaderboard for standardized model comparison | [Space](https://huggingface.co/spaces/Video-Reason/VBVR-Bench-Leaderboard) | ## Links - **Website**: https://video-reason.com - **Paper**: https://arxiv.org/abs/2602.20159 - **Slack**: https://join.slack.com/t/video-reason/shared_invite/zt-3qqf23icm-UC29fatWWYsIuzRNBR1lgg - **Hugging Face**: https://huggingface.co/Video-Reason - **Contact**: hokinxqdeng@gmail.com ## Citation If you use VBVR in your research, please cite: ```bibtex @article{vbvr2026, title = {A Very Big Video Reasoning Suite}, author = {Wang, Maijunxian and Wang, Ruisi and Lin, Juyi and Ji, Ran and Wiedemer, Thadd{\"a}us and Gao, Qingying and Luo, Dezhi and Qian, Yaoyao and Huang, Lianyu and Hong, Zelong and Ge, Jiahui and Ma, Qianli and He, Hang and Zhou, Yifan and Guo, Lingzi and Mei, Lantao and Li, Jiachen and Xing, Hanwen and Zhao, Tianqi and Yu, Fengyuan and Xiao, Weihang and Jiao, Yizheng and Hou, Jianheng and Zhang, Danyang and Xu, Pengcheng and Zhong, Boyang and Zhao, Zehong and Fang, Gaoyun and Kitaoka, John and Xu, Yile and Xu, Hua and Blacutt, Kenton and Nguyen, Tin and Song, Siyuan and Sun, Haoran and Wen, Shaoyue and He, Linyang and Wang, Runming and Wang, Yanzhi and Yang, Mengyue and Ma, Ziqiao and Milli{\`e}re, Rapha{\"e}l and Shi, Freda and Vasconcelos, Nuno and Khashabi, Daniel and Yuille, Alan and Du, Yilun and Liu, Ziming and Lin, Dahua and Liu, Ziwei and Kumar, Vikash and Li, Yijiang and Yang, Lei and Cai, Zhongang and Deng, Hokin}, journal = {arXiv preprint arXiv:2602.20159}, year = {2026}, url = {https://arxiv.org/abs/2602.20159} } ```