Hokin commited on
Commit
710dbc0
·
verified ·
1 Parent(s): 96221bf

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +51 -5
README.md CHANGED
@@ -1,10 +1,56 @@
1
  ---
2
- title: README
3
- emoji: 😻
4
- colorFrom: green
5
- colorTo: red
6
  sdk: static
7
  pinned: false
 
 
 
 
 
8
  ---
9
 
10
- Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ title: Video-Reason
3
+ emoji: 🎬
4
+ colorFrom: blue
5
+ colorTo: purple
6
  sdk: static
7
  pinned: false
8
+ tags:
9
+ - video-reasoning
10
+ - benchmark
11
+ - video-generation
12
+ - reasoning
13
  ---
14
 
15
+ <div align="center">
16
+
17
+ # Video-Reason
18
+
19
+ ### We bet on a future where video reasoning is the next foundational intelligence paradigm.
20
+
21
+ [🌐 Website](https://video-reason.com) | [🤗 Leaderboard](https://huggingface.co/spaces/Video-Reason/VBVR-Bench-Leaderboard) | [📄 Paper](https://video-reason.com) | [💻 GitHub](https://github.com/Video-Reason)
22
+
23
+ </div>
24
+
25
+ ---
26
+
27
+ ## About Us
28
+
29
+ **Video-Reason** is a research initiative dedicated to advancing **video reasoning** as the next foundational intelligence paradigm — where spatiotemporal, embodied world experiences are more naturally captured than through text alone. We build large-scale datasets, benchmarks, and models to systematically study and scale video reasoning capabilities.
30
+
31
+ ## VBVR: A Very Big Video Reasoning Suite
32
+
33
+ Our flagship project, **VBVR (Very Big Video Reasoning)**, introduces an unprecedentedly large-scale resource for video reasoning research:
34
+
35
+ - **200** curated reasoning tasks across 5 domains: **Perception**, **Abstraction**, **Spatiality**, **Transformation**, and **Knowledge**
36
+ - **1,000,000+** video clips — approximately three orders of magnitude larger than existing datasets
37
+ - **Verifiable evaluation** via rule-based, human-aligned scorers (no model-based judging)
38
+ - **Early signs of emergent generalization** to unseen reasoning tasks through large-scale scaling studies
39
+
40
+ ## Releases
41
+
42
+ | Resource | Description | Link |
43
+ |---|---|---|
44
+ | **VBVR-Wan2.2** | Strong baseline model fine-tuned from Wan2.2-I2V-A14B on the VBVR Dataset | [Model](https://huggingface.co/Video-Reason/VBVR-Wan2.2) |
45
+ | **VBVR-Bench-Data** | Official benchmark test set (500 samples across in-domain and out-of-domain splits) | [Dataset](https://huggingface.co/datasets/Video-Reason/VBVR-Bench-Data) |
46
+ | **VBVR-Bench-Leaderboard** | Public leaderboard for standardized model comparison | [Space](https://huggingface.co/spaces/Video-Reason/VBVR-Bench-Leaderboard) |
47
+
48
+ ## Citation
49
+
50
+ ```bibtex
51
+ @article{vbvr2026,
52
+ title={A Very Big Video Reasoning Suite},
53
+ author={Wang, Maijunxian and Wang, Ruisi and Lin, Juyi and Ji, Ran and Wiedemer, Thaddäus and Gao, Qingying and Luo, Dezhi and Qian, Yaoyao and Huang, Lianyu and Hong, Zelong and Ge, Jiahui and Ma, Qianli and He, Hang and Zhou, Yifan and Guo, Lingzi and Mei, Lantao and Li, Jiachen and Xing, Hanwen and Zhao, Tianqi and Yu, Fengyuan and Xiao, Weihang and Jiao, Yizheng and Hou, Jianheng and Zhang, Danyang and Xu, Pengcheng and Zhong, Boyang and Zhao, Zehong and Fang, Gaoyun and Kitaoka, John and Xu, Yile and Xu, Hua and Blacutt, Kenton and Nguyen, Tin and Song, Siyuan and Sun, Haoran and Wen, Shaoyue and He, Linyang and Wang, Runming and Wang, Yanzhi and Yang, Mengyue and Ma, Ziqiao and Millière, Raphaël and Shi, Freda and Vasconcelos, Nuno and Khashabi, Daniel and Yuille, Alan and Du, Yilun and Liu, Ziming and Lin, Dahua and Liu, Ziwei and Kumar, Vikash and Li, Yijiang and Yang, Lei and Cai, Zhongang and Deng, Hokin},
54
+ year={2026}
55
+ }
56
+ ```