xunshuhang commited on
Commit
559f804
verified
1 Parent(s): 21bfa83

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +46 -0
README.md ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ task_categories:
4
+ - video-text-to-text
5
+ language:
6
+ - en
7
+ size_categories:
8
+ - 1K<n<10K
9
+ ---
10
+ <div align="center">
11
+ <h1>RTV-Bench: Benchmarking MLLM Continuous Perception, Understanding and Reasoning through Real-Time Video</h1>
12
+ </div>
13
+
14
+ [![hf_checkpoint](https://img.shields.io/badge/馃-RTV--Bench-9C276A.svg)](https://huggingface.co/datasets/xunsh/RTV-Bench) [![ms_checkpoint](https://img.shields.io/badge/馃-RTV--Bench-8A2BE2.svg)](https://www.modelscope.cn/datasets/Jungang/RTV-Bench)
15
+
16
+
17
+ ## 馃敟 News
18
+ * **`2025.05.03`** 馃専 We are happy to release the RTV-Bench. You can find the RTV-Bench from [![hf_checkpoint](https://img.shields.io/badge/馃-RTV--Bench-9C276A.svg)](https://huggingface.co/datasets/xunsh/RTV-Bench) or [![ms_checkpoint](https://img.shields.io/badge/馃-RTV--Bench-8A2BE2.svg)](https://www.modelscope.cn/datasets/Jungang/RTV-Bench).
19
+
20
+ ## TODO
21
+ - [ ] Release the final label json.
22
+ - [ ] Release the evaluation code.
23
+ - [ ] Construct a more comprehensive benchmark for real-time video analysis.
24
+ - [ ] 路路路
25
+ ## 馃憖 $\mathcal{RTV}\text{-}Bench$ Overview
26
+
27
+ We introduce RTV-Bench, a fine-grained benchmark for MLLM real-time video analysis, which contains **552** videos (167.2 hours) and **4,631** high-quality QA pairs. We evaluated leading MLLMs, including proprietary (GPT-4o, Gemini 2.0), open-source offline (Qwen2.5-VL, VideoLLaMA3), and open-source real-time (VITA-1.5, InternLM-XComposer2.5-OmniLive) models. Experiment results show open-source real-time models largely outperform offline ones but still trail top proprietary models. Our analysis also reveals that larger model size or higher frame sampling rates do not significantly boost $\mathcal{RTV}\text{-}Bench$ performance, sometimes causing slight decreases. This underscores the need for better model architectures optimized for video stream processing and long sequences to advance real-time video analysis with MLLMs. $\mathcal{RTV}\text{-}Bench$ includes three key principles:
28
+ * **Multi-Timestamp Question Answering (MTQA)**, where answers evolve with scene changes;
29
+ * **Hierarchical Question Structure**, combining basic and advanced queries; and
30
+ * **Multi-dimensional Evaluation**, assessing the ability of continuous perception, understanding, and reasoning.
31
+
32
+
33
+ ## 馃専 Star History
34
+
35
+ [![Star History Chart](https://api.star-history.com/svg?repos=LJungang/RTV-Bench&type=Date)](https://star-history.com/#LJungang/RTV-Bench&Date)
36
+
37
+ If you find our work helpful for your research, please consider citing our work.
38
+
39
+ ```bibtex
40
+ @article{xun2025rtv,
41
+ title={RTV-Bench: Benchmarking MLLM Continuous Perception, Understanding and Reasoning through Real-Time Video},
42
+ author={Xun, Shuhang and Tao, Sicheng and Li, Jungang and Shi, Yibo and Lin, Zhixin and Zhu, Zhanhui and Yan, Yibo and Li, Hanqian and Zhang, Linghao and Wang, Shikang and others},
43
+ journal={arXiv preprint arXiv:2505.02064},
44
+ year={2025}
45
+ }
46
+ ```