ProactiveVideoQA / README.md
wangyueqian's picture
Update README.md
7f59bce verified
ProactiveVideoQA: A Comprehensive Benchmark Evaluating Proactive Interactions in Video Large Language Models
---
<div align="center">
<div style="margin: 30px 0">
<a href="https://arxiv.org/abs/2507.09313" style="margin: 0 10px">📄 arXiv Paper</a> |
<a href="https://github.com/yellow-binary-tree/ProactiveVideoQA" style="margin: 0 10px"> 🖥️ Github Code </a> |
<a href="https://huggingface.co/datasets/wangyueqian/ProactiveVideoQA" style="margin: 0 10px">📦 Data</a>
</div>
</div>
## Introduction
ProactiveVideoQA is the first comprehensive benchmark designed to evaluate a system's ability to engage in proactive interaction in multimodal dialogue settings.
Unlike traditional turn-by-turn dialogue systems, in proactive intraction model need to determine when to repsond during the playback, so both response timing and response textual content are important points for evaluation.
## Dataset Statistics
ProactiveVideoQA contains 4 tasks:
1. **Proactive web-video QA** `[WEB]`: centering on general web-video understanding.
1. **Proactive ego-centric video QA** `[EGO]`: centering on first-person-view video comprehension, particularly relevant in robotics and daily assistant applications.
1. **Proactive TV-series video QA** `[TV]`: emphasizing dialogue and social relationship understanding with speech input, and
1. **Proactive video anomaly detection** `[VAD]` targeting surveillance video monitoring and alerting.
- **1377** videos from different sources
- **1427** different qeustions, and **3510** ground truth reply turns
- Fully proactive questions and open-ended answers ✅
## Data Format
Each test example in `{dataset}/anno.json` has the following format:
```json
{
"question_id": "OSfMU69X3C4.7.mp4", // unique identifier for this test example
"video": "OSfMU69X3C4.7.mp4", // video file name in `video` folder
"conversation": [ // model input
{"role": "user", "time": 0, "content": "What are the people doing in the office?"}
],
"answer": [ // expected model output
{ // model are expected to reply with the content in the reply timespan
"role": "assistant", "content": "People are working at workstations.",
"reply_timespan": [0.0, 9.88]
},
{ ... }
]
}
```
## Citation
```bibtex
@misc{wang2025proactivevideoqacomprehensivebenchmarkevaluating,
title={ProactiveVideoQA: A Comprehensive Benchmark Evaluating Proactive Interactions in Video Large Language Models},
author={Yueqian Wang and Xiaojun Meng and Yifan Wang and Huishuai Zhang and Dongyan Zhao},
year={2025},
eprint={2507.09313},
archivePrefix={arXiv},
primaryClass={cs.CV},
url={https://arxiv.org/abs/2507.09313},
}
```