--- license: mit task_categories: - video-text-to-text --- # Video-Browsecomp [**Project Page**](https://liang-zhengyang.github.io/video-browsecomp/) | [**Paper**](https://huggingface.co/papers/2512.23044) | [**Code**](https://github.com/chrisx599/Video-Browser) Video-Browsecomp is a benchmark designed for evaluating open-ended agentic browsing tasks that enforce a mandatory dependency on videos. It was introduced in the paper "Video-Browser: Towards Agentic Open-web Video Browsing". The benchmark challenges autonomous agents to research, watch, and analyze video content to answer complex, open-ended user queries. Along with the benchmark, the authors propose **Video-Browser**, a novel agent leveraging *Pyramidal Perception* to efficiently filter information using metadata and perform fine-grained visual verification only when necessary. ## Citation If you use the Video-Browsecomp benchmark or the Video-Browser agent in your research, please cite: ```bibtex @misc{liang2026videobrowseragenticopenwebvideo, title={Video-Browser: Towards Agentic Open-web Video Browsing}, author={Zhengyang Liang and Yan Shu and Xiangrui Liu and Minghao Qin and Kaixin Liang and Paolo Rota and Nicu Sebe and Zheng Liu and Lizi Liao}, year={2026}, eprint={2512.23044}, archivePrefix={arXiv}, primaryClass={cs.CV}, url={https://arxiv.org/abs/2512.23044}, } ```