VCBench / README.md
nielsr's picture
nielsr HF Staff
Add paper link, GitHub link, and update metadata for VCBench
b13331d verified
|
raw
history blame
3.04 kB
---
language:
- en
license: cc-by-4.0
size_categories:
- 1K<n<10K
task_categories:
- video-text-to-text
tags:
- video-understanding
- temporal-reasoning
- counting
- benchmark
---
# VCBench: A Streaming Counting Benchmark for Spatial-Temporal State Maintenance in Long Videos
[**Paper**](https://huggingface.co/papers/2603.12703) | [**Code**](https://github.com/buaaplay/VCBench) | [**Dataset**](https://huggingface.co/datasets/buaaplay/VCBench)
VCBench is a streaming counting benchmark that repositions counting as a minimal probe for diagnosing **spatial-temporal state maintenance** capability in video-language models. By querying models at multiple timepoints during video playback, VCBench observes how model predictions evolve rather than checking isolated answers.
## Task Taxonomy
VCBench decomposes state maintenance into 8 fine-grained subcategories across two dimensions:
### Object Counting (tracking entities)
| Subcategory | Description |
|-------------|-------------|
| **O1-Snap** | How many objects are visible *at this moment*? |
| **O1-Delta** | How many objects appeared in the *past N seconds*? |
| **O2-Unique** | How many *different* individuals have appeared so far? |
| **O2-Gain** | How many *new* individuals appeared in the past N seconds? |
### Event Counting (tracking actions)
| Subcategory | Description |
|-------------|-------------|
| **E1-Action** | How many times has an atomic action occurred so far? |
| **E1-Transit** | How many scene transitions have occurred so far? |
| **E2-Episode** | How many activity segments have occurred so far? |
| **E2-Periodic** | How many complete cycles of a periodic action so far? |
## Dataset Summary
- **Total Videos**: 406 source videos (generating 4,574 clipped segments)
- **Total Size**: ~80 GB
- **Annotations**: 1,000 counting questions with 4,576 streaming query points and 10,071 frame-by-frame annotations.
- **Sources**: YouTube, ARKitScenes, ScanNet, ScanNet++, Ego4D, RoomTour3D, CODa, OmniWorld, and physics simulations.
## Usage
### Download via CLI
You can download the dataset using the `huggingface-cli`:
```bash
huggingface-cli download buaaplay/VCBench --repo-type dataset --local-dir data/videos
```
The `chunkedVideos/` directory contains 4,576 video clips (one per query point), each truncated to the query timestamp.
### Evaluation
To compute metrics (GPA, MoC, UDA) on results using the official evaluation scripts:
```bash
# Compute metrics on provided results
python eval/compute_metrics.py results/vcbench_gemini3flash_unified.jsonl data/vcbench_eval.jsonl
```
## Citation
```bibtex
@article{vcbench2025,
title={VCBench: A Streaming Counting Benchmark for Spatial-Temporal State Maintenance in Long Videos},
author={Liu, Pengyiang and Shi, Zhongyue and Hao, Hongye and Fu, Qi and Bi, Xueting and Zhang, Siwei and Hu, Xiaoyang and Wang, Zitian and Huang, Linjiang and Liu, Si},
year={2026}
}
```
## License
This dataset and code are released under [CC BY 4.0](https://creativecommons.org/licenses/by/4.0/).