| | --- |
| | license: mit |
| | task_categories: |
| | - video-classification |
| | - question-answering |
| | language: |
| | - en |
| | tags: |
| | - video-understanding |
| | - temporal-reasoning |
| | - counting |
| | - benchmark |
| | size_categories: |
| | - 1K<n<10K |
| | --- |
| | |
| | # VCBench: Clipped Videos Dataset |
| |
|
| | ## Dataset Description |
| |
|
| | This dataset contains **4,574 clipped video segments** from the VCBench (Video Counting Benchmark), designed for evaluating spatial-temporal state maintenance capabilities in video understanding models. |
| |
|
| | ### Dataset Summary |
| |
|
| | - **Total Videos**: 4,574 clips |
| | - **Total Size**: ~80 GB |
| | - **Video Format**: MP4 (H.264) |
| | - **Categories**: 8 subcategories across object counting and event counting tasks |
| |
|
| | ### Categories |
| |
|
| | **Object Counting (2,297 clips)**: |
| | - `O1-Snap`: Current-state snapshot (252 clips) |
| | - `O1-Delta`: Current-state delta (98 clips) |
| | - `O2-Unique`: Global unique counting (1,869 clips) |
| | - `O2-Gain`: Windowed gain counting (78 clips) |
| |
|
| | **Event Counting (2,277 clips)**: |
| | - `E1-Action`: Instantaneous action (1,281 clips) |
| | - `E1-Transit`: State transition (205 clips) |
| | - `E2-Periodic`: Periodic action (280 clips) |
| | - `E2-Episode`: Episodic segment (511 clips) |
| |
|
| | ## File Naming Convention |
| |
|
| | ### Multi-query clips |
| | Format: `{category}_{question_id}_{query_index}.mp4` |
| |
|
| | Example: `e1action_0000_00.mp4`, `e1action_0000_01.mp4` |
| |
|
| | ### Single-query clips |
| | Format: `{category}_{question_id}.mp4` |
| |
|
| | Example: `o1delta_0007.mp4`, `o2gain_0000.mp4` |
| |
|
| | ## Video Properties |
| |
|
| | - **Encoding**: H.264 (using `-c copy` for lossless clipping) |
| | - **Frame Rates**: Preserved from source (3fps, 24fps, 25fps, 30fps, 60fps) |
| | - **Duration Accuracy**: ±0.1s from annotation timestamps |
| | - **Quality**: Original quality maintained (no re-encoding) |
| |
|
| | ## Source Datasets |
| |
|
| | Videos are clipped from multiple source datasets: |
| | - YouTube walking tours and sports videos |
| | - RoomTour3D (indoor navigation) |
| | - Ego4D (first-person view) |
| | - ScanNet, ScanNetPP, ARKitScenes (3D indoor scenes) |
| | - TOMATO, CODa, OmniWorld (temporal reasoning) |
| | - Simulated physics videos |
| |
|
| | ## Usage |
| |
|
| | ### Loading with Python |
| |
|
| | ```python |
| | from huggingface_hub import hf_hub_download |
| | import cv2 |
| | |
| | # Download a specific video |
| | video_path = hf_hub_download( |
| | repo_id="YOUR_USERNAME/VCBench", |
| | filename="e1action_0000_00.mp4", |
| | repo_type="dataset" |
| | ) |
| | |
| | # Load with OpenCV |
| | cap = cv2.VideoCapture(video_path) |
| | ``` |
| |
|
| | ### Batch Download |
| |
|
| | ```bash |
| | # Install huggingface-cli |
| | pip install huggingface_hub |
| | |
| | # Download entire dataset |
| | huggingface-cli download YOUR_USERNAME/VCBench --repo-type dataset --local-dir ./vcbench_videos |
| | ``` |
| |
|
| | ## Annotations |
| |
|
| | For complete annotations including questions, query points, and ground truth answers, please refer to the original VCBench repository: |
| | - Object counting annotations: `object_count_data/*.json` |
| | - Event counting annotations: `event_counting_data/*.json` |
| |
|
| | Each annotation file contains: |
| | - `id`: Question identifier |
| | - `source_dataset`: Original video source |
| | - `video_path`: Original video filename |
| | - `question`: Counting question |
| | - `query_time` or `query_points`: Timestamp(s) for queries |
| | - `count`: Ground truth answer(s) |
| |
|
| | ## Quality Validation |
| |
|
| | All videos have been validated for: |
| | - ✓ Duration accuracy (100% within ±0.1s) |
| | - ✓ Frame rate preservation (original fps maintained) |
| | - ✓ No frame drops or speed changes |
| | - ✓ Lossless clipping (no re-encoding artifacts) |
| |
|
| | ## Citation |
| |
|
| | If you use this dataset, please cite the VCBench paper: |
| |
|
| | ```bibtex |
| | @article{vcbench2026, |
| | title={VCBench: A Streaming Counting Benchmark for Spatial-Temporal State Maintenance}, |
| | author={[Authors]}, |
| | journal={[Journal/Conference]}, |
| | year={2026} |
| | } |
| | ``` |
| |
|
| | ## License |
| |
|
| | MIT License - See LICENSE file for details. |
| |
|
| | ## Dataset Statistics |
| |
|
| | | Category | Clips | Avg Duration | Total Size | |
| | |----------|-------|--------------|------------| |
| | | O1-Snap | 252 | ~2min | ~4.3 GB | |
| | | O1-Delta | 98 | ~1min | ~1.7 GB | |
| | | O2-Unique | 1,869 | ~3min | ~32 GB | |
| | | O2-Gain | 78 | ~1min | ~1.3 GB | |
| | | E1-Action | 1,281 | ~4min | ~28 GB | |
| | | E1-Transit | 205 | ~2min | ~3.5 GB | |
| | | E2-Periodic | 280 | ~3min | ~8.7 GB | |
| | | E2-Episode | 511 | ~2min | ~4.8 GB | |
| | | **Total** | **4,574** | - | **~80 GB** | |
| |
|
| | ## Contact |
| |
|
| | For questions or issues, please open an issue in the dataset repository. |
| |
|