File size: 4,163 Bytes
809abcc
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
cc670e4
 
809abcc
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
cc670e4
 
809abcc
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
---
license: mit
task_categories:
- video-classification
- question-answering
language:
- en
tags:
- video-understanding
- temporal-reasoning
- counting
- benchmark
size_categories:
- 1K<n<10K
---

# VCBench: Clipped Videos Dataset

## Dataset Description

This dataset contains **4,574 clipped video segments** from the VCBench (Video Counting Benchmark), designed for evaluating spatial-temporal state maintenance capabilities in video understanding models.

### Dataset Summary

- **Total Videos**: 4,574 clips
- **Total Size**: ~80 GB
- **Video Format**: MP4 (H.264)
- **Categories**: 8 subcategories across object counting and event counting tasks

### Categories

**Object Counting (2,297 clips)**:
- `O1-Snap`: Current-state snapshot (252 clips)
- `O1-Delta`: Current-state delta (98 clips)
- `O2-Unique`: Global unique counting (1,869 clips)
- `O2-Gain`: Windowed gain counting (78 clips)

**Event Counting (2,277 clips)**:
- `E1-Action`: Instantaneous action (1,281 clips)
- `E1-Transit`: State transition (205 clips)
- `E2-Periodic`: Periodic action (280 clips)
- `E2-Episode`: Episodic segment (511 clips)

## File Naming Convention

### Multi-query clips
Format: `{category}_{question_id}_{query_index}.mp4`

Example: `e1action_0000_00.mp4`, `e1action_0000_01.mp4`

### Single-query clips
Format: `{category}_{question_id}.mp4`

Example: `o1delta_0007.mp4`, `o2gain_0000.mp4`

## Video Properties

- **Encoding**: H.264 (using `-c copy` for lossless clipping)
- **Frame Rates**: Preserved from source (3fps, 24fps, 25fps, 30fps, 60fps)
- **Duration Accuracy**: ±0.1s from annotation timestamps
- **Quality**: Original quality maintained (no re-encoding)

## Source Datasets

Videos are clipped from multiple source datasets:
- YouTube walking tours and sports videos
- RoomTour3D (indoor navigation)
- Ego4D (first-person view)
- ScanNet, ScanNetPP, ARKitScenes (3D indoor scenes)
- TOMATO, CODa, OmniWorld (temporal reasoning)
- Simulated physics videos

## Usage

### Loading with Python

```python
from huggingface_hub import hf_hub_download
import cv2

# Download a specific video
video_path = hf_hub_download(
    repo_id="YOUR_USERNAME/VCBench",
    filename="e1action_0000_00.mp4",
    repo_type="dataset"
)

# Load with OpenCV
cap = cv2.VideoCapture(video_path)
```

### Batch Download

```bash
# Install huggingface-cli
pip install huggingface_hub

# Download entire dataset
huggingface-cli download YOUR_USERNAME/VCBench --repo-type dataset --local-dir ./vcbench_videos
```

## Annotations

For complete annotations including questions, query points, and ground truth answers, please refer to the original VCBench repository:
- Object counting annotations: `object_count_data/*.json`
- Event counting annotations: `event_counting_data/*.json`

Each annotation file contains:
- `id`: Question identifier
- `source_dataset`: Original video source
- `video_path`: Original video filename
- `question`: Counting question
- `query_time` or `query_points`: Timestamp(s) for queries
- `count`: Ground truth answer(s)

## Quality Validation

All videos have been validated for:
- ✓ Duration accuracy (100% within ±0.1s)
- ✓ Frame rate preservation (original fps maintained)
- ✓ No frame drops or speed changes
- ✓ Lossless clipping (no re-encoding artifacts)

## Citation

If you use this dataset, please cite the VCBench paper:

```bibtex
@article{vcbench2026,
  title={VCBench: A Streaming Counting Benchmark for Spatial-Temporal State Maintenance},
  author={[Authors]},
  journal={[Journal/Conference]},
  year={2026}
}
```

## License

MIT License - See LICENSE file for details.

## Dataset Statistics

| Category | Clips | Avg Duration | Total Size |
|----------|-------|--------------|------------|
| O1-Snap | 252 | ~2min | ~4.3 GB |
| O1-Delta | 98 | ~1min | ~1.7 GB |
| O2-Unique | 1,869 | ~3min | ~32 GB |
| O2-Gain | 78 | ~1min | ~1.3 GB |
| E1-Action | 1,281 | ~4min | ~28 GB |
| E1-Transit | 205 | ~2min | ~3.5 GB |
| E2-Periodic | 280 | ~3min | ~8.7 GB |
| E2-Episode | 511 | ~2min | ~4.8 GB |
| **Total** | **4,574** | - | **~80 GB** |

## Contact

For questions or issues, please open an issue in the dataset repository.