VPoS / README.md
ghazishazan's picture
Update dataset card: Correct license, task category, and add descriptive tags (#3)
b25c3fe verified
---
license: cc-by-nc-4.0
task_categories:
- image-segmentation
tags:
- video
- multimodal
- segmentation
- pointing
- spatio-temporal-grounding
- robotics
- autonomous-driving
- cell-tracking
- egocentric-vision
- gui-interaction
---
# VPoS-Bench: Video Pointing and Segmentation Benchmark
**VPoS-Bench** is a challenging out-of-distribution benchmark designed to evaluate the spatio-temporal pointing and reasoning capabilities of video-language models. It covers a diverse set of five real-world application domains, with fine-grained point-level and segmentation annotations that enable robust evaluation of multimodal models under realistic, temporally complex scenarios.
> **Webpage**: [VideoMolmo](https://mbzuai-oryx.github.io/VideoMolmo/)
> **Paper**: [VideoMolmo: Spatio-Temporal Grounding meets Pointing](https://arxiv.org/pdf/2506.05336)
> **Model**: [VideoMolmo on Hugging Face](https://huggingface.co/ghazishazan/VideoMolmo)
> **Code**: [VideoMolmo on Github](https://github.com/mbzuai-oryx/VideoMolmo)
---
## 🌍 Benchmark Overview
VPoS-Bench tests the **generalization** of models in five diverse real-world scenarios:
1. **Cell Tracking**
Track the trajectory of biological entities (e.g., nuclei or cells) across microscopy video frames.
> Applications: developmental biology, disease modeling
2. **Egocentric Vision**
Identify and follow objects or hands in first-person camera footage.
> Applications: activity recognition, assistive tech
3. **Autonomous Driving**
Point to traffic participants (pedestrians, vehicles, lights) under varying conditions.
> Applications: self-driving systems, urban scene understanding
4. **Video-GUI Interaction**
Follow on-screen elements (e.g., cursors, buttons) across software interface recordings.
> Applications: AI-assisted UI navigation, screen agents
5. **Robotics**
Track manipulable objects or robotic end-effectors as they interact in structured environments.
> Applications: robot learning, manipulation planning
---
## 📁 Dataset Structure
The dataset is organized by domain. Each domain folder contains three subdirectories:
- `frames/` – Extracted video frames.
- `masks/` – Segmentation masks corresponding to frames.
- `annotations/` – JSON files containing text descriptions and point-level annotations.
```text
vpos-bench/
├── cell-tracking/
│ ├── frames/ # Extracted video frames (e.g., frame_0001.jpg, ...)
│ ├── masks/ # Segmentation masks per frame (optional)
│ └── annotations/ # Point coordinates + caption in JSON format
├── autonomous-driving/
...
---
├──
```
## 📁 Annotation Format
Each annotation is keyed by a unique video ID and consists of:
```json
{
"video_id": {
"caption": "natural language instruction here",
"frames": [
{
"frame_path": "domain/frames/video_id/frame_00001.jpg",
"mask_path": "domain/masks/video_id/0.png",
"points": [[x, y], ...]
},
{
"frame_path": "domain/frames/video_id/frame_00002.jpg",
"mask_path": "domain/masks/video_id/1.png",
"points": [[x, y], ...]
}
]
}
}
```