|
|
--- |
|
|
license: mit |
|
|
task_categories: |
|
|
- visual-question-answering |
|
|
- video-classification |
|
|
tags: |
|
|
- spatial-reasoning |
|
|
- vision-language |
|
|
- video-generation |
|
|
size_categories: |
|
|
- 10K<n<100K |
|
|
--- |
|
|
|
|
|
# VR-Bench Dataset |
|
|
|
|
|
VR-Bench is a benchmark dataset for evaluating spatial reasoning capabilities of Vision-Language Models (VLMs) and Video Generation Models. |
|
|
|
|
|
## Dataset Structure |
|
|
|
|
|
The dataset is split into two subsets: |
|
|
|
|
|
``` |
|
|
dataset_VR_split/ |
|
|
├── train/ # Training set (96 cases) |
|
|
│ ├── maze/ |
|
|
│ ├── maze3d/ |
|
|
│ ├── pathfinder/ |
|
|
│ ├── sokoban/ |
|
|
│ └── trapfield/ |
|
|
└── eval/ # Evaluation set (24 cases) |
|
|
├── maze/ |
|
|
├── maze3d/ |
|
|
├── pathfinder/ |
|
|
├── sokoban/ |
|
|
└── trapfield/ |
|
|
``` |
|
|
|
|
|
Each game directory contains: |
|
|
- `images/`: Initial state images (PNG) |
|
|
- `states/`: Game state metadata (JSON) |
|
|
- `videos/`: Solution trajectory videos (MP4) |
|
|
|
|
|
## Games |
|
|
|
|
|
- **Maze**: 2D grid-based navigation with walls |
|
|
- **TrapField**: 2D grid-based navigation with traps |
|
|
- **Sokoban**: Box-pushing puzzle game |
|
|
- **PathFinder**: Irregular maze with curved paths |
|
|
- **Maze3D**: 3D maze with vertical navigation |
|
|
|
|
|
## Usage |
|
|
|
|
|
### For Video Model Evaluation |
|
|
```python |
|
|
from datasets import load_dataset |
|
|
|
|
|
dataset = load_dataset("your-username/VR-Bench") |
|
|
train_data = dataset["train"] |
|
|
eval_data = dataset["eval"] |
|
|
``` |
|
|
|
|
|
Each video file shows the optimal solution trajectory for the corresponding game state. |
|
|
|
|
|
## Citation |
|
|
|
|
|
If you use this dataset, please cite: |
|
|
|
|
|
```bibtex |
|
|
@article{yang2025vrbench, |
|
|
title={Reasoning via Video: The First Evaluation of Video Models' Reasoning Abilities through Maze-Solving Tasks}, |
|
|
author={Cheng Yang and Haiyuan Wan and Yiran Peng and Xin Cheng and Zhaoyang Yu and Jiayi Zhang and Junchi Yu and Xinlei Yu and Xiawu Zheng and Dongzhan Zhou and Chenglin Wu}, |
|
|
journal={arXiv preprint arXiv:2511.15065}, |
|
|
year={2025} |
|
|
} |
|
|
``` |
|
|
|
|
|
## License |
|
|
|
|
|
MIT License |
|
|
|
|
|
|