MultihopSpatial / README.md
ywlee88's picture
Update README.md
2e5e955 verified
---
license: apache-2.0
task_categories:
- visual-question-answering
- image-text-to-text
language:
- en
tags:
- spatial-reasoning
- multi-hop
- grounding
- vision-language
- benchmark
- VQA
- bounding-box
pretty_name: MultihopSpatial
size_categories:
- 10K<n<100K
configs:
- config_name: default
data_files:
- split: train
path: data/multihop_train_6791.json
- split: test
path: data/multihop_test_4500.json
---
# MultihopSpatial: Multi-hop Compositional Spatial Reasoning Benchmark for Vision-Language Models
<p align="center">
<img src="teaser_2.png" width="100%" alt="MultihopSpatial Benchmark Overview">
</p>
<p align="center">
<a href="https://youngwanlee.github.io/multihopspatial"><b>Project Page</b></a> |
<a href="https://arxiv.org/abs/2603.18892"><b>Paper</b></a> |
<a href="https://huggingface.co/etri-vilab/MultiHopSpatial-Qwen3-VL-4B-Instruct"><b>Model</b></a>
</p>
## Overview
**MultihopSpatial** is a benchmark designed to evaluate whether vision-language models (VLMs) demonstrate robustness in **multi-hop compositional spatial reasoning**. Unlike existing benchmarks that only assess single-step spatial relations, MultihopSpatial features queries with **1 to 3 reasoning hops** paired with **visual grounding evaluation**, exposing a critical blind spot: models achieving high multiple-choice accuracy often lack proper spatial localization.
All 4,500 benchmark QA pairs and bounding boxes are **strictly annotated by ten trained human experts** with an inter-rater agreement of 90% (Krippendorff's α = 0.90).
## Key Features
- **Multi-hop Composition**: Tests 1-hop, 2-hop, and 3-hop sequential spatial reasoning, mirroring real-world embodied AI needs.
- **Grounded Evaluation**: Addresses the "lucky guess" problem — models must both select the correct answer AND localize it via bounding box (Acc@50IoU).
- **Perspective-taking**: Includes both ego-centric and exo-centric viewpoints.
- **Three Spatial Categories**: Attribute (ATT), Position (POS), and Relation (REL), composable into multi-hop questions.
- **Training Data**: MultihopSpatial-Train (6,791 samples) supports post-training via reinforcement learning (e.g., GRPO).
## Dataset Statistics
### MultihopSpatial
| | **Ego-centric** | **Exo-centric** | **Total** |
|---|:---:|:---:|:---:|
| **1-hop** | 750 | 750 | 1,500 |
| **2-hop** | 750 | 750 | 1,500 |
| **3-hop** | 750 | 750 | 1,500 |
| **Total** | 2,250 | 2,250 | **4,500** |
### Spatial Reasoning Compositions
| **Hop** | **Categories** |
|---|---|
| 1-hop | ATT, POS, REL |
| 2-hop | ATT+POS, ATT+REL, POS+REL |
| 3-hop | ATT+POS+REL |
## Data Fields
| Field | Type | Description |
|---|---|---|
| `id` | `int` | Unique sample identifier |
| `image_path` | `string` | Image filename (e.g., `000000303219.jpg` or `01ce4fd6-..._002114.jpeg`) |
| `image_resolution` | `string` | Image resolution in `WxH` format |
| `view` | `string` | Viewpoint type: `"ego"` (ego-centric) or `"exo"` (exo-centric) |
| `hop` | `string` | Reasoning complexity: `"1hop"`, `"2hop"`, or `"3hop"` |
| `question` | `string` | The spatial reasoning question in plain text with multiple-choice options |
| `question_tag` | `string` | Same question with spatial reasoning type tags (`<ATT>`, `<POS>`, `<REL>`) annotated inline |
| `answer` | `string` | The correct answer choice (e.g., `"(c) frame of the reed picture"`) |
| `bbox` | `list[float]` | Bounding box `[x, y, width, height]` of the answer object in pixel coordinates |
### `question` vs `question_tag`
- **`question`**: Clean natural language question, e.g.,
> *"From the perspective of the woman holding the remote control, which object is on her right?"*
- **`question_tag`**: Same question with spatial reasoning tags marking which type of reasoning each part requires, e.g.,
> *"From the perspective of the woman holding the remote control, which object is **\<POS\>on her right\</POS\>**?"*
Tags: `<ATT>...</ATT>` (Attribute), `<POS>...</POS>` (Position), `<REL>...</REL>` (Relation)
## Data Structure
```
MultihopSpatial/
├── README.md
├── teaser_2.png
├── data/
│ ├── multihop_test_4500.json
│ ├── multihop_train_6791.json
│ └── images/
│ ├── 000000303219.jpg
│ ├── 000000022612.jpg
│ ├── 01ce4fd6-197a-4792-8778-775b03780369_002114.jpeg
│ └── ...
```
## Usage
```python
from datasets import load_dataset
dataset = load_dataset("etri-vilab/MultihopSpatial")
# Access splits
test_data = dataset["test"]
train_data = dataset["train"]
# Example
sample = test_data[0]
print(sample["question"])
# "From the perspective of the woman holding the remote control, which object is on her right? ..."
print(sample["answer"])
# "(c) frame of the reed picture"
print(sample["bbox"])
# [52.86, 38.7, 70.95, 97.83]
print(sample["hop"])
# "1hop"
```
## Image Sources & License
| Component | License | Source |
|---|---|---|
| **VQA Annotations** (questions, answers, bounding boxes) | [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0) | MultihopSpatial (this work) |
| **COCO Images** | [COCO Terms of Use](https://cocodataset.org/#termsofuse) | [MS-COCO](https://cocodataset.org/) |
| **PACO-Ego4D Images** | [Ego4D License](https://ego4ddataset.com/ego4d-data/license/) | [PACO](https://github.com/facebookresearch/paco) / [Ego4D](https://ego4ddataset.com/) |
> The images retain their original licenses. Our VQA annotations (questions, answers, bounding boxes, and metadata) are released under the Apache 2.0 License.
## Citation
```bibtex
@article{lee2025multihopspatial,
title={MultihopSpatial: Multi-hop Compositional Spatial Reasoning Benchmark for Vision-Language Models},
author={Lee, Youngwan and Jang, Soojin and Cho, Yoorhim and Lee, Seunghwan and Lee, Yong-Ju and Hwang, Sung Ju},
journal={arXiv preprint arXiv:2603.18892},
year={2025}
}
```
## Contact
For questions or issues, please visit the [Project Page](https://youngwanlee.github.io/multihopspatial_private) or open an issue in this repository.