Datasets:
File size: 6,110 Bytes
65541e9 1272358 65541e9 2e5e955 65541e9 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 | ---
license: apache-2.0
task_categories:
- visual-question-answering
- image-text-to-text
language:
- en
tags:
- spatial-reasoning
- multi-hop
- grounding
- vision-language
- benchmark
- VQA
- bounding-box
pretty_name: MultihopSpatial
size_categories:
- 10K<n<100K
configs:
- config_name: default
data_files:
- split: train
path: data/multihop_train_6791.json
- split: test
path: data/multihop_test_4500.json
---
# MultihopSpatial: Multi-hop Compositional Spatial Reasoning Benchmark for Vision-Language Models
<p align="center">
<img src="teaser_2.png" width="100%" alt="MultihopSpatial Benchmark Overview">
</p>
<p align="center">
<a href="https://youngwanlee.github.io/multihopspatial"><b>Project Page</b></a> |
<a href="https://arxiv.org/abs/2603.18892"><b>Paper</b></a> |
<a href="https://huggingface.co/etri-vilab/MultiHopSpatial-Qwen3-VL-4B-Instruct"><b>Model</b></a>
</p>
## Overview
**MultihopSpatial** is a benchmark designed to evaluate whether vision-language models (VLMs) demonstrate robustness in **multi-hop compositional spatial reasoning**. Unlike existing benchmarks that only assess single-step spatial relations, MultihopSpatial features queries with **1 to 3 reasoning hops** paired with **visual grounding evaluation**, exposing a critical blind spot: models achieving high multiple-choice accuracy often lack proper spatial localization.
All 4,500 benchmark QA pairs and bounding boxes are **strictly annotated by ten trained human experts** with an inter-rater agreement of 90% (Krippendorff's α = 0.90).
## Key Features
- **Multi-hop Composition**: Tests 1-hop, 2-hop, and 3-hop sequential spatial reasoning, mirroring real-world embodied AI needs.
- **Grounded Evaluation**: Addresses the "lucky guess" problem — models must both select the correct answer AND localize it via bounding box (Acc@50IoU).
- **Perspective-taking**: Includes both ego-centric and exo-centric viewpoints.
- **Three Spatial Categories**: Attribute (ATT), Position (POS), and Relation (REL), composable into multi-hop questions.
- **Training Data**: MultihopSpatial-Train (6,791 samples) supports post-training via reinforcement learning (e.g., GRPO).
## Dataset Statistics
### MultihopSpatial
| | **Ego-centric** | **Exo-centric** | **Total** |
|---|:---:|:---:|:---:|
| **1-hop** | 750 | 750 | 1,500 |
| **2-hop** | 750 | 750 | 1,500 |
| **3-hop** | 750 | 750 | 1,500 |
| **Total** | 2,250 | 2,250 | **4,500** |
### Spatial Reasoning Compositions
| **Hop** | **Categories** |
|---|---|
| 1-hop | ATT, POS, REL |
| 2-hop | ATT+POS, ATT+REL, POS+REL |
| 3-hop | ATT+POS+REL |
## Data Fields
| Field | Type | Description |
|---|---|---|
| `id` | `int` | Unique sample identifier |
| `image_path` | `string` | Image filename (e.g., `000000303219.jpg` or `01ce4fd6-..._002114.jpeg`) |
| `image_resolution` | `string` | Image resolution in `WxH` format |
| `view` | `string` | Viewpoint type: `"ego"` (ego-centric) or `"exo"` (exo-centric) |
| `hop` | `string` | Reasoning complexity: `"1hop"`, `"2hop"`, or `"3hop"` |
| `question` | `string` | The spatial reasoning question in plain text with multiple-choice options |
| `question_tag` | `string` | Same question with spatial reasoning type tags (`<ATT>`, `<POS>`, `<REL>`) annotated inline |
| `answer` | `string` | The correct answer choice (e.g., `"(c) frame of the reed picture"`) |
| `bbox` | `list[float]` | Bounding box `[x, y, width, height]` of the answer object in pixel coordinates |
### `question` vs `question_tag`
- **`question`**: Clean natural language question, e.g.,
> *"From the perspective of the woman holding the remote control, which object is on her right?"*
- **`question_tag`**: Same question with spatial reasoning tags marking which type of reasoning each part requires, e.g.,
> *"From the perspective of the woman holding the remote control, which object is **\<POS\>on her right\</POS\>**?"*
Tags: `<ATT>...</ATT>` (Attribute), `<POS>...</POS>` (Position), `<REL>...</REL>` (Relation)
## Data Structure
```
MultihopSpatial/
├── README.md
├── teaser_2.png
├── data/
│ ├── multihop_test_4500.json
│ ├── multihop_train_6791.json
│ └── images/
│ ├── 000000303219.jpg
│ ├── 000000022612.jpg
│ ├── 01ce4fd6-197a-4792-8778-775b03780369_002114.jpeg
│ └── ...
```
## Usage
```python
from datasets import load_dataset
dataset = load_dataset("etri-vilab/MultihopSpatial")
# Access splits
test_data = dataset["test"]
train_data = dataset["train"]
# Example
sample = test_data[0]
print(sample["question"])
# "From the perspective of the woman holding the remote control, which object is on her right? ..."
print(sample["answer"])
# "(c) frame of the reed picture"
print(sample["bbox"])
# [52.86, 38.7, 70.95, 97.83]
print(sample["hop"])
# "1hop"
```
## Image Sources & License
| Component | License | Source |
|---|---|---|
| **VQA Annotations** (questions, answers, bounding boxes) | [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0) | MultihopSpatial (this work) |
| **COCO Images** | [COCO Terms of Use](https://cocodataset.org/#termsofuse) | [MS-COCO](https://cocodataset.org/) |
| **PACO-Ego4D Images** | [Ego4D License](https://ego4ddataset.com/ego4d-data/license/) | [PACO](https://github.com/facebookresearch/paco) / [Ego4D](https://ego4ddataset.com/) |
> The images retain their original licenses. Our VQA annotations (questions, answers, bounding boxes, and metadata) are released under the Apache 2.0 License.
## Citation
```bibtex
@article{lee2025multihopspatial,
title={MultihopSpatial: Multi-hop Compositional Spatial Reasoning Benchmark for Vision-Language Models},
author={Lee, Youngwan and Jang, Soojin and Cho, Yoorhim and Lee, Seunghwan and Lee, Yong-Ju and Hwang, Sung Ju},
journal={arXiv preprint arXiv:2603.18892},
year={2025}
}
```
## Contact
For questions or issues, please visit the [Project Page](https://youngwanlee.github.io/multihopspatial_private) or open an issue in this repository.
|