File size: 6,632 Bytes
7599d73
 
59f0912
 
 
 
 
 
 
 
 
7599d73
 
372a1b9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
475348b
372a1b9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2164085
51b1b30
 
 
 
372a1b9
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
---
license: cc-by-4.0
configs:
- config_name: capnav_v0
  data_files:
  - split: train
    path: capnav_v0_with_answer.parquet
- config_name: agent_profiles
  data_files:
  - split: train
    path: agent_profiles.parquet
---

# CapNav: Benchmarking Vision Language Models on Capability-conditioned Indoor Navigation

CapNav is a benchmark for Vision Language Models evaluating on **capability-conditioned navigation reasoning** in indoor environments.
The benchmark focuses on determining whether an embodied agent with specific physical constraints and abilities
can navigate from a start area to a target area within a complex indoor scene.

This repository contains **two complementary datasets**:
1. The main **CapNav Benchmark Dataset**, which provides navigation questions paired with scene graph nodes and agent identifiers.
2. The **Agent Profiles Dataset**, which defines the physical dimensions and capabilities of each agent type.

---

## Quickstart

```python
from datasets import load_dataset

capnav = load_dataset("RichardC0216/CapNav", "capnav_v0", split="train")
agents = load_dataset("RichardC0216/CapNav", "agent_profiles", split="train")

print(len(capnav), capnav.column_names)
print(len(agents), agents.column_names)
```

## Dataset Overview

### 1. CapNav Benchmark Dataset (`capnav_v0`)
- **File:** `capnav_v0_with_answer.parquet`
- **Format:** Parquet
- **Split:** `train`
- **Rows:** 2,330 (question × agent combinations)
This dataset contains natural language navigation questions grounded in indoor scenes.
Each question is evaluated under a specific agent profile, enabling systematic analysis
of how agent capabilities affect navigation feasibility.

#### Data Fields
Below one can find the description of each field in the dataset.
- **`question_id` (str)**  
  Unique identifier for each question–agent pair.
- **`question` (str)**  
  Natural language navigation question describing whether an agent can move from a start area to a goal area.
- **`scene_id` (str)**  
  Identifier of the indoor scene (e.g., `HM3D00000`, `MP3D00027`).
- **`scene_type` (str)**  
  High-level category of the scene (e.g., `home`).
- **`scene_nodes` (list[dict])**  
  List of scene graph nodes available in the environment.  
  Each node contains:
  - `node_id` (str): Unique identifier of the node  
  - `name` (str): Human-readable node name (e.g., room or area label)
- **`agent_name` (str)**  
  Name of the agent under which the navigation question should be evaluated.
  This field links to an entry in the Agent Profiles Dataset.
- **`answer` ((bool))**  
  Binary ground-truth navigability label (`True` / `False`) for the (`question`, `scene`, `agent`) triple.
  > Note: `answer` only provides a **binary feasibility** signal. For scene graphs, route-level traversability,
  > and detailed ground-truth rationale, please refer to the full annotations in `ground_truth/`(see below).

---

### 2. Agent Profiles Dataset
- **File:** `agent_profiles.parquet`
- **Format:** Parquet
- **Rows:** One per agent type
This dataset defines the physical properties and functional capabilities of each agent
used in the CapNav benchmark.

#### Data Fields
- **`agent_name` (str)**  
  Unique identifier of the agent (e.g., `HUMAN`, `WHEELCHAIR`, `SWEEPER`).
- **`body_shape` (str)**  
  Abstract geometric representation of the agent’s body (e.g., `cylinder`, `box`).
- **`body_height_m` (float)**  
  Agent height in meters.
- **`body_width_m` (float)**  
  Agent width in meters.
- **`body_depth_m` (float, optional)**  
  Agent depth in meters.  
  May be `null` for rotationally symmetric agents.
- **`max_vertical_cross_height_m` (float)**  
  Maximum vertical obstacle height the agent can cross.
- **`can_go_up_or_down_stairs` (bool)**  
  Indicates whether the agent is capable of traversing stairs.
- **`can_operate_elevator` (bool)**  
  Indicates whether the agent can operate and use elevators.
- **`can_open_the_door` (bool)**  
  Indicates whether the agent can open doors independently.
- **`description` (str)**  
  Natural language description summarizing the agent’s physical and functional characteristics.
---

## Full Ground Truth Annotations (Graphs and Traversability)
This repository also includes a `ground_truth/` directory that provides the
**complete ground-truth annotations** used to derive the binary `answer` labels
in the main CapNav benchmark.

While the benchmark dataset exposes only a binary navigability outcome for each
`(question, scene, agent)` triple, the annotations in `ground_truth/` contain
the underlying **structural and traversability information** that supports
more detailed inspection and analysis.

Specifically, this directory includes:
- **Scene graph annotations** (`ground_truth/graphs/`)  
  Manually constructed graph representations for each indoor environment,
  encoding nodes, connectivity, and true spatial structure.
- **Agent-conditioned traversability annotations** (`ground_truth/traverse/`)  
  Route-level annotations that specify whether and why a path is traversable
  for a given agent, including agent-specific constraints and failure rationales.

The binary `answer` field in the benchmark dataset is a **distilled signal**
derived from these annotations. Researchers interested in path validity,
failure cases, or the reasons behind infeasible navigation decisions should
refer to the files in `ground_truth/`.

Detailed descriptions of file formats and annotation semantics can be found in:
- `ground_truth/README.md`

The annotation pipeline, tooling, and quality control procedures are documented
in the CapNav GitHub repository:
- https://github.com/makeabilitylab/CapNav  
  (see the annotation code and documentation)

---

## Intended Use
The CapNav benchmark is intended for:
- Evaluating multimodal or vision-language models on embodied navigation reasoning
- Studying how physical capabilities affect navigation feasibility
- Benchmarking agent-aware reasoning and planning systems
The dataset is **not** intended to provide low-level control signals or precise geometric navigation paths.

---

## License
- **Dataset:** Creative Commons Attribution 4.0 International (CC BY 4.0)
---

## Citation

If you find it useful for your research and applications, please cite our paper using this BibTeX:
```bibtex
@article{su2026capnav,
    title={CapNav: Benchmarking Vision Language Models on Capability-conditioned Indoor Navigation},
    author={Su, Xia and Chen, Ruiqi and Liu, Benlin and Ma, Jingwei and Di, Zonglin and Krishna, Ranjay and Froehlich, Jon},
    journal={arXiv preprint arXiv:2602.18424},
    year={2026}
}
```