|
|
--- |
|
|
license: cc-by-4.0 |
|
|
--- |
|
|
<div align="center"> |
|
|
<h1>Spatial Reasoning with Vision-Language Models in Ego-Centric Multi-view Scenes</h1> |
|
|
<a href="https://arxiv.org/abs/2509.06266" target="_blank"> |
|
|
<img alt="arXiv" src="https://img.shields.io/badge/arXiv-red?logo=arxiv" height="20" /> |
|
|
</a> |
|
|
<a href="https://vbdi.github.io/Ego3D-Bench-webpage/" target="_blank"> |
|
|
<img alt="Website" src="https://img.shields.io/badge/🌎_Website-blue.svg" height="20" /> |
|
|
</a> |
|
|
<a href="https://github.com/vbdi/Ego3D-Bench" target="_blank"> |
|
|
<img alt="Code: Code" src="https://img.shields.io/badge/Code-100000?logo=github&logoColor=white" height="20" /> |
|
|
</a> |
|
|
</div> |
|
|
|
|
|
--- |
|
|
|
|
|
### ⚖️ **Ego3D-Bench Overview** |
|
|
We introduce Ego3D-Bench, a benchmark designed to evaluate the spatial understanding of VLMs in ego-centric multi-view scenarios. Images are collected from three different datasets: NuScenes, Argoverse, and Waymo. Questions are designed to require cross-view reseasoning. We define question from the ego-perspective and from the perspective of objects in the scene. To clearly indicate the perspective of each question, we categorize them into ego-centric or object-centric. In total we have 10 questions: 8 multi-choice QAs and 2 exact number QAs. Figure |
|
|
|
|
|
 |
|
|
|
|
|
--- |
|
|
|
|
|
### ⚖️ **Ego3D-Bench vs. other Spatial Reasoning Benchmarks:** |
|
|
<div align="center"> |
|
|
<img src="figs/benchmarks.png" width="600" height="100"> |
|
|
</div> |
|
|
|
|
|
--- |
|
|
|
|
|
📄 **Dataset Access and License Notice:** |
|
|
|
|
|
This dataset includes a subsample of the Waymo Open Dataset (WOD) and is governed by the Waymo Open Dataset License Agreement. |
|
|
Please review the full license terms at: https://waymo.com/open/terms |
|
|
|
|
|
🔒 **Access and Usage Conditions** |
|
|
|
|
|
- License Compliance: This dataset is derived from the Waymo Open Dataset (WOD). All use of this dataset must comply with the terms outlined in the WOD license. |
|
|
|
|
|
- Non-Commercial Use Only:This dataset is made available exclusively for non-commercial research purposes. Any commercial use is strictly prohibited. |
|
|
|
|
|
- Access Agreement: Requesting or accessing this dataset constitutes your agreement to the Waymo Open Dataset License. |
|
|
|
|
|
--- |
|
|
|
|
|
|
|
|
### 📌 Benchmarking on Ego3D-Bench: |
|
|
|
|
|
Refer to the GitHub page (https://github.com/vbdi/Ego3D-Bench) to perform benchmarking using this dataset. |
|
|
|
|
|
--- |
|
|
|
|
|
|
|
|
### Citation: |
|
|
If you find our paper and code useful in your research, please consider giving us a star ⭐ and citing our work 📝 :) |
|
|
|
|
|
``` |
|
|
@misc{gholami2025spatialreasoningvisionlanguagemodels, |
|
|
title={Spatial Reasoning with Vision-Language Models in Ego-Centric Multi-View Scenes}, |
|
|
author={Mohsen Gholami and Ahmad Rezaei and Zhou Weimin and Sitong Mao and Shunbo Zhou and Yong Zhang and Mohammad Akbari}, |
|
|
year={2025}, |
|
|
eprint={2509.06266}, |
|
|
archivePrefix={arXiv}, |
|
|
primaryClass={cs.CV}, |
|
|
url={https://arxiv.org/abs/2509.06266}, |
|
|
} |
|
|
``` |