Datasets:

Modalities:
Image
Text
Formats:
arrow
ArXiv:
Libraries:
Datasets
License:
mgholami commited on
Commit
3f7106f
·
verified ·
1 Parent(s): 940b0a4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -15
README.md CHANGED
@@ -2,9 +2,7 @@
2
  license: cc-by-4.0
3
  ---
4
  <div align="center">
5
- <h1>Spatial Reasoning with Vision-Language Models in Ego-Centric Multi-view Scenes (Code Comes Soon!)</h1>
6
- <p><i>Benchmarking and Improving 3D Spatial Reasoning in Vision-Language Models</i></p>
7
-
8
  <a href="https://arxiv.org/abs/2509.06266" target="_blank">
9
  <img alt="arXiv" src="https://img.shields.io/badge/arXiv-red?logo=arxiv" height="20" />
10
  </a>
@@ -17,19 +15,9 @@ license: cc-by-4.0
17
  </div>
18
 
19
  ---
20
- ![Sample](figs/Fig2_v4.png)
21
-
22
- ### 📌 Key Highlights
23
-
24
- - 📊 **Ego3D-Bench**: A benchmark of **8,600+ human-verified QA pairs** for evaluating VLMs in **ego-centric, multi-view outdoor environments**.
25
- - 🧠 **Ego3D-VLM**: A **post-training framework** that builds cognitive maps from global 3D coordinates, achieving **+12% QA accuracy** and **+56% distance estimation** improvements.
26
- - 🚀 **Impact**: Together, Ego3D-Bench and Ego3D-VLM move VLMs closer to **human-level 3D spatial understanding** in real-world settings.
27
-
28
- ---
29
-
30
 
31
- ### ⚖️ **Ego3D-Bench**
32
- Benchmark Overview: We introduce Ego3D-Bench, a benchmark designed to evaluate the spatial understanding of VLMs in ego-centric multi-view scenarios. Images are collected from three different datasets: NuScenes, Argoverse, and Waymo. Questions are designed to require cross-view reseasoning. We define question from the ego-perspective and from the perspective of objects in the scene. To clearly indicate the perspective of each question, we categorize them into ego-centric or object-centric. In total we have 10 questions: 8 multi-choice QAs and 2 exact number QAs. Figure
33
 
34
  ![Sample](figs/Fig5_v2.png)
35
 
 
2
  license: cc-by-4.0
3
  ---
4
  <div align="center">
5
+ <h1>Spatial Reasoning with Vision-Language Models in Ego-Centric Multi-view Scenes</h1>
 
 
6
  <a href="https://arxiv.org/abs/2509.06266" target="_blank">
7
  <img alt="arXiv" src="https://img.shields.io/badge/arXiv-red?logo=arxiv" height="20" />
8
  </a>
 
15
  </div>
16
 
17
  ---
 
 
 
 
 
 
 
 
 
 
18
 
19
+ ### ⚖️ **Ego3D-Bench Overview**
20
+ We introduce Ego3D-Bench, a benchmark designed to evaluate the spatial understanding of VLMs in ego-centric multi-view scenarios. Images are collected from three different datasets: NuScenes, Argoverse, and Waymo. Questions are designed to require cross-view reseasoning. We define question from the ego-perspective and from the perspective of objects in the scene. To clearly indicate the perspective of each question, we categorize them into ego-centric or object-centric. In total we have 10 questions: 8 multi-choice QAs and 2 exact number QAs. Figure
21
 
22
  ![Sample](figs/Fig5_v2.png)
23