Datasets:

scottzhao commited on
Commit
9800a9a
·
verified ·
1 Parent(s): ebb0c24

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -7
README.md CHANGED
@@ -78,11 +78,11 @@ dataset_info:
78
 
79
 
80
  <p align="center">
81
- <img src="images/TDBench.jpg" width="500"></a>
82
  </p>
83
- <p align="justify"> <b> Figure</b>: 8 Representative VLMs on 10 dimensions in TDBench
84
 
85
- **<p align="justify"> Abstract:** The rapid emergence of Vision-Language Models (VLMs) has significantly advanced multimodal understanding, enabling applications in scene comprehension and visual reasoning. While these models have been primarily evaluated and developed for front-view image understanding, their capabilities in interpreting top-down images have received limited attention, partly due to the scarcity of diverse top-down datasets and the challenges in collecting such data. In contrast, top-down vision provides explicit spatial overviews and improved contextual understanding of scenes, making it particularly valuable for tasks like autonomous navigation, aerial imaging, and spatial planning. In this work, we address this gap by introducing TDBench, a comprehensive benchmark for VLMs in top-down image understanding. TDBench is constructed from public top-down view datasets and high-quality simulated images, including diverse real-world and synthetic scenarios. TDBench consists of visual question-answer pairs across ten evaluation dimensions of image understanding. Moreover, we conduct four case studies that commonly happen in real-world scenarios but are less explored. By revealing the strengths and limitations of existing VLM through evaluation results, we hope TDBench to provide insights for motivating future research. </p>
86
 
87
 
88
  ## 📢 Latest Updates
@@ -110,7 +110,7 @@ dataset_info:
110
  </p>
111
 
112
 
113
- > <p align="justify"> <b> <span style="color: blue;">Table</span></b>: Overview performance of 8 open source VLMs and 6 propriety VLMs on 10 dimensions with RotationalEval method. </p>
114
 
115
  <hr />
116
 
@@ -222,11 +222,9 @@ If you find our work and this repository useful, please consider giving our repo
222
  ```
223
 
224
 
225
-
226
-
227
  ## 📨 Contact
228
  If you have any questions, please create an issue on this repository or contact at kh3119@columbia.edu or
229
  mz2866@columbia.edu.
230
 
231
  ---
232
- [<img src="images/ICSL_Logo.png" height="80">](http://icsl.ee.columbia.edu/)
 
78
 
79
 
80
  <p align="center">
81
+ <img src="images/TDBench.jpg" width="600"></a>
82
  </p>
83
+ <p align="center"> 8 Representative VLMs on 10 dimensions in TDBench </p>
84
 
85
+ **Abstract:** The rapid emergence of Vision-Language Models (VLMs) has significantly advanced multimodal understanding, enabling applications in scene comprehension and visual reasoning. While these models have been primarily evaluated and developed for front-view image understanding, their capabilities in interpreting top-down images have received limited attention, partly due to the scarcity of diverse top-down datasets and the challenges in collecting such data. In contrast, top-down vision provides explicit spatial overviews and improved contextual understanding of scenes, making it particularly valuable for tasks like autonomous navigation, aerial imaging, and spatial planning. In this work, we address this gap by introducing TDBench, a comprehensive benchmark for VLMs in top-down image understanding. TDBench is constructed from public top-down view datasets and high-quality simulated images, including diverse real-world and synthetic scenarios. TDBench consists of visual question-answer pairs across ten evaluation dimensions of image understanding. Moreover, we conduct four case studies that commonly happen in real-world scenarios but are less explored. By revealing the strengths and limitations of existing VLM through evaluation results, we hope TDBench to provide insights for motivating future research.
86
 
87
 
88
  ## 📢 Latest Updates
 
110
  </p>
111
 
112
 
113
+ <p align="center"> Overview performance of 8 open source VLMs and 6 propriety VLMs on 10 dimensions with RotationalEval method. </p>
114
 
115
  <hr />
116
 
 
222
  ```
223
 
224
 
 
 
225
  ## 📨 Contact
226
  If you have any questions, please create an issue on this repository or contact at kh3119@columbia.edu or
227
  mz2866@columbia.edu.
228
 
229
  ---
230
+ [<img src="images/ICSL_Logo.png" width="500"/>](http://icsl.ee.columbia.edu/)