Add comprehensive dataset card for ZoomEye evaluation data

#2
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +101 -0
README.md ADDED
@@ -0,0 +1,101 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ task_categories:
3
+ - image-text-to-text
4
+ license: cc-by-nc-4.0
5
+ tags:
6
+ - multimodal
7
+ - llm
8
+ - vision-language
9
+ - visual-reasoning
10
+ - tree-search
11
+ ---
12
+
13
+ # ZoomEye: Enhancing Multimodal LLMs with Human-Like Zooming Capabilities through Tree-Based Image Exploration
14
+
15
+ This repository contains the evaluation data for **ZoomEye**, a method presented in the paper [ZoomEye: Enhancing Multimodal LLMs with Human-Like Zooming Capabilities through Tree-Based Image Exploration](https://huggingface.co/papers/2411.16044).
16
+
17
+ ZoomEye proposes a training-free, model-agnostic tree search algorithm tailored for vision-level reasoning. It addresses the limitations of existing Multimodal Large Language Models (MLLMs) that operate on fixed visual inputs, especially when dealing with images containing numerous fine-grained elements. By treating an image as a hierarchical tree structure, ZoomEye enables MLLMs to simulate human-like zooming behavior, navigating from root to leaf nodes to gather detailed visual cues necessary for accurate decision-making.
18
+
19
+ This dataset supports the evaluation of MLLMs on a series of high-resolution benchmarks, demonstrating consistent performance improvements for various models.
20
+
21
+ * **Paper:** [https://huggingface.co/papers/2411.16044](https://huggingface.co/papers/2411.16044)
22
+ * **Project Page:** [https://szhanz.github.io/zoomeye/](https://szhanz.github.io/zoomeye/)
23
+ * **Code:** [https://github.com/om-ai-lab/ZoomEye](https://github.com/om-ai-lab/ZoomEye)
24
+
25
+ ## Evaluation Data Preparation
26
+
27
+ The core evaluation data (including V* Bench and HR-Bench) used in the ZoomEye paper has been packaged together.
28
+
29
+ 1. **Download Data**: The evaluation data is provided [here](https://huggingface.co/datasets/omlab/zoom_eye_data). After downloading, please unzip it. The path to the unzipped data is referred to as **`<anno path>`**.
30
+
31
+ 2. **[Optional] MME-RealWorld Benchmark**: If you wish to evaluate ZoomEye on the MME-RealWorld Benchmark, follow these steps:
32
+ * Follow the instructions in [this repository](https://github.com/yfzhang114/MME-RealWorld) to download the images.
33
+ * Extract the images to the `<anno path>/mme-realworld` directory.
34
+ * Place the `annotation_mme-realworld.json` file from [this link](https://huggingface.co/datasets/omlab/zoom_eye_data) into `<anno path>/mme-realworld`.
35
+
36
+ ### Folder Tree
37
+ The expected folder structure after preparation is as follows:
38
+
39
+ ```
40
+ zoom_eye_data
41
+ β”œβ”€β”€ hr-bench_4k
42
+ β”‚Β Β  └── annotation_hr-bench_4k.json
43
+ β”‚Β Β  └── images/
44
+ β”‚ └── some.jpg
45
+ β”‚Β Β  ...
46
+ β”œβ”€β”€ hr-bench_8k
47
+ β”‚Β Β  └── annotation_hr-bench_8k.json
48
+ β”‚Β Β  └── images/
49
+ β”‚ └── some.jpg
50
+ β”‚Β Β  ...
51
+ β”œβ”€β”€ vstar
52
+ β”‚Β Β  └── annotation_vstar.json
53
+ β”‚Β Β  └── direct_attributes/
54
+ β”‚ └── some.jpg
55
+ β”‚Β Β  ...
56
+ β”‚Β Β  └── relative_positions/
57
+ β”‚ └── some.jpg
58
+ β”‚Β Β  ...
59
+ β”œβ”€β”€ mme-realworld
60
+ β”‚Β Β  └── annotation_mme-realworld.json
61
+ β”‚Β Β  └── AutonomousDriving/
62
+ β”‚ └── MME-HD-CN/
63
+ β”‚ └── monitoring_images/
64
+ β”‚ └── ocr_cc/
65
+ β”‚ └── remote_sensing/
66
+ ...
67
+ ```
68
+
69
+ ## Sample Usage
70
+
71
+ ### 1. Run the Python Demo
72
+
73
+ We provide a demo file of Zoom Eye accepting any input Image-Question pair. The zoomed views of Zoom Eye will be saved into the demo folder.
74
+
75
+ ```bash
76
+ python ZoomEye/demo.py \
77
+ --model-path lmms-lab/llava-onevision-qwen2-7b-ov \
78
+ --input_image demo/demo.jpg \
79
+ --question "What is the color of the soda can?"
80
+ ```
81
+
82
+ ### 2. Run the Gradio Demo
83
+
84
+ We also provide a Gradio Demo. Run the script and open `http://127.0.0.1:7860/` in your browser.
85
+
86
+ ```bash
87
+ python gdemo_gradio.py
88
+ ```
89
+
90
+ ## Citation
91
+
92
+ If you find this repository helpful to your research, please cite our paper:
93
+
94
+ ```bibtex
95
+ @article{shen2024zoomeye,
96
+ title={ZoomEye: Enhancing Multimodal LLMs with Human-Like Zooming Capabilities through Tree-Based Image Exploration},
97
+ author={Shen, Haozhan and Zhao, Kangjia and Zhao, Tiancheng and Xu, Ruochen and Zhang, Zilun and Zhu, Mingwei and Yin, Jianwei},
98
+ journal={arXiv preprint arXiv:2411.16044},
99
+ year={2024}
100
+ }
101
+ ```