Datasets:

Modalities:
Image
Text
Languages:
English
Size:
< 1K
ArXiv:
Libraries:
Datasets
License:
BenchLMM / README.md
Ziruibest's picture
Update README.md
4902993
---
license: apache-2.0
task_categories:
- visual-question-answering
language:
- en
pretty_name: BenchLMM
size_categories:
- n<1K
---
# Dataset Card for BenchLMM
BenchLMM is a benchmarking dataset focusing on the cross-style visual capability of large multimodal models. It evaluates these models' performance in various visual contexts.
## Dataset Details
### Dataset Description
- **Curated by:** Rizhao Cai, Zirui Song, Dayan Guan, Zhenhao Chen, Xing Luo, Chenyu Yi, and Alex Kot.
- **Funded by :** Supported in part by the Rapid-Rich Object Search (ROSE) Lab of Nanyang Technological University and the NTU-PKU Joint Research Institute.
- **Shared by :** AIFEG.
- **Language(s) (NLP):** English.
- **License:** Apache-2.0.
### Dataset Sources
- **Repository:** [GitHub - AIFEG/BenchLMM](https://github.com/AIFEG/BenchLMM)
- **Paper :** Cai, R., Song, Z., Guan, D., et al. (2023). BenchLMM: Benchmarking Cross-style Visual Capability of Large Multimodal Models. arXiv:2312.02896.
## Uses
### Direct Use
The dataset can be used to benchmark large multimodal models, especially focusing on their capability to interpret and respond to different visual styles.
## Dataset Structure
- **Directory Structure:**
- `baseline/`: Baseline code for LLaVA and InstructBLIP.
- `evaluate/`: Python code for model evaluation.
- `evaluate_results/`: Evaluation results of baseline models.
- `jsonl/`: JSONL files with questions, image locations, and answers.
## Dataset Creation
### Curation Rationale
Developed to assess large multimodal models' performance in diverse visual contexts, helping to understand their capabilities and limitations.
### Source Data
#### Data Collection and Processing
The dataset consists of various visual questions and corresponding answers, structured to evaluate multimodal model performance.
## Bias, Risks, and Limitations
Users should consider the specific visual contexts and question types included in the dataset when interpreting model performance.
## Citation
**BibTeX:**
@misc{cai2023benchlmm,
title={BenchLMM: Benchmarking Cross-style Visual Capability of Large Multimodal Models},
author={Rizhao Cai and Zirui Song and Dayan Guan and Zhenhao Chen and Xing Luo and Chenyu Yi and Alex Kot},
year={2023},
eprint={2312.02896},
archivePrefix={arXiv},
primaryClass={cs.CV}
}
**APA:**
Cai, R., Song, Z., Guan, D., Chen, Z., Luo, X., Yi, C., & Kot, A. (2023). BenchLMM: Benchmarking Cross-style Visual Capability of Large Multimodal Models. arXiv preprint arXiv:2312.02896.
## Acknowledgements
This research is supported in part by the Rapid-Rich Object Search (ROSE) Lab of Nanyang Technological University and the NTU-PKU Joint Research Institute.