---
dataset_info:
features:
- name: question_id
dtype: string
- name: question
dtype: string
- name: answer
dtype: string
- name: query_image_0
dtype: image
- name: query_image_1
dtype: image
- name: query_image_2
dtype: image
- name: query_image_3
dtype: image
- name: query_image_4
dtype: image
- name: query_image_5
dtype: image
- name: query_image_6
dtype: image
- name: query_image_7
dtype: image
- name: answer_image_a
dtype: image
- name: answer_image_b
dtype: image
- name: answer_image_c
dtype: image
- name: answer_image_d
dtype: image
- name: answer_image_e
dtype: image
- name: answer_image_f
dtype: image
splits:
- name: test
num_bytes: 12321743.0
num_examples: 50
download_size: 10640175
dataset_size: 12321743.0
configs:
- config_name: default
data_files:
- split: test
path: data/test-*
---
# Dataset Card for "IQ50"
# Large-scale Multi-modality Models Evaluation Suite
> Accelerating the development of large-scale multi-modality models (LMMs) with `lmms-eval`
🏠 [Homepage](https://lmms-lab.github.io/) | 📚 [Documentation](docs/README.md) | 🤗 [Huggingface Datasets](https://huggingface.co/lmms-lab)
# This Dataset
This is a formatted version of [IQ50](https://github.com/microsoft/unilm/issues/1265). It is used in our `lmms-eval` pipeline to allow for one-click evaluations of large multi-modality models.
```
@article{huang2023language,
title={Language is not all you need: Aligning perception with language models},
author={Huang, Shaohan and Dong, Li and Wang, Wenhui and Hao, Yaru and Singhal, Saksham and Ma, Shuming and Lv, Tengchao and Cui, Lei and Mohammed, Owais Khan and Liu, Qiang and others},
journal={arXiv preprint arXiv:2302.14045},
volume={1},
number={2},
pages={3},
year={2023}
}
```
[More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)