--- dataset_info: features: - name: Index dtype: int32 - name: Question dtype: string - name: Options dtype: string - name: Correct Answer dtype: string - name: image dtype: image splits: - name: train num_bytes: 3145724 num_examples: 300 download_size: 3063257 dataset_size: 3145724 configs: - config_name: default data_files: - split: train path: data/train-* license: mit task_categories: - visual-question-answering tags: - multimodal - vision-language - clip - benchmark --- # MMVP (Multimodal Visual Patterns) Benchmark This is a corrected version of the [MMVP benchmark](https://huggingface.co/datasets/MMVP/MMVP), re-hosted by [lmms-lab-eval](https://huggingface.co/lmms-lab-eval) for use with [lmms-eval](https://github.com/EvolvingLMMs-Lab/lmms-eval). ## Why this copy? The original `MMVP/MMVP` dataset was uploaded in `imagefolder` format, which only exposes the `image` column. The text annotations (`Question`, `Options`, `Correct Answer`, `Index`) from the accompanying `Questions.csv` were not loaded into the dataset, making it unusable for evaluation. This version reconstructs the complete dataset by merging the images with `Questions.csv` from the original repository. ## Ground Truth Corrections Based on verification in [lmms-eval issue #1018](https://github.com/EvolvingLMMs-Lab/lmms-eval/issues/1018) and the [original MMVP issue #30](https://github.com/tsb0601/MMVP/issues/30), we found that two pairs of samples had their answers swapped. The corrections are applied directly in this dataset: | Index | Question | Original GT | Corrected GT | Reason | |:-----:|:---------|:-----------:|:------------:|:-------| | 99 | Does the elephant have long or short tusks? | (a) Long | **(b) Short** | Image shows short tusks | | 100 | Does the elephant have long or short tusks? | (b) Short | **(a) Long** | Image shows long tusks | | 279 | Is the elderly person standing or sitting? | (a) Standing | **(b) Sitting** | Image shows person sitting on bench | | 280 | Is the elderly person standing or sitting? | (b) Sitting | **(a) Standing** | Image shows person standing | ## Dataset Structure | Field | Type | Description | |-------|------|-------------| | `Index` | int32 | 1-based sample index (1–300) | | `Question` | string | The visual question | | `Options` | string | Answer choices in format `(a) ... (b) ...` | | `Correct Answer` | string | Ground truth: `(a)` or `(b)` | | `image` | image | 224×224 RGB image | - **300 samples** organized as **150 pairs** - Each pair has the same question but opposite correct answers - Tests 9 visual patterns: orientation, direction, color, counting, etc. ## References - **Paper**: [Eyes Wide Shut? Exploring the Visual Shortcomings of Multimodal LLMs](https://arxiv.org/abs/2401.06209) - **Original Repository**: https://github.com/tsb0601/MMVP - **Original Dataset**: https://huggingface.co/datasets/MMVP/MMVP - **lmms-eval Task**: https://github.com/EvolvingLMMs-Lab/lmms-eval/tree/main/lmms_eval/tasks/mmvp ## Citation ```bibtex @inproceedings{tong2024eyes, title={Eyes Wide Shut? Exploring the Visual Shortcomings of Multimodal LLMs}, author={Tong, Shengbang and Liu, Zhuang and Zhai, Yuexiang and Ma, Yi and LeCun, Yann and Xie, Saining}, booktitle={CVPR}, year={2024} } ```