--- configs: - config_name: default data_files: - split: test path: - Objects.csv - Segmentations.json - visual_patterns.csv task_categories: - image-referring-segmentation - image-segmentation --- # PixMMVP Benchmark [Project Page](https://msiam.github.io/PixFoundationSeries/) | [Paper](https://huggingface.co/papers/2502.04192) | [GitHub](https://github.com/msiam/pixfoundation) The PixMMVP dataset augments the [MMVP](https://huggingface.co/datasets/MMVP/MMVP) benchmark with referring expressions and corresponding segmentation masks for the objects of interest in their respective questions within the original VQA task. The goal of this benchmark is to evaluate the pixel-level visual grounding and visual question answering capabilities of recent pixel-level MLLMs (e.g., OMG-Llava, Llava-G, GLAMM, and LISA). # Acknowledgements I acknowledge the use of MMVP dataset's images and questions/choices part of building this dataset, the original [MMVP](https://huggingface.co/datasets/MMVP/MMVP). # Citation Please cite the following work if you find the dataset useful: ```bibtex @article{siam2025pixfoundation, title={PixFoundation: Are We Heading in the Right Direction with Pixel-level Vision Foundation Models?}, author={Siam, Mennatullah}, journal={arXiv preprint arXiv:2502.04192}, year={2025} } ```