--- license: cc-by-nc-4.0 --- # ๐Ÿงช Multimodal Benchmark This repository provides a benchmark suite for evaluating Multimodal Large Language Models (MLLMs) across a variety of visual-language tasks. --- ## ๐Ÿ“ Directory Structure ### `/data` This folder contains all benchmark images and task-specific JSON files. Each JSON file defines the input and expected output format for a given task. ### `/run` This folder includes example scripts that demonstrate how to run different MLLMs on the benchmark tasks. --- ## ๐Ÿ“„ Result Collection After inference, all task JSON outputs should be merged into a single file named `result.json`. Each entry in `result.json` includes a `response` field that stores the model's prediction. --- ## ๐Ÿ“Š Evaluation The predictions stored in `result.json` can be evaluated using `metric.py`. This script computes performance metrics by comparing the predicted responses with the reference answers. --- ## ๐Ÿ’ก Ad Understanding Task The **Ad Understanding** task requires an additional LLM-based preprocessing step before evaluation. An example of deploying a language model for this purpose is provided in [`gpt_judge.py`](./gpt_judge.py). ---