File size: 2,417 Bytes
9e3b395 f898e80 9e3b395 7453ed9 9e3b395 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 |
# CVQA for VLMEvalKit
- [Original dataset:](https://huggingface.co/datasets/afaji/cvqa) ported to [VLMEvalKit](https://github.com/open-compass/VLMEvalKit)
- From the original authors:
> CVQA is a culturally diverse multilingual VQA benchmark consisting of over 10,000 questions from 39 country-language pairs. The questions in CVQA are written in both the native languages and English, and are categorized into 10 diverse categories.
```
{'image': <PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=2048x1536 at 0x7C3E0EBEEE00>,
'ID': '5919991144272485961_0',
'Subset': "('Japanese', 'Japan')",
'Question': '写真に写っているキャラクターの名前は? ',
'Translated Question': 'What is the name of the object in the picture? ',
'Options': ['コスモ星丸', 'ミャクミャク', ' フリービー ', 'ハイバオ'],
'Translated Options': ['Cosmo Hoshimaru','MYAKU-MYAKU','Freebie ','Haibao'],
'Label': -1,
'Category': 'Objects / materials / clothing',
'Image Type': 'Self',
'Image Source': 'Self-open',
'License': 'CC BY-SA'
}
```
- To support VLMEvalKit, two TSV files were created to represent the two versions of CVQA:
1. The localised **(LOC)** version. The questions and answer options are in the subset's original native language. For evaluating with multilingual LLMs.
2. The english **(ENG)** version. Questions and answers are asked in translated English, although the topics of the question involve cultures other than English. For evaluating on LLMs trained primarily on English.
- TSV row data columns for **LOC** and **ENG** [VLMEvalKit](https://github.com/timothycdc/VLMEvalKit/blob/main/docs/en/Development.md):
- index (int, based on dataset order. Does not follow CVQA ids since they are of type str)
- image (base64)
- question
- A option
- B option
- C option
- D option
- l2-category (`Subset`)
- split (always called `test`)
## Info
- Proposed method of evaluation:
- Prompt the model to answer only with the correct option letter (one of `[A,B,C,D]`)
- Use regex or string search to locate the correct letter
- Alternatively, use LLM-as-a-judge to identify the correct answer letter. Although, this is a bit of an overkill.
- The original CVQA dataset numbers the options as `[0,1,2,3]`, however this has been changed to`[A,B,C,D]` to follow the VLMEvalKit standard. This shouldn't have much effect on performance. |