File size: 2,417 Bytes
9e3b395
 
f898e80
 
9e3b395
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7453ed9
 
9e3b395
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
# CVQA for VLMEvalKit
- [Original dataset:](https://huggingface.co/datasets/afaji/cvqa) ported to [VLMEvalKit](https://github.com/open-compass/VLMEvalKit)
- From the original authors: 
> CVQA is a culturally diverse multilingual VQA benchmark consisting of over 10,000 questions from 39 country-language pairs. The questions in CVQA are written in both the native languages and English, and are categorized into 10 diverse categories.
```
{'image': <PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=2048x1536 at 0x7C3E0EBEEE00>, 
 'ID': '5919991144272485961_0', 
 'Subset': "('Japanese', 'Japan')", 
 'Question': '写真に写っているキャラクターの名前は?  ', 
 'Translated Question': 'What is the name of the object in the picture? ', 
 'Options': ['コスモ星丸', 'ミャクミャク', ' フリービー ', 'ハイバオ'], 
 'Translated Options': ['Cosmo Hoshimaru','MYAKU-MYAKU','Freebie ','Haibao'], 
 'Label': -1, 
 'Category': 'Objects / materials / clothing', 
 'Image Type': 'Self', 
 'Image Source': 'Self-open', 
 'License': 'CC BY-SA'
}
```

- To support VLMEvalKit, two TSV files were created to represent the two versions of CVQA:
  1. The localised **(LOC)** version. The questions and answer options are in the subset's original native language. For evaluating with multilingual LLMs.
  2. The english **(ENG)** version. Questions and answers are asked in translated English, although the topics of the question involve cultures other than English. For evaluating on LLMs trained primarily on English.

- TSV row data columns for **LOC** and **ENG** [VLMEvalKit](https://github.com/timothycdc/VLMEvalKit/blob/main/docs/en/Development.md):
  - index (int, based on dataset order. Does not follow CVQA ids since they are of type str)
  - image (base64)
  - question 
  - A option
  - B option
  - C option
  - D option
  - l2-category (`Subset`)
  - split (always called `test`)
  

## Info
- Proposed method of evaluation:
  - Prompt the model to answer only with the correct option letter (one of `[A,B,C,D]`)
  - Use regex or string search to locate the correct letter
  - Alternatively, use LLM-as-a-judge to identify the correct answer letter. Although, this is a bit of an overkill.
- The original CVQA dataset numbers the options as `[0,1,2,3]`, however this has been changed to`[A,B,C,D]` to follow the VLMEvalKit standard. This shouldn't have much effect on performance.