| | from vlmeval.dataset.utils import build_judge |
| | from vlmeval.smp import * |
| | from .image_base import ImageBaseDataset |
| | from ..utils import track_progress_rich |
| | from ..smp import load, dump, decode_base64_to_image |
| | from .utils import DEBUG_MESSAGE |
| |
|
| | import zipfile |
| | from random import shuffle, seed |
| |
|
| |
|
| | RANDOM_SEED = 0 |
| |
|
| | VQA_SYSTEM_PROMPT = json.dumps({ |
| | 'task': 'Answer the question presented to you truthfully.', |
| | 'requirements': [ |
| | 'Analyze the image(s) first, then answer the question. If you are given a list of possible answers, you must choose from it.', |
| | 'You must answer in the following json format: {"analysis": "(write your analysis here)", "answer": "(your answer)"}' |
| | ] |
| | }) |
| |
|
| | EVAL_SYSTEM_PROMPT = json.dumps({ |
| | 'task': 'Evaluate whether the answer to a question is correct.', |
| | 'requirements': [ |
| | 'Compare an answer to a question with the ground truth answer. Determine whether it is correct.', |
| | 'You must ignore any analysis of the problem if present. You must focus only on the final answer.', |
| | 'You must answer in the following json format: {"verdict": "(1 for correct, 0 for incorrect)"}' |
| | ] |
| | }) |
| |
|
| | def str2json(s: str): |
| | try: |
| | return json.loads(s) |
| | except: |
| | if s.startswith('```json') and s.endswith('```'): |
| | return json.loads(s[6:-3]) |
| | return s |
| |
|
| | class MOAT(ImageBaseDataset): |
| | TYPE = 'VQA' |
| | DATASET_URL = { |
| | 'MOAT': "https://huggingface.co/datasets/waltsun/MOAT/resolve/main/MOAT.tsv", |
| | } |
| | DATASET_MD5 = { |
| | 'MOAT': '803b5a176a5b01aa1b8094fae73532a2', |
| | } |
| |
|
| | def __init__(self, dataset, **kwargs): |
| | super().__init__(dataset, **kwargs) |
| | seed(RANDOM_SEED) |
| |
|
| | def post_build(self, dataset): |
| | assert dataset == "MOAT", f"Wrong dataset name {dataset}" |
| | ROOT = LMUDataRoot() |
| | os.makedirs(ROOT, exist_ok=True) |
| |
|
| | |
| | zip_url = "https://huggingface.co/datasets/waltsun/MOAT/resolve/main/MOAT_images.zip" |
| | zip_md5 = 'c0818a3e0a3f0bc7ee2be89ff04d73a6' |
| | zip_path = osp.join(ROOT, "MOAT_images.zip") |
| | if osp.exists(zip_path) and md5(zip_path) == zip_md5: |
| | pass |
| | else: |
| | warnings.warn('The dataset tsv is not downloaded') |
| | download_file(zip_url, zip_path) |
| | |
| | |
| | self.img_root = osp.join(ROOT, 'MOAT_images') |
| | if not osp.exists(self.img_root): |
| | with zipfile.ZipFile(osp.join(ROOT, 'MOAT_images.zip'), 'r') as zip_ref: |
| | zip_ref.extractall(self.img_root) |
| |
|
| | def build_prompt(self, line): |
| | if isinstance(line, int): |
| | line = self.data.iloc[line] |
| |
|
| | question, choices, images, outside_knowledge_text, outside_knowledge_images = line['question'], line['choices'], line['images'], line['outside_knowledge_text'], line['outside_knowledge_images'] |
| | choices, images, outside_knowledge_images = toliststr(choices), toliststr(images), toliststr(outside_knowledge_images) |
| |
|
| | if len(choices): |
| | shuffle(choices) |
| | question += f'\nThe choices are: {choices}' |
| | msgs = [ |
| | { |
| | 'type': 'text', |
| | 'value': VQA_SYSTEM_PROMPT + '\n' + question, |
| | }, |
| | ] |
| | for img in images: |
| | msgs.append({'type': 'image', 'value': osp.join(self.img_root, img)}) |
| | if not pd.isna(outside_knowledge_text): |
| | msgs.append({'type': 'text', 'value': 'Hint:\n' + outside_knowledge_text}) |
| | for img in outside_knowledge_images: |
| | msgs.append({'type': 'image', 'value': osp.join(self.img_root, img)}) |
| | return msgs |
| | |
| | @classmethod |
| | def evaluate(self, eval_file, **judge_kwargs): |
| | model = judge_kwargs['model'] |
| | suffix = eval_file.split('.')[-1] |
| | result_path = eval_file.replace(f'.{suffix}', f"_{model}.xlsx") |
| | nproc = judge_kwargs.pop('nproc', 4) |
| |
|
| | if not osp.exists(result_path): |
| | data = load(eval_file) |
| | model = build_judge(**judge_kwargs) |
| | assert model.working(), ('MOAT evaluation requires a working OPENAI API\n' + DEBUG_MESSAGE) |
| |
|
| | def extract_prediction(s: str): |
| | try: |
| | return json.loads(s)['answer'] |
| | except: |
| | pass |
| | try: |
| | if s.startswith('```json') and s.endswith('```'): |
| | return json.loads(s[7:-3])['answer'] |
| | except: |
| | print(">>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ", s) |
| | return s |
| |
|
| | def extract_verdict(s: str): |
| | try: |
| | return json.loads(s)['verdict'] |
| | except: |
| | if s.startswith('```json') and s.endswith('```'): |
| | return json.loads(s[7:-3])['verdict'] |
| | raise |
| |
|
| | def verdict_one(model, line): |
| | prediction = extract_prediction(line['prediction']) |
| | answer = line['answer'] |
| | prompt = EVAL_SYSTEM_PROMPT + '\n' + f'The answer to evaluate is {prediction}\nThe ground truth answer is {answer}' |
| | res = model.generate(prompt) |
| | return extract_verdict(res) |
| |
|
| | verdict_list = track_progress_rich( |
| | lambda line: verdict_one(model, line), |
| | [data.iloc[i] for i in range(len(data))], |
| | nproc=nproc, |
| | chunksize=nproc, |
| | ) |
| | data['verdict'] = verdict_list |
| | dump(data, result_path) |
| | |
| | data = load(result_path) |
| | overall_acc = data['verdict'].mean() |
| | capability_set = set() |
| | for i in range(len(data)): |
| | capability_set.update(toliststr(data.iloc[i]['capability'])) |
| | capability_score_map = {capability: (0, 0) for capability in capability_set} |
| | for i in range(len(data)): |
| | line = data.iloc[i] |
| | capabilities = toliststr(line['capability']) |
| | verdict = line['verdict'] |
| | for capability in capabilities: |
| | capability_score_map[capability] = (capability_score_map[capability][0] + verdict, capability_score_map[capability][1] + 1) |
| | capability_score_map = {capability: score[0] / score[1] for capability, score in capability_score_map.items()} |
| | metrics = { |
| | 'overall_acc': overall_acc, |
| | 'result_path': result_path, |
| | 'capability_acc': capability_score_map, |
| | } |
| | score_pth = eval_file.replace(f'.{suffix}', f"_score.json") |
| | dump(metrics, score_pth) |
| | |
| | return metrics |
| |
|