| | import os |
| | import re |
| | import tempfile |
| | from functools import partial |
| |
|
| | import pandas as pd |
| |
|
| | from .image_base import ImageBaseDataset |
| | from .utils import build_judge, DEBUG_MESSAGE |
| | from ..smp import * |
| | from ..utils import track_progress_rich |
| |
|
| |
|
| | class ImageVQADataset(ImageBaseDataset): |
| | TYPE = 'VQA' |
| |
|
| | DATASET_URL = { |
| | 'OCRVQA_TEST': 'https://opencompass.openxlab.space/utils/VLMEval/OCRVQA_TEST.tsv', |
| | 'OCRVQA_TESTCORE': 'https://opencompass.openxlab.space/utils/VLMEval/OCRVQA_TESTCORE.tsv', |
| | 'TextVQA_VAL': 'https://opencompass.openxlab.space/utils/VLMEval/TextVQA_VAL.tsv', |
| | 'DocVQA_VAL': 'https://opencompass.openxlab.space/utils/VLMEval/DocVQA_VAL.tsv', |
| | 'DocVQA_TEST': 'https://opencompass.openxlab.space/utils/VLMEval/DocVQA_TEST.tsv', |
| | 'InfoVQA_VAL': 'https://opencompass.openxlab.space/utils/VLMEval/InfoVQA_VAL.tsv', |
| | 'InfoVQA_TEST': 'https://opencompass.openxlab.space/utils/VLMEval/InfoVQA_TEST.tsv', |
| | 'ChartQA_TEST': 'https://opencompass.openxlab.space/utils/VLMEval/ChartQA_TEST.tsv', |
| | 'GQA_TestDev_Balanced': 'https://opencompass.openxlab.space/utils/VLMEval/GQA_TestDev_Balanced.tsv', |
| | } |
| |
|
| | DATASET_MD5 = { |
| | 'OCRVQA_TEST': 'ca46a6d74b403e9d6c0b670f6fc00db9', |
| | 'OCRVQA_TESTCORE': 'c5239fe77db8bdc1f2ad8e55e0d1fe97', |
| | 'TextVQA_VAL': 'b233b31f551bbf4056f2f955da3a92cd', |
| | 'DocVQA_VAL': 'd5ee77e1926ff10690d469c56b73eabf', |
| | 'DocVQA_TEST': '6a2f28cac26ef2d3447374e8c6f6c8e9', |
| | 'InfoVQA_VAL': '2342e9c225222f0ef4dec545ebb126fe', |
| | 'InfoVQA_TEST': 'df535bf51b88dc9718252c34131a6227', |
| | 'ChartQA_TEST': 'c902e0aa9be5582a7aad6dcf52734b42', |
| | 'GQA_TestDev_Balanced': '99b62f22e224d9b2f32dcbe41359d1c9', |
| | } |
| |
|
| | def build_prompt(self, line): |
| | msgs = super().build_prompt(line) |
| | assert msgs[-1]['type'] == 'text' |
| | msgs[-1]['value'] += '\nAnswer the question using a single word or phrase.' |
| | return msgs |
| |
|
| | |
| | def evaluate(self, eval_file, **judge_kwargs): |
| | from .utils.vqa_eval import hit_calculate, process_line |
| |
|
| | data = load(eval_file) |
| | dataset = self.dataset_name |
| | assert 'answer' in data and 'prediction' in data |
| | data['prediction'] = [str(x) for x in data['prediction']] |
| | data['answer'] = [str(x) for x in data['answer']] |
| | lt = len(data) |
| | pool = mp.Pool(16) |
| | lines = [data.iloc[i] for i in range(lt)] |
| | if listinstr(['TextVQA'], dataset): |
| | res = pool.map(partial(process_line, method='vqa_score'), lines) |
| | elif listinstr(['ChartQA'], dataset): |
| | res = pool.map(partial(process_line, method='relaxed_accuracy'), lines) |
| | elif listinstr(['OCRVQA', 'GQA'], dataset): |
| | res = pool.map(partial(process_line, method='accuracy'), lines) |
| | elif listinstr(['DocVQA', 'InfoVQA'], dataset): |
| | res = pool.map(partial(process_line, method='anls'), lines) |
| | else: |
| | res = pool.map(process_line, lines) |
| | hit = hit_calculate(res, dataset) |
| | ret = dict() |
| | if 'split' in data: |
| | splits = set(data['split']) |
| | for sp in splits: |
| | sub = [r for l, r in zip(lines, res) if l['split'] == sp] |
| | |
| | hit = hit_calculate(sub, dataset) |
| | ret[sp] = np.mean(hit) * 100 |
| | sub = [r for l, r in zip(lines, res)] |
| | hit = hit_calculate(sub, dataset) |
| | ret['Overall'] = np.mean(hit) * 100 |
| | else: |
| | ret['Overall'] = np.mean(hit) * 100 |
| | if 'category' in data: |
| | cates = list(set(data['category'])) |
| | cates.sort() |
| | for c in cates: |
| | sub = [r for l, r in zip(lines, res) if l['category'] == c] |
| | |
| | hit = hit_calculate(sub, dataset) |
| | ret[c] = np.mean(hit) * 100 |
| | ret = d2df(ret) |
| | ret.round(2) |
| |
|
| | suffix = eval_file.split('.')[-1] |
| | result_file = eval_file.replace(f'.{suffix}', '_acc.csv') |
| | dump(ret, result_file) |
| | return ret |
| |
|
| |
|
| | class VizWiz(ImageBaseDataset): |
| | TYPE = 'VQA' |
| | DATASET_URL = { |
| | 'VizWiz': 'https://opencompass.openxlab.space/utils/VLMEval/VizWiz.tsv' |
| | } |
| | DATASET_MD5 = { |
| | 'VizWiz': 'fa4ac4164467563ed2fac6eac6631bd0' |
| | } |
| |
|
| | @classmethod |
| | def evaluate(self, eval_file, **judge_kwargs): |
| | from .utils.vqa_eval import hit_calculate, process_line |
| |
|
| | suffix = eval_file.split('.')[-1] |
| | result_file = eval_file.replace(f'.{suffix}', '_acc.csv') |
| |
|
| | if not osp.exists(result_file): |
| | data = load(eval_file) |
| | assert 'answers' in data and 'prediction' in data |
| | data['prediction'] = [str(x) for x in data['prediction']] |
| | data['answer'] = [str(x) for x in data['answers']] |
| |
|
| | lt = len(data) |
| | pool = mp.Pool(16) |
| | lines = [data.iloc[i] for i in range(lt)] |
| | res = pool.map(process_line, lines) |
| |
|
| | hit = hit_calculate(res, 'VizWiz') |
| | ret = dict() |
| |
|
| | ret['Overall'] = np.mean(hit) * 100 |
| | ret = d2df(ret) |
| | ret.round(2) |
| |
|
| | dump(ret, result_file) |
| |
|
| | retz = pd.read_csv(result_file) |
| | return retz |
| |
|
| |
|
| | class OCRBench(ImageBaseDataset): |
| | TYPE = 'VQA' |
| | DATASET_URL = { |
| | 'OCRBench': 'https://opencompass.openxlab.space/utils/VLMEval/OCRBench.tsv', |
| | |
| | 'OCRBench_MINI': 'https://opencompass.openxlab.space/utils/TEST/OCRBench_MINI.tsv' |
| | } |
| | DATASET_MD5 = {'OCRBench': 'e953d98a987cc6e26ef717b61260b778'} |
| |
|
| | |
| | @classmethod |
| | def evaluate(self, eval_file, **judge_kwargs): |
| | OCRBench_score = { |
| | 'Regular Text Recognition': 0, |
| | 'Irregular Text Recognition': 0, |
| | 'Artistic Text Recognition': 0, |
| | 'Handwriting Recognition': 0, |
| | 'Digit String Recognition': 0, |
| | 'Non-Semantic Text Recognition': 0, |
| | 'Scene Text-centric VQA': 0, |
| | 'Doc-oriented VQA': 0, |
| | 'Key Information Extraction': 0, |
| | 'Handwritten Mathematical Expression Recognition': 0, |
| | } |
| |
|
| | data = load(eval_file) |
| | lt = len(data) |
| | lines = [data.iloc[i] for i in range(lt)] |
| | for i in tqdm(range(len(lines))): |
| | line = lines[i] |
| | predict = str(line['prediction']) |
| | answers = eval(line['answer']) |
| | category = line['category'] |
| | if category == 'Handwritten Mathematical Expression Recognition': |
| | for j in range(len(answers)): |
| | answer = answers[j].strip().replace('\n', ' ').replace(' ', '') |
| | predict = predict.strip().replace('\n', ' ').replace(' ', '') |
| | if answer in predict: |
| | OCRBench_score[category] += 1 |
| | break |
| | else: |
| | for j in range(len(answers)): |
| | answer = answers[j].lower().strip().replace('\n', ' ') |
| | predict = predict.lower().strip().replace('\n', ' ') |
| | if answer in predict: |
| | OCRBench_score[category] += 1 |
| | break |
| |
|
| | final_score_dict = {} |
| | final_score_dict['Text Recognition'] = \ |
| | (OCRBench_score['Regular Text Recognition'] + OCRBench_score['Irregular Text Recognition'] |
| | + OCRBench_score['Artistic Text Recognition'] + OCRBench_score['Handwriting Recognition'] |
| | + OCRBench_score['Digit String Recognition'] + OCRBench_score['Non-Semantic Text Recognition']) |
| | final_score_dict['Scene Text-centric VQA'] = OCRBench_score['Scene Text-centric VQA'] |
| | final_score_dict['Doc-oriented VQA'] = OCRBench_score['Doc-oriented VQA'] |
| | final_score_dict['Key Information Extraction'] = OCRBench_score['Key Information Extraction'] |
| | final_score_dict['Handwritten Mathematical Expression Recognition'] = \ |
| | (OCRBench_score['Handwritten Mathematical Expression Recognition']) |
| | final_score_dict['Final Score'] = \ |
| | (final_score_dict['Text Recognition'] + final_score_dict['Scene Text-centric VQA'] |
| | + final_score_dict['Doc-oriented VQA'] + final_score_dict['Key Information Extraction'] |
| | + final_score_dict['Handwritten Mathematical Expression Recognition']) |
| | final_score_dict['Final Score Norm'] = (float(final_score_dict['Final Score']) / 10) |
| | score_pth = eval_file.replace('.xlsx', '_score.json') |
| | dump(final_score_dict, score_pth) |
| | return final_score_dict |
| |
|
| |
|
| | class MathVista(ImageBaseDataset): |
| | TYPE = 'VQA' |
| | DATASET_URL = { |
| | 'MathVista_MINI': 'https://opencompass.openxlab.space/utils/VLMEval/MathVista_MINI.tsv' |
| | } |
| | DATASET_MD5 = {'MathVista_MINI': 'f199b98e178e5a2a20e7048f5dcb0464'} |
| |
|
| | |
| | @classmethod |
| | def evaluate(self, eval_file, **judge_kwargs): |
| | from .utils.mathvista import MathVista_auxeval, MathVista_acc |
| |
|
| | model = judge_kwargs['model'] |
| | suffix = eval_file.split('.')[-1] |
| | storage = eval_file.replace(f'.{suffix}', f'_{model}.xlsx') |
| | tmp_file = eval_file.replace(f'.{suffix}', f'_{model}.pkl') |
| | nproc = judge_kwargs.pop('nproc', 4) |
| |
|
| | if not osp.exists(storage): |
| | data = load(eval_file) |
| | model = build_judge(max_tokens=128, **judge_kwargs) |
| | assert model.working(), ('MathVista evaluation requires a working OPENAI API\n' + DEBUG_MESSAGE) |
| | lt = len(data) |
| | lines = [data.iloc[i] for i in range(lt)] |
| | tups = [(model, line) for line in lines] |
| | indices = [line['index'] for line in lines] |
| |
|
| | ans = {} |
| | if osp.exists(tmp_file): |
| | ans = load(tmp_file) |
| | tups = [x for x, i in zip(tups, indices) if i not in ans] |
| | indices = [i for i in indices if i not in ans] |
| |
|
| | if len(indices): |
| | new_results = track_progress_rich( |
| | MathVista_auxeval, |
| | tups, |
| | nproc=nproc, |
| | chunksize=nproc, |
| | keys=indices, |
| | save=tmp_file, |
| | ) |
| | ans = load(tmp_file) |
| | for k, v in zip(indices, new_results): |
| | assert k in ans |
| | assert ans[k]['log'] == v['log'] and ans[k]['res'] == v['res'] |
| |
|
| | data['res'] = [ans[idx]['res'] for idx in data['index']] |
| | data['log'] = [ans[idx]['log'] for idx in data['index']] |
| | dump(data, storage) |
| |
|
| | score = MathVista_acc(storage) |
| | score_pth = storage.replace('.xlsx', '_score.csv') |
| | dump(score, score_pth) |
| | return score |
| |
|
| |
|
| | class MathVerse(ImageBaseDataset): |
| | TYPE = 'VQA' |
| | DATASET_URL = { |
| | 'MathVerse_MINI': 'http://opencompass.openxlab.space/utils/benchmarks/MathVerse/MathVerse_MINIV.tsv', |
| | 'MathVerse_MINI_Vision_Only': 'http://opencompass.openxlab.space/utils/benchmarks/MathVerse/MathVerse_MINIVOnly.tsv', |
| | 'MathVerse_MINI_Vision_Only_cot': 'http://opencompass.openxlab.space/utils/benchmarks/MathVerse/MathVerse_MINIVOnly.tsv', |
| | 'MathVerse_MINI_Vision_Dominant': 'http://opencompass.openxlab.space/utils/benchmarks/MathVerse/MathVerse_MINIVDom.tsv', |
| | 'MathVerse_MINI_Vision_Intensive': 'http://opencompass.openxlab.space/utils/benchmarks/MathVerse/MathVerse_MINIVInt.tsv', |
| | 'MathVerse_MINI_Text_Lite': 'http://opencompass.openxlab.space/utils/benchmarks/MathVerse/MathVerse_MINITLite.tsv', |
| | 'MathVerse_MINI_Text_Dominant': 'http://opencompass.openxlab.space/utils/benchmarks/MathVerse/MathVerse_MINITDom.tsv', |
| | } |
| | DATASET_MD5 = { |
| | 'MathVerse_MINI': '5017caca32b7fa110c350a1bea861b65', |
| | 'MathVerse_MINI_Vision_Only': '68a11d4680014ac881fa37adeadea3a4', |
| | 'MathVerse_MINI_Vision_Only_cot': '68a11d4680014ac881fa37adeadea3a4', |
| | 'MathVerse_MINI_Vision_Dominant': 'b8fb63852d261ab2aaefba29cc2414d3', |
| | 'MathVerse_MINI_Vision_Intensive': '01cbd35be202bb0c4873a4186a63bc19', |
| | 'MathVerse_MINI_Text_Lite': '19e4b13bdd30b89a03b2e358bcfefa04', |
| | 'MathVerse_MINI_Text_Dominant': '4f5cd2fa6630ea00bb11d6fde1f6fe6a', |
| | } |
| |
|
| | |
| | def build_prompt(self, line): |
| | if isinstance(line, int): |
| | line = self.data.iloc[line] |
| |
|
| | if self.meta_only: |
| | tgt_path = toliststr(line['image_path']) |
| | else: |
| | tgt_path = self.dump_image(line) |
| | if 'cot' in self.dataset_name: |
| | question = line['query_cot'] |
| | else: |
| | question = line['question'] |
| |
|
| | msgs = [] |
| | if isinstance(tgt_path, list): |
| | msgs.extend([dict(type='image', value=p) for p in tgt_path]) |
| | else: |
| | msgs = [dict(type='image', value=tgt_path)] |
| | msgs.append(dict(type='text', value=question)) |
| | return msgs |
| |
|
| | |
| | @classmethod |
| | def evaluate(self, eval_file, **judge_kwargs): |
| | from .utils.mathverse import MathVerse_auxeval_extract, MathVerse_auxeval_score, MathVerse_acc |
| |
|
| | model = judge_kwargs['model'] |
| | suffix = eval_file.split('.')[-1] |
| | storage_extract = eval_file.replace(f'.{suffix}', f'_{model}_extract.xlsx') |
| | tmp_file_extract = eval_file.replace(f'.{suffix}', f'_{model}_extract.pkl') |
| | storage_score = eval_file.replace(f'.{suffix}', f'_{model}_score.xlsx') |
| | tmp_file_score = eval_file.replace(f'.{suffix}', f'_{model}_score.pkl') |
| | nproc = judge_kwargs.pop('nproc', 4) |
| | |
| | if not osp.exists(storage_extract): |
| | data = load(eval_file) |
| | model = build_judge(max_tokens=128, **judge_kwargs) |
| | assert model.working(), ('MathVerse evaluation requires a working OPENAI API\n' + DEBUG_MESSAGE) |
| | lt = len(data) |
| | lines = [data.iloc[i] for i in range(lt)] |
| | tups = [(model, line) for line in lines] |
| | indices = [line['index'] for line in lines] |
| |
|
| | ans = {} |
| | if osp.exists(tmp_file_extract): |
| | ans = load(tmp_file_extract) |
| | tups = [x for x, i in zip(tups, indices) if i not in ans] |
| | indices = [i for i in indices if i not in ans] |
| |
|
| | if len(indices): |
| | new_results = track_progress_rich( |
| | MathVerse_auxeval_extract, |
| | tups, |
| | nproc=nproc, |
| | chunksize=nproc, |
| | keys=indices, |
| | save=tmp_file_extract, |
| | ) |
| | ans = load(tmp_file_extract) |
| | for k, v in zip(indices, new_results): |
| | assert k in ans |
| | assert ans[k]['log_extract'] == v['log_extract'] and ans[k]['extract'] == v['extract'] |
| |
|
| | data['extract'] = [ans[idx]['extract'] for idx in data['index']] |
| | data['log_extract'] = [ans[idx]['log_extract'] for idx in data['index']] |
| | dump(data, storage_extract) |
| |
|
| | |
| | if not osp.exists(storage_score): |
| | data = load(storage_extract) |
| | model = build_judge(max_tokens=128, **judge_kwargs) |
| | assert model.working(), ('MathVerse evaluation requires a working OPENAI API\n' + DEBUG_MESSAGE) |
| | lt = len(data) |
| | lines = [data.iloc[i] for i in range(lt)] |
| | tups = [(model, line) for line in lines] |
| | indices = [line['index'] for line in lines] |
| |
|
| | ans = {} |
| | if osp.exists(tmp_file_score): |
| | ans = load(tmp_file_score) |
| | tups = [x for x, i in zip(tups, indices) if i not in ans] |
| | indices = [i for i in indices if i not in ans] |
| |
|
| | if len(indices): |
| | new_results = track_progress_rich( |
| | MathVerse_auxeval_score, |
| | tups, |
| | nproc=nproc, |
| | chunksize=nproc, |
| | keys=indices, |
| | save=tmp_file_score, |
| | ) |
| | ans = load(tmp_file_score) |
| | for k, v in zip(indices, new_results): |
| | assert k in ans |
| | assert ans[k]['log_score'] == v['log_score'] and ans[k]['score'] == v['score'] |
| |
|
| | data['score'] = [ans[idx]['score'] for idx in data['index']] |
| | data['log_score'] = [ans[idx]['log_score'] for idx in data['index']] |
| | dump(data, storage_score) |
| |
|
| | score = MathVerse_acc(storage_score) |
| | score_pth = storage_score.replace('.xlsx', '.csv') |
| | dump(score, score_pth) |
| | return score |
| |
|
| |
|
| | class MathVision(ImageBaseDataset): |
| | TYPE = 'VQA' |
| | DATASET_URL = { |
| | 'MathVision': 'https://opencompass.openxlab.space/utils/VLMEval/MathVision.tsv', |
| | 'MathVision_MINI': 'https://opencompass.openxlab.space/utils/VLMEval/MathVision_MINI.tsv' |
| | } |
| | DATASET_MD5 = { |
| | 'MathVision': '93f6de14f7916e598aa1b7165589831e', |
| | 'MathVision_MINI': '060fe4fa5d868987ce179307bd5f8a33' |
| | } |
| |
|
| | |
| | @classmethod |
| | def evaluate(self, eval_file, **judge_kwargs): |
| | from .utils.mathv import MATH_V_auxeval, MATH_V_acc |
| |
|
| | if 'model' in judge_kwargs: |
| | model = judge_kwargs['model'] |
| | else: |
| | model = os.path.basename(os.environ.get('LOCAL_LLM')) |
| | suffix = eval_file.split('.')[-1] |
| | storage = eval_file.replace(f'.{suffix}', f'_{model}.xlsx') |
| | tmp_file = eval_file.replace(f'.{suffix}', f'_{model}.pkl') |
| | nproc = judge_kwargs.pop('nproc', 4) |
| |
|
| | if not osp.exists(storage): |
| | data = load(eval_file) |
| | model = build_judge(max_tokens=128, **judge_kwargs) |
| | assert model.working(), ('MATH-Vision evaluation requires a working OPENAI API\n' + DEBUG_MESSAGE) |
| | lt = len(data) |
| | lines = [data.iloc[i] for i in range(lt)] |
| | tups = [(model, line) for line in lines] |
| | indices = [line['index'] for line in lines] |
| |
|
| | ans = {} |
| | if osp.exists(tmp_file): |
| | ans = load(tmp_file) |
| | tups = [x for x, i in zip(tups, indices) if i not in ans] |
| | indices = [i for i in indices if i not in ans] |
| |
|
| | if len(indices): |
| | new_results = track_progress_rich( |
| | MATH_V_auxeval, |
| | tups, |
| | nproc=nproc, |
| | chunksize=nproc, |
| | keys=indices, |
| | save=tmp_file, |
| | ) |
| | ans = load(tmp_file) |
| | for k, v in zip(indices, new_results): |
| | assert k in ans |
| | assert ans[k]['log'] == v['log'] and ans[k]['res'] == v['res'] |
| |
|
| | data['res'] = [ans[idx]['res'] for idx in data['index']] |
| | data['log'] = [ans[idx]['log'] for idx in data['index']] |
| | dump(data, storage) |
| |
|
| | score = MATH_V_acc(storage) |
| | score_pth = storage.replace('.xlsx', '_score.csv') |
| | dump(score, score_pth) |
| | return score |
| |
|
| |
|
| | class OlympiadBench(ImageBaseDataset): |
| | TYPE = 'VQA_ex_prompt' |
| | DATASET_URL = { |
| | 'OlympiadBench': 'https://opencompass.openxlab.space/utils/VLMEval/OlympiadBench.tsv', |
| | 'OlympiadBench_EN': 'https://opencompass.openxlab.space/utils/VLMEval/OlympiadBench_EN.tsv', |
| | 'OlympiadBench_CN': 'https://opencompass.openxlab.space/utils/VLMEval/OlympiadBench_CN.tsv' |
| | } |
| | DATASET_MD5 = { |
| | 'OlympiadBench': '9735ae0f0299eae1e7d07f5a7feab914', |
| | 'OlympiadBench_EN': '5c68e100d394351fc7049f29d4d4efed', |
| | 'OlympiadBench_CN': 'ea01b16788955702c79650c701e5b623' |
| | } |
| |
|
| | def dump_image(self, line): |
| | os.makedirs(self.img_root, exist_ok=True) |
| |
|
| | tgt_path_z = [] |
| | if isinstance(line['image'], list): |
| | for i in range(len(line['image'])): |
| | tgt_path = osp.join(self.img_root, f"{line['index']}--{i + 1}.jpg") |
| | if not read_ok(tgt_path): |
| | decode_base64_to_image_file(line['image'][i], tgt_path) |
| | tgt_path_z.append(tgt_path) |
| | else: |
| | tgt_path = osp.join(self.img_root, f"{line['index']}.jpg") |
| | if not read_ok(tgt_path): |
| | decode_base64_to_image_file(line['image'], tgt_path) |
| | tgt_path_z.append(tgt_path) |
| | return tgt_path_z |
| |
|
| | def build_prompt(self, line): |
| |
|
| | from .utils.olympiadbench import get_answer_type_text, make_input |
| |
|
| | self.is_chinese = 'zh' in line['source'] |
| | self.is_math = 'maths' in line['source'] |
| | self.is_theorem_proving = 'TP' in line['source'] |
| |
|
| | if self.is_chinese: |
| | subject_content = '数学' if self.is_math else '物理' |
| | if self.is_theorem_proving: |
| | prompt = ( |
| | f"以下是中国{subject_content}竞赛中的证明题。请根据题目的要求,运用逻辑推理及常用定理证明题目中的命题。" |
| | "证明过程中使用的变量和公式请使用LaTeX格式表示。" |
| | ) |
| | else: |
| | answer_type_text = get_answer_type_text(line['answer_type'], is_chinese=True, |
| | multiple_answer=line['is_multiple_answer']) |
| | if line['is_multiple_answer']: |
| | multiple_answer_text = '\\boxed{用英文逗号连接的多个答案}' |
| | else: |
| | multiple_answer_text = '\\boxed{答案}' |
| | unit_text = '' |
| | if line['unit']: |
| | multiple_answer_text += '(单位)' |
| | unit_text = ',注意答案的单位不要放在\\boxed{}中' |
| | prompt = ( |
| | f'以下是中国{subject_content}竞赛中的解答题{answer_type_text}。请根据题目的要求和所提供的信息计算得出答案。' |
| | f'解答过程和结果中使用的变量和公式请使用LaTeX格式表示。请在最后以“所以最终答案是{multiple_answer_text}。”' |
| | f'显式给出结果{unit_text}。' |
| | ) |
| | else: |
| | subject_content = 'Math' if self.is_math else 'Physics' |
| | if self.is_theorem_proving: |
| | prompt = ( |
| | f'The following is a theorem proving problem from an International {subject_content} competition. ' |
| | 'Please use logical reasoning and common theorems to prove the proposition in the problem ' |
| | 'according to the given requirements. ' |
| | 'Please use LaTeX format to represent the variables and formulas used in the proof.' |
| | ) |
| | else: |
| | if line['is_multiple_answer']: |
| | multiple_answer_text = '\\boxed{multiple answers connected with commas}' |
| | else: |
| | multiple_answer_text = '\\boxed{answer}' |
| | unit_text = '' |
| | if line['unit']: |
| | multiple_answer_text += '(unit)' |
| | unit_text = ', note that the unit of the answer should not be included in \\boxed{}' |
| | answer_type_text = get_answer_type_text(line['answer_type'], is_chinese=False, |
| | multiple_answer=line['is_multiple_answer']) |
| | prompt = ( |
| | f'The following is an open-ended problem from an International {subject_content} competition. ' |
| | f'{answer_type_text}Please calculate the answer according to the given requirements and ' |
| | 'the information provided. Please use LaTeX format to represent the variables and formulas ' |
| | 'used in the solution process and results. Please end your solution with "So the final answer ' |
| | f'is {multiple_answer_text}." and give the result explicitly{unit_text}.' |
| | ) |
| |
|
| | if self.is_math: |
| | input = make_input(prompt, line['question']) |
| | else: |
| | if 'context' in line.keys() and str(line['context']) != 'nan': |
| | input = make_input(prompt, line['context'] + '\n' + line['question']) |
| | else: |
| | input = make_input(prompt, line['question']) |
| |
|
| | ret = [dict(type='text', value=input)] |
| | tgt_path = self.dump_image(line) |
| |
|
| | ret.extend([dict(type='image', value=s) for s in tgt_path]) |
| |
|
| | return ret |
| |
|
| | @classmethod |
| | def evaluate(self, eval_file, **judge_kwargs): |
| | from .utils.olympiadbench import MathJudger, extract_answer |
| | judger = MathJudger() |
| |
|
| | suffix = eval_file.split('.')[-1] |
| | name_str1 = 'judge' |
| | name_str2 = 'score' |
| | result_file = eval_file.replace(f'.{suffix}', f'_{name_str1}_result.xlsx') |
| | score_file = eval_file.replace(f'.{suffix}', f'_{name_str2}_result.csv') |
| |
|
| | if not osp.exists(result_file): |
| | data = load(eval_file) |
| | scorez = [] |
| |
|
| | for i in tqdm(data.iterrows()): |
| | line = i[1] |
| | model_answer = line['prediction'] |
| | is_chinese = 'zh' in line['source'] |
| | model_answer = extract_answer(is_chinese, model_answer, is_deepseek=False) |
| | answer_type = line['answer_type'] |
| |
|
| | final_answer = line['final_answer'][2:-2] |
| |
|
| | if str(answer_type) != 'nan' and 'Tuple' in answer_type: |
| | judge_result = judger.judge(model_answer, final_answer) |
| | else: |
| | if str(line['error']) != 'nan': |
| | if ',' in line['error']: |
| | precisions = line['error'].split(',') |
| | precisions = [float(p) if p else 1e-8 for p in precisions] |
| | judge_result = judger.judge(model_answer, final_answer, precisions) |
| | else: |
| | precision = float(line['error']) |
| | judge_result = judger.judge(model_answer, final_answer, precision) |
| | else: |
| | judge_result = judger.judge(model_answer, final_answer) |
| | scorez.append(judge_result) |
| |
|
| | data['score'] = scorez |
| | dump(data, result_file) |
| |
|
| | judge_file = load(result_file) |
| |
|
| | if not osp.exists(score_file): |
| | name_list = ['OE_MM_maths_en_COMP', 'OE_MM_maths_zh_CEE', 'OE_MM_maths_zh_COMP', 'OE_MM_physics_en_COMP', |
| | 'OE_MM_physics_zh_CEE','OE_TO_maths_en_COMP', 'OE_TO_maths_zh_CEE', 'OE_TO_maths_zh_COMP', |
| | 'OE_TO_physics_en_COMP', 'OE_TO_physics_zh_CEE'] |
| |
|
| | sample_list = [[] for _ in range(len(name_list))] |
| | for i in judge_file.iterrows(): |
| | line = i[1] |
| | for j in range(len(name_list)): |
| | if line['source'] == name_list[j]: |
| | sample_list[j].append(line['score']) |
| |
|
| | acc_dict = {} |
| | correct_list = [] |
| |
|
| | |
| | for i in range(len(name_list)): |
| | correct_num = 0 |
| | for j in sample_list[i]: |
| | if j: |
| | correct_num += 1 |
| | correct_list.append(correct_num) |
| | acc = 100 * correct_num / len(sample_list[i]) |
| | acc_dict[name_list[i]] = [acc] |
| |
|
| | |
| | labela = ['zh', 'en'] |
| | labelb = ['maths', 'physics'] |
| |
|
| | grain_list = [[x,y] for x in labela for y in labelb] |
| | for j in grain_list: |
| | dict_name = j[0] + "_" + j[1] |
| | correct_num = 0 |
| | full_num = 0 |
| | for i in range(len(name_list)): |
| | if all(k in name_list[i] for k in j): |
| | correct_num += correct_list[i] |
| | full_num += len(sample_list[i]) |
| | acc = 100 * correct_num / full_num |
| | acc_dict[dict_name] = [acc] |
| |
|
| | |
| | grain_list = ['maths', 'physics'] |
| | for j in grain_list: |
| | dict_name = j |
| | correct_num = 0 |
| | full_num = 0 |
| | for i in range(len(name_list)): |
| | if j in name_list[i]: |
| | correct_num += correct_list[i] |
| | full_num += len(sample_list[i]) |
| | acc = 100 * correct_num / full_num |
| | acc_dict[dict_name] = [acc] |
| |
|
| | |
| | correct_num = sum(correct_list) |
| | acc = 100 * correct_num / len(judge_file) |
| | acc_dict['AVG'] = [acc] |
| |
|
| | acc_pd = pd.DataFrame(acc_dict) |
| | acc_pd.to_csv(score_file, index=False, encoding='gbk') |
| |
|
| | accdz = pd.read_csv(score_file) |
| | return accdz |
| |
|
| |
|
| | class LogicVista(ImageBaseDataset): |
| | TYPE = 'VQA' |
| | DATASET_URL = { |
| | 'LogicVista': 'https://opencompass.openxlab.space/utils/VLMEval/LogicVista.tsv' |
| | } |
| | DATASET_MD5 = {'LogicVista': '41c5d33adf33765c399e0e6ae588c061'} |
| |
|
| | def evaluate(self, eval_file, **judge_kwargs): |
| | from .utils.logicvista import LogicVista_auxeval, evaluate_logicvista |
| |
|
| | |
| | model = judge_kwargs.get('model', 'exact_matching') |
| | assert model in ['exact_matching', 'gpt-4-0125', 'gpt-4-turbo', 'gpt-4o-mini'], model |
| | name_str_map = {'gpt-4-0125': 'gpt4', 'gpt-4-turbo': 'gpt4-turbo', 'gpt-4o-mini': 'gpt4o-mini'} |
| | name_str = name_str_map[model] if model in name_str_map else model |
| |
|
| | if model == 'exact_matching': |
| | model = None |
| | elif gpt_key_set(): |
| | model = build_judge(**judge_kwargs) |
| | if not model.working(): |
| | warnings.warn('OPENAI API is not working properly, will use exact matching for evaluation') |
| | warnings.warn(DEBUG_MESSAGE) |
| | model = None |
| | else: |
| | warnings.warn('OPENAI_API_KEY is not set properly, will use exact matching for evaluation') |
| | model = None |
| |
|
| | suffix = eval_file.split('.')[-1] |
| | storage = eval_file.replace(f'.{suffix}', f'_{name_str}.xlsx') |
| | tmp_file = eval_file.replace(f'.{suffix}', f'_{name_str}.pkl') |
| | nproc = judge_kwargs.pop('nproc', 4) |
| |
|
| | if not osp.exists(storage) and model is not None: |
| | data = load(eval_file) |
| | model = build_judge(max_tokens=128, **judge_kwargs) |
| | assert model.working(), ('LogicVista evaluation requires a working OPENAI API\n' + DEBUG_MESSAGE) |
| | lt = len(data) |
| | lines = [data.iloc[i] for i in range(lt)] |
| | tups = [(model, line) for line in lines] |
| | indices = [line['index'] for line in lines] |
| |
|
| | ans = {} |
| | if osp.exists(tmp_file): |
| | ans = load(tmp_file) |
| | tups = [x for x, i in zip(tups, indices) if i not in ans] |
| | indices = [i for i in indices if i not in ans] |
| |
|
| | if len(indices): |
| | new_results = track_progress_rich( |
| | LogicVista_auxeval, |
| | tups, |
| | nproc=nproc, |
| | chunksize=nproc, |
| | keys=indices, |
| | save=tmp_file, |
| | ) |
| | ans = load(tmp_file) |
| | for k, v in zip(indices, new_results): |
| | assert k in ans |
| | assert ans[k]['log'] == v['log'] and ans[k]['res'] == v['res'] and ans[k]['hit'] == v['hit'] |
| |
|
| | data['res'] = [ans[idx]['res'] for idx in data['index']] |
| | data['log'] = [ans[idx]['log'] for idx in data['index']] |
| | data['hit'] = [ans[idx]['hit'] for idx in data['index']] |
| |
|
| | dump(data, storage) |
| | if osp.exists(storage): |
| | accuracy_scores = evaluate_logicvista(storage) |
| | score_pth = storage.replace('.xlsx', '_score.csv') |
| | dump(accuracy_scores, score_pth) |
| |
|
| | return accuracy_scores |
| |
|
| | class MME_CoT(ImageBaseDataset): |
| | TYPE = 'VQA' |
| | DATASET_URL = { |
| | 'MME_CoT_TEST': 'https://huggingface.co/datasets/CaraJ/MME-CoT_VLMEvalKit/resolve/main/MME-CoT.tsv' |
| | } |
| | DATASET_MD5 = { |
| | 'MME_CoT_TEST': 'a612dee0f2d702e01fe50267201302e0', |
| | } |
| |
|
| | def split_MME_CoT(self, msgs): |
| | text, images = None, [] |
| |
|
| | |
| | for s in msgs: |
| | if s['type'] == 'image': |
| | images.append(s['value']) |
| | elif s['type'] == 'text': |
| | assert text is None |
| | text = s['value'] |
| |
|
| | |
| | text_segs = text.split('<image>') |
| |
|
| | |
| | segs = [] |
| | |
| | for i, seg in enumerate(text_segs): |
| | |
| | if i > 0 and i - 1 < len(images): |
| | segs.append(dict(type='image', value=images[i - 1])) |
| | |
| | if len(seg.strip()) > 0: |
| | segs.append(dict(type='text', value=seg)) |
| |
|
| | return segs |
| | |
| | def dump_image(self, line): |
| | os.makedirs(self.img_root, exist_ok=True) |
| |
|
| | if 'image' in line: |
| | if isinstance(line['image'], list): |
| | tgt_path = [] |
| | if 'image_path' in line: |
| | image_path_list = line['image_path'] |
| | else: |
| | image_path_list = [f"{line['index']}--{i + 1}.jpg" for i in range(len(line['image']))] |
| | for img, im_name in zip(line['image'], image_path_list): |
| | path = osp.join(self.img_root, im_name) |
| | if not read_ok(path): |
| | decode_base64_to_image_file(img, path) |
| | tgt_path.append(path) |
| | else: |
| | tgt_path = osp.join(self.img_root, f"{line['index']}.jpg") |
| | if not read_ok(tgt_path): |
| | decode_base64_to_image_file(line['image'], tgt_path) |
| | tgt_path = [tgt_path] |
| | else: |
| | assert 'image_path' in line |
| | tgt_path = toliststr(line['image_path']) |
| |
|
| | return tgt_path |
| |
|
| | def build_prompt(self, line): |
| |
|
| | if isinstance(line, int): |
| | line = self.data.iloc[line] |
| |
|
| | tgt_path = self.dump_image(line) |
| |
|
| | prompt = line['question'] |
| |
|
| | options = { |
| | cand: line[cand] |
| | for cand in string.ascii_uppercase |
| | if cand in line and not pd.isna(line[cand]) |
| | } |
| | prompt = prompt + '\n' + '\n'.join([f'{key}. {item}' for key, item in options.items()]) |
| | |
| | |
| | if os.environ.get('USE_COT_PROMPT', '1') == '1': |
| | prompt += "\nPlease generate a step by step answer, include all your intermediate reasoning process, and provide the final answer at the end." |
| | else: |
| | prompt += "\nPlease directly provide the final answer without any other output." |
| |
|
| | msgs = [] |
| | if isinstance(tgt_path, list): |
| | msgs.extend([dict(type='image', value=p) for p in tgt_path]) |
| | else: |
| | msgs = [dict(type='image', value=tgt_path)] |
| | msgs.append(dict(type='text', value=prompt)) |
| |
|
| | msgs = self.split_MME_CoT(msgs) |
| | return msgs |
| |
|
| | |
| | @classmethod |
| | def evaluate(self, eval_file, **judge_kwargs): |
| | print("\033[1;31;40m" + "[MME-CoT Evaluation]: Please refer to the official repository for evaluation: https://github.com/CaraJ7/MME-CoT/tree/main" + "\033[0m") |
| | dummy_result = dict( |
| | dummy_result=0 |
| | ) |
| | return pd.DataFrame(dummy_result, index=[0]) |
| |
|
| |
|
| |
|
| | class LLaVABench(ImageBaseDataset): |
| | TYPE = 'VQA' |
| | DATASET_URL = {'LLaVABench': 'https://opencompass.openxlab.space/utils/VLMEval/LLaVABench.tsv'} |
| | DATASET_MD5 = {'LLaVABench': 'd382a093f749a697820d3dadd61c8428'} |
| |
|
| | |
| | @classmethod |
| | def evaluate(self, eval_file, **judge_kwargs): |
| | from .utils.llavabench import ( |
| | build_prompt, |
| | LLaVABench_atomeval, |
| | LLaVABench_score, |
| | ) |
| |
|
| | suffix = '.' + eval_file.split('.')[-1] |
| | record_file = eval_file.replace(suffix, '_openai_result' + suffix) |
| | score_file = eval_file.replace(suffix, '_score.csv') |
| | nproc = judge_kwargs.pop('nproc', 4) |
| | system_prompt = 'You are a helpful and precise assistant for checking the quality of the answer.' |
| |
|
| | if not osp.exists(record_file): |
| | data = load(eval_file) |
| | lines = [data.iloc[i] for i in range(len(data))] |
| | model = build_judge(temperature=0.2, system_prompt=system_prompt, **judge_kwargs) |
| | assert model.working(), ('LLaVABench evaluation requires a working OPENAI API\n' + DEBUG_MESSAGE) |
| |
|
| | prompts = [build_prompt(line) for line in lines] |
| | tups = [(model, prompt) for prompt in prompts] |
| | scores = track_progress_rich(LLaVABench_atomeval, tups, nproc=nproc, chunksize=nproc) |
| | data['gpt4_score'] = [x[0] for x in scores] |
| | data['score'] = [x[1] for x in scores] |
| | dump(data, record_file) |
| |
|
| | data = load(record_file) |
| | ret = LLaVABench_score(data).round(1) |
| | dump(ret, score_file) |
| | return ret |
| |
|
| |
|
| | class MMVet(ImageBaseDataset): |
| | TYPE = 'VQA' |
| | DATASET_URL = { |
| | 'MMVet': 'https://opencompass.openxlab.space/utils/VLMEval/MMVet.tsv', |
| | 'MMVet_Hard': 'http://opencompass.openxlab.space/utils/VLMEval/MMVet_Hard.tsv' |
| | } |
| | DATASET_MD5 = {'MMVet': '748aa6d4aa9d4de798306a63718455e3', 'MMVet_Hard': '63a598819a936a2e77c410a78a21ff16'} |
| |
|
| | |
| | @classmethod |
| | def evaluate(self, eval_file, **judge_kwargs): |
| | from .utils.mmvet import MMVet_auxeval, MMVet_acc |
| |
|
| | suffix = eval_file.split('.')[-1] |
| | model = judge_kwargs['model'] |
| | storage = eval_file.replace(f'.{suffix}', f'_{model}.xlsx') |
| | tmp_file = eval_file.replace(f'.{suffix}', f'_{model}.pkl') |
| | nproc = judge_kwargs.pop('nproc', 4) |
| | if not osp.exists(storage): |
| | data = load(eval_file) |
| | model = build_judge(max_tokens=3, **judge_kwargs) |
| | assert model.working(), ('MMVet evaluation requires a working OPENAI API\n' + DEBUG_MESSAGE) |
| |
|
| | lt = len(data) |
| | lines = [data.iloc[i] for i in range(lt)] |
| | tups = [(model, line) for line in lines] |
| | indices = [line['index'] for line in lines] |
| |
|
| | ans = load(tmp_file) if osp.exists(tmp_file) else {} |
| | tups = [x for x, i in zip(tups, indices) if i not in ans] |
| | indices = [i for i in indices if i not in ans] |
| |
|
| | if len(indices): |
| | new_results = track_progress_rich( |
| | MMVet_auxeval, |
| | tups, |
| | nproc=nproc, |
| | chunksize=nproc, |
| | keys=indices, |
| | save=tmp_file, |
| | ) |
| | ans = load(tmp_file) |
| | for k, v in zip(indices, new_results): |
| | assert k in ans |
| | assert ans[k]['log'] == v['log'] and ans[k]['score'] == v['score'] |
| | data['score'] = [ans[idx]['score'] for idx in data['index']] |
| | data['log'] = [ans[idx]['log'] for idx in data['index']] |
| | dump(data, storage) |
| |
|
| | score, score_fine = MMVet_acc(storage) |
| | score_pth = storage.replace('.xlsx', '_score.csv') |
| | score_fine_pth = storage.replace('.xlsx', '_score_fine.csv') |
| | dump(score, score_pth) |
| | dump(score_fine, score_fine_pth) |
| | return score |
| |
|
| |
|
| | class MTVQADataset(ImageBaseDataset): |
| | TYPE = 'VQA' |
| | DATASET_URL = {'MTVQA_TEST': 'https://opencompass.openxlab.space/utils/VLMEval/MTVQA_TEST.tsv'} |
| | DATASET_MD5 = {'MTVQA_TEST': 'd87c17dbab934b7cd89c0a3c1c5657f4'} |
| |
|
| | @classmethod |
| | def evaluate(self, eval_file, **judge_kwargs): |
| | data = load(eval_file) |
| | assert 'answer' in data and 'prediction' in data and 'category' in data |
| | data['prediction'] = [str(x) for x in data['prediction']] |
| | data['answer'] = [str(x) for x in data['answer']] |
| | if 'split' in data: |
| | assert np.all([x.lower() == 'test' for x in data['split']]), 'We only support MTVQA_TEST for now. ' |
| | lt = len(data) |
| | category_scores = defaultdict(list) |
| | for i in range(lt): |
| | line = data.iloc[i] |
| | ans = line['answer'].strip().lower().replace('.', '') |
| | pred = line['prediction'].strip().lower().replace('.', '') |
| | cate = line['category'] |
| | score = 1.0 if ans in pred else 0.0 |
| | category_scores[cate].append(score) |
| | category_scores['Average'].append(score) |
| | |
| | category_averages = {category: np.mean(scores) * 100 for category, scores in category_scores.items()} |
| |
|
| | suffix = eval_file.split('.')[-1] |
| | result_file = eval_file.replace(f'.{suffix}', '_acc.json') |
| | dump(category_averages, result_file) |
| |
|
| | return category_averages |
| |
|
| | |
| | def build_prompt(self, line): |
| | msgs = super().build_prompt(line) |
| | assert sum([x['type'] == 'text' for x in msgs]) == 1 |
| | for item in msgs: |
| | if item['type'] == 'text': |
| | item['value'] += '\nAnswer the question using a word or phrase in the language of the question.' |
| | return msgs |
| |
|
| |
|
| | class TableVQABench(ImageBaseDataset): |
| | TYPE = 'VQA' |
| | DATASET_URL = { |
| | 'TableVQABench': 'https://pai-aigc-photog.oss-cn-hangzhou.aliyuncs.com/mentor-vil/datasets/tablevqa-bench.tsv' |
| | } |
| | DATASET_MD5 = {'TableVQABench': '2550adc61bdc82d8e62f3b003de7c62d'} |
| |
|
| | from .utils.tablevqabench import FINTABNETQA_PROMPT, VTABFACT_PROMPT, VWTQ_PROMPT |
| |
|
| | |
| | @classmethod |
| | def evaluate(self, eval_file, **judge_kwargs): |
| | import pandas as pd |
| | from .utils.tablevqabench import evaluate_fintabnet, evaluate_tabfact, evaluate_wtq |
| |
|
| | data = load(eval_file) |
| | assert 'answer' in data and 'prediction' in data |
| |
|
| | data['prediction'] = data['prediction'].str.replace('^Answer: ', '', regex=True) |
| | data_group = dict(tuple(data.groupby('split'))) |
| | eval_result = {'split': [], 'average_scores': []} |
| | for split in ['fintabnetqa', 'vtabfact', 'vwtq', 'vwtq_syn']: |
| | data_split = data_group[split].to_dict(orient='records') |
| | if split == 'fintabnetqa': |
| | split_eval_meta = evaluate_fintabnet(data_split, ['accuracy']) |
| | elif split == 'vtabfact': |
| | split_eval_meta = evaluate_tabfact(data_split, ['accuracy']) |
| | elif split == 'vwtq' or split == 'vwtq_syn': |
| | split_eval_meta = evaluate_wtq(data_split, ['accuracy']) |
| | eval_result['split'].append(split) |
| | eval_result['average_scores'].append(split_eval_meta['average_scores']) |
| |
|
| | suffix = eval_file.split('.')[-1] |
| | result_file = eval_file.replace(f'.{suffix}', '_acc.csv') |
| | eval_result = pd.DataFrame(eval_result) |
| | dump(eval_result, result_file) |
| |
|
| | return eval_result |
| |
|
| | |
| | def build_prompt(self, line): |
| | msgs = super().build_prompt(line) |
| | assert sum([x['type'] == 'text' for x in msgs]) == 1 |
| | for item in msgs: |
| | if item['type'] == 'text': |
| | if line['split'] == 'fintabnetqa': |
| | item['value'] = self.FINTABNETQA_PROMPT.format_map({'question': item['value']}) |
| | elif line['split'] == 'vtabfact': |
| | item['value'] = self.VTABFACT_PROMPT.format_map({'question': item['value']}) |
| | elif line['split'] == 'vwtq_syn' or line['split'] == 'vwtq': |
| | item['value'] = self.VWTQ_PROMPT.format_map({'question': item['value']}) |
| | return msgs |
| |
|
| |
|
| | class CustomVQADataset(ImageBaseDataset): |
| | TYPE = 'VQA' |
| |
|
| | def load_data(self, dataset): |
| | data_path = osp.join(LMUDataRoot(), f'{dataset}.tsv') |
| |
|
| | if file_size(data_path, 'GB') > 1: |
| | local_path = data_path.replace('.tsv', '_local.tsv') |
| | if not osp.exists(local_path) or os.environ.get('FORCE_LOCAL', None): |
| | from ..tools import LOCALIZE |
| |
|
| | LOCALIZE(data_path, local_path) |
| | data_path = local_path |
| | return load(data_path) |
| |
|
| | def evaluate(self, eval_file, **judge_kwargs): |
| | raise NotImplementedError |
| |
|
| |
|
| | class CRPE(ImageBaseDataset): |
| | TYPE = 'VQA' |
| | DATASET_URL = { |
| | 'CRPE_EXIST': 'https://huggingface.co/datasets/petter12321/crpe_vlmevalkit/resolve/main/CRPE_EXIST.tsv', |
| | 'CRPE_RELATION': 'https://huggingface.co/datasets/petter12321/crpe_vlmevalkit/resolve/main/CRPE_RELATION.tsv' |
| | } |
| | DATASET_MD5 = { |
| | 'CRPE_EXIST': '315584e23ac1ff7f8719ed3b7ad90f08', |
| | 'CRPE_RELATION': 'bad7094cde0b572288f4b119c2d0c656'} |
| |
|
| | @classmethod |
| | def evaluate(self, eval_file, **judge_kwargs): |
| | from .utils.crpe import is_correct |
| | |
| | |
| | score = { |
| | 'exist': 0, |
| | 'subject': 0, |
| | 'predicate': 0, |
| | 'object': 0, |
| | 'total': 0, |
| | } |
| | num = { |
| | 'exist': 0, |
| | 'subject': 0, |
| | 'predicate': 0, |
| | 'object': 0, |
| | 'total': 0, |
| | } |
| | final_score_dict = { |
| | 'exist': 0, |
| | 'subject': 0, |
| | 'predicate': 0, |
| | 'object': 0, |
| | 'total': 0, |
| | } |
| | data = load(eval_file) |
| | lt = len(data) |
| | lines = [data.iloc[i] for i in range(lt)] |
| | for i in tqdm(range(len(lines))): |
| | line = lines[i] |
| | predict = str(line['prediction']) |
| | answers = str(line['answer']) |
| | |
| | |
| | category = line['category'] |
| | if is_correct(answers, predict): |
| | score[category] += 1 |
| | score['total'] += 1 |
| | num[category] += 1 |
| | num['total'] += 1 |
| |
|
| | for category in ['exist', 'subject', 'predicate', 'object', 'total']: |
| | if num[category] != 0: |
| | final_score_dict[category] = score[category] / num[category] |
| | else: |
| | final_score_dict[category] = None |
| |
|
| | score_pth = eval_file.replace('.xlsx', '_score.json') |
| | dump(final_score_dict, score_pth) |
| | return final_score_dict |
| |
|
| | def build_prompt(self, line): |
| | ROOT = LMUDataRoot() |
| | msgs = super().build_prompt(line) |
| | for msg in msgs: |
| | if msg['type'] == 'image': |
| | msg['value'] = osp.join(osp.join(ROOT, 'images', self.dataset_name), msg['value']) |
| | return msgs |
| |
|
| |
|
| | class QSpatial(ImageBaseDataset): |
| | TYPE = 'VQA' |
| | DATASET_URL = { |
| | 'QSpatial_plus': '', |
| | 'QSpatial_scannet': '' |
| | } |
| |
|
| | |
| | |
| | |
| | qspatial_root = "TO_BE_REPLACED_WITH_THE_PATH_TO_QSPATIAL_DATASET" |
| | url = "https://raw.githubusercontent.com/andrewliao11/Q-Spatial-Bench-code/refs/heads/main/prompt_templates/" |
| |
|
| | def post_build(self, dataset): |
| | |
| |
|
| | links = [ |
| | self.url + "system_prompt.txt", |
| | self.url + "spatial_prompt_single.txt", |
| | self.url + "spatial_prompt_steps.txt", |
| | self.url + "standard_prompt.txt", |
| | self.url + "zero_shot_prompt.txt" |
| | ] |
| | with tempfile.TemporaryDirectory() as temp_dir: |
| | for link in links: |
| | tgt_path = os.path.join(temp_dir, link.split("/")[-1]) |
| | os.system(f"wget {link} -O {tgt_path}") |
| |
|
| | self.system_prompt = open(os.path.join(temp_dir, "system_prompt.txt")).read() |
| | self._prompt_templates = dict( |
| | spatial_prompt_single=open(os.path.join(temp_dir, "spatial_prompt_single.txt")).read(), |
| | spatial_prompt_steps=open(os.path.join(temp_dir, "spatial_prompt_steps.txt")).read(), |
| | standard_prompt=open(os.path.join(temp_dir, "standard_prompt.txt")).read(), |
| | zero_shot_prompt=open(os.path.join(temp_dir, "zero_shot_prompt.txt")).read(), |
| | ) |
| |
|
| | |
| | def build_prompt(self, line): |
| | from jinja2.sandbox import SandboxedEnvironment |
| | text_prompt_template = self._prompt_templates["spatial_prompt_single"] |
| | env = SandboxedEnvironment() |
| | text_prompt = env.from_string(text_prompt_template).render(question=line["question"]) |
| | tgt_path = self.dump_image(line) |
| |
|
| | msgs = [] |
| | if isinstance(tgt_path, list): |
| | msgs.extend([dict(type='image', value=p) for p in tgt_path]) |
| | else: |
| | msgs = [dict(type='image', value=tgt_path)] |
| |
|
| | msgs.append(dict(type='text', value=f"{self.system_prompt}\n{text_prompt}")) |
| | return msgs |
| |
|
| | |
| | def load_data(self, dataset): |
| | import io |
| | import pandas as pd |
| | from datasets import load_dataset |
| |
|
| | hf_dataset = load_dataset("andrewliao11/Q-Spatial-Bench", split=dataset) |
| | df = hf_dataset.to_pandas() |
| |
|
| | df.reset_index(drop=True, inplace=True) |
| | df['index'] = df.index |
| | df['answer'] = list(zip(df['answer_value'], df['answer_unit'])) |
| | df = df[['index'] + [col for col in df.columns if col != 'index']] |
| |
|
| | if dataset == "QSpatial_scannet": |
| | df = df.drop(columns=["image"]) |
| | df["image"] = [Image.open(os.path.join(self.qspatial_root, image_path)) for image_path in df["image_path"]] |
| | else: |
| | df["image"] = [Image.open(io.BytesIO(image_dict["bytes"])) for image_dict in df["image"]] |
| |
|
| | df["image"] = [encode_image_to_base64(image) for image in df["image"]] |
| | return df |
| |
|
| | @classmethod |
| | def get_multiplier(self, unit): |
| |
|
| | unit = unit.lower() |
| | if unit in ["meters", "meter", "m", "metre", "metres"]: |
| | multiplier = 100 |
| | elif unit in ["centimeters", "centimeter", "cm"]: |
| | multiplier = 1 |
| | elif unit in ["feet", "foot", "ft"]: |
| | multiplier = 30.48 |
| | elif unit in ["inch", "inches", "in"]: |
| | multiplier = 2.54 |
| | elif unit in ["mm"]: |
| | multiplier = 0.1 |
| | else: |
| | print(f"Unknown unit: {unit}") |
| | multiplier = 0. |
| |
|
| | return multiplier |
| |
|
| | @classmethod |
| | def parse_string(self, input_str): |
| | |
| | match = re.match(r'\(([\d.-]+), (.+)\)', input_str) |
| | if match: |
| | number_part = match.group(1) |
| | text = match.group(2) |
| |
|
| | if '-' in number_part: |
| | start, end = map(float, number_part.split('-')) |
| | number = (start + end) / 2 |
| | else: |
| | number = float(number_part) |
| |
|
| | return number * self.get_multiplier(text) |
| | else: |
| | print(f"Unable to parse the input string {input_str}") |
| | return 0 |
| |
|
| | @classmethod |
| | def parse_prediction(self, vlm_response): |
| | |
| | pattern = r'scalar{([^}]*)}' |
| | str_inside_scalar_boxes = re.findall(pattern, vlm_response)[-1] |
| | scalar_list = re.findall(r'\d+\.?\d*', str_inside_scalar_boxes) |
| | parsed_scalar = np.array(scalar_list).astype(float).mean() |
| |
|
| | |
| | pattern = r'distance_unit{([^}]*)}' |
| | str_inside_unit_boxes = re.findall(pattern, vlm_response) |
| | parsed_unit = str_inside_unit_boxes[-1] |
| |
|
| | pred_value_in_cms = parsed_scalar * self.get_multiplier(parsed_unit) |
| | return pred_value_in_cms |
| |
|
| | |
| | @classmethod |
| | def evaluate(self, eval_file, **judge_kwargs): |
| |
|
| | data = load(eval_file) |
| | if "model" in judge_kwargs: |
| | from .utils.qspatial import QSpatial_auxeval |
| |
|
| | |
| | model = judge_kwargs['model'] |
| | suffix = eval_file.split('.')[-1] |
| | storage = eval_file.replace(f'.{suffix}', f'_{model}.xlsx') |
| | tmp_file = eval_file.replace(f'.{suffix}', f'_{model}.pkl') |
| | nproc = judge_kwargs.pop('nproc', 4) |
| |
|
| | if not osp.exists(storage): |
| | model = build_judge(max_tokens=128, **judge_kwargs) |
| |
|
| | assert model.working(), ('Evaluation requires a working OPENAI API\n' + DEBUG_MESSAGE) |
| | lt = len(data) |
| | lines = [data.iloc[i] for i in range(lt)] |
| | tups = [(model, line) for line in lines] |
| | indices = [line['index'] for line in lines] |
| |
|
| | ans = {} |
| | if osp.exists(tmp_file): |
| | ans = load(tmp_file) |
| | tups = [x for x, i in zip(tups, indices) if i not in ans] |
| | indices = [i for i in indices if i not in ans] |
| |
|
| | if len(indices): |
| | new_results = track_progress_rich( |
| | QSpatial_auxeval, |
| | tups, |
| | nproc=nproc, |
| | chunksize=nproc, |
| | keys=indices, |
| | save=tmp_file, |
| | ) |
| | ans = load(tmp_file) |
| | for k, v in zip(indices, new_results): |
| | assert k in ans |
| | assert ans[k]['log'] == v['log'] and ans[k]['res'] == v['res'] |
| |
|
| | data['res'] = [ans[idx]['res'] for idx in data['index']] |
| | data['log'] = [ans[idx]['log'] for idx in data['index']] |
| | dump(data, storage) |
| |
|
| | data = load(storage) |
| |
|
| | pred_value_in_cms = [] |
| | for res in data["res"]: |
| | try: |
| | pred_value_in_cms.append(self.parse_string(res)) |
| | except ValueError: |
| | pred_value_in_cms.append(0.) |
| |
|
| | pred_value_in_cms = np.array(pred_value_in_cms) + 1e-8 |
| | else: |
| | |
| | pred_value_in_cms = [] |
| | n_errors_in_parsing = 0 |
| | for pred in data["prediction"]: |
| | try: |
| | parsed_value = self.parse_prediction(pred) |
| | except IndexError: |
| | n_errors_in_parsing += 1 |
| | parsed_value = 1e-8 |
| |
|
| | pred_value_in_cms.append(parsed_value) |
| |
|
| | print(f"Encounter {n_errors_in_parsing} errors in parsing") |
| | pred_value_in_cms = np.array(pred_value_in_cms) + 1e-8 |
| |
|
| | |
| | ground_truth_value_in_cms = [] |
| | for answer in data["answer"]: |
| | value, unit = eval(answer) |
| | ground_truth_value_in_cms.append(value * self.get_multiplier(unit)) |
| | ground_truth_value_in_cms = np.array(ground_truth_value_in_cms) + 1e-8 |
| |
|
| | |
| | pred_gt = pred_value_in_cms / ground_truth_value_in_cms |
| | gt_pred = ground_truth_value_in_cms / pred_value_in_cms |
| | delta_2 = np.stack([pred_gt, gt_pred]).max(0) < 2. |
| | delta_1_point_5 = np.stack([pred_gt, gt_pred]).max(0) < 1.5 |
| |
|
| | data["eval_score_delta_2"] = delta_2 |
| | data["eval_score_delta_1_point_5"] = delta_1_point_5 |
| |
|
| | final_score_dict = { |
| | "delta_2": delta_2.mean(), |
| | "delta_1_point_5": delta_1_point_5.mean() |
| | } |
| | for question_type in set(data["question_type"]): |
| | filtered_data = data[data["question_type"] == question_type] |
| | delta_2_per_question_type = filtered_data["eval_score_delta_2"].mean() |
| | delta_1_point_5_per_question_type = filtered_data["eval_score_delta_1_point_5"].mean() |
| | final_score_dict.update({f"{question_type}_delta_2": delta_2_per_question_type}) |
| | final_score_dict.update({f"{question_type}_delta_1_point_5": delta_1_point_5_per_question_type}) |
| |
|
| | score_pth = eval_file.replace('.xlsx', '_score.json') |
| | dump(final_score_dict, score_pth) |
| | return final_score_dict |
| |
|
| |
|
| | class MMNIAH(ImageBaseDataset): |
| | TYPE = 'VQA' |
| | DATASET_URL = { |
| | 'MM_NIAH_VAL': |
| | 'https://huggingface.co/datasets/petter12321/MM-NIAH-VLMEvalKit/resolve/main/MM_NIAH_VAL.tsv', |
| | 'MM_NIAH_TEST': |
| | ['https://huggingface.co/datasets/petter12321/MM-NIAH-VLMEvalKit/resolve/main/part-aa', |
| | 'https://huggingface.co/datasets/petter12321/MM-NIAH-VLMEvalKit/resolve/main/part-ab', |
| | 'https://huggingface.co/datasets/petter12321/MM-NIAH-VLMEvalKit/resolve/main/part-ac', |
| | 'https://huggingface.co/datasets/petter12321/MM-NIAH-VLMEvalKit/resolve/main/part-ad', |
| | 'https://huggingface.co/datasets/petter12321/MM-NIAH-VLMEvalKit/resolve/main/part-ae']} |
| | DATASET_MD5 = {'MM_NIAH_VAL': '27e5a8c3cef7746cb38f89cd86c474c5', |
| | 'MM_NIAH_TEST': 'f490eb2a43096307465fe9e7ef13497c'} |
| |
|
| | def prepare_tsv(self, url, file_md5=None): |
| | import os |
| | data_root = LMUDataRoot() |
| | os.makedirs(data_root, exist_ok=True) |
| | update_flag = False |
| | file_name = 'MM_NIAH_VAL.tsv' if 'MM_NIAH_VAL' in url else 'MM_NIAH_TEST.tsv' |
| | data_path = osp.join(data_root, file_name) |
| | if osp.exists(data_path) and (file_md5 is None or md5(data_path) == file_md5): |
| | pass |
| | elif file_name == 'MM_NIAH_TEST.tsv': |
| | warnings.warn('The dataset tsv is not downloaded') |
| | for i in range(len(url)): |
| | if osp.exists(osp.join(data_root, 'part-a' + chr(ord('a') + i))): |
| | print('part_a' + chr(ord('a') + i) + ' is existed') |
| | continue |
| | download_file(url[i], data_path) |
| | file_prefix = 'part-' |
| | output_file = data_path |
| | split_files = sorted([f for f in os.listdir(data_root) if f.startswith(file_prefix)]) |
| | with open(output_file, 'wb') as outfile: |
| | |
| | for filename in split_files: |
| | with open(osp.join(data_root, filename), 'rb') as infile: |
| | outfile.write(infile.read()) |
| | update_flag = True |
| | else: |
| | warnings.warn('The dataset tsv is not downloaded') |
| | download_file(url, data_path) |
| | update_flag = True |
| |
|
| | if file_size(data_path, 'GB') > 1: |
| | local_path = data_path.replace('.tsv', '_local.tsv') |
| | if not osp.exists(local_path) or os.environ.get('FORCE_LOCAL', None) or update_flag: |
| | from ..tools import LOCALIZE |
| | LOCALIZE(data_path, local_path) |
| | data_path = local_path |
| | return load(data_path) |
| |
|
| | @classmethod |
| | def evaluate(self, eval_file, **judge_kwargs): |
| | from .utils.mmniah import is_correct |
| | |
| | |
| | MMNIAH_score = { |
| | 'count-text': 0, |
| | 'find-image': 0, |
| | 'find-text': 0, |
| | 'infer-choose': 0, |
| | 'count-image': 0, |
| | 'visual-reasoning': 0, |
| | 'total': 0, |
| | } |
| | MMNIAH_num = { |
| | 'count-text': 0, |
| | 'find-image': 0, |
| | 'find-text': 0, |
| | 'infer-choose': 0, |
| | 'count-image': 0, |
| | 'visual-reasoning': 0, |
| | 'total': 0, |
| | } |
| | final_score_dict = { |
| | 'count-text': 0, |
| | 'find-image': 0, |
| | 'find-text': 0, |
| | 'infer-choose': 0, |
| | 'count-image': 0, |
| | 'visual-reasoning': 0, |
| | 'total': 0, |
| | } |
| | data = load(eval_file) |
| | lt = len(data) |
| | lines = [data.iloc[i] for i in range(lt)] |
| | for i in tqdm(range(len(lines))): |
| | line = lines[i] |
| | predict = line['prediction'] |
| | answers = line['answer'] |
| | category = line['category'] |
| | if category in ['visual-reasoning', 'find-image']: |
| | answers = int(answers) |
| | if is_correct(answers, predict): |
| | MMNIAH_score[category] += 1 |
| | MMNIAH_score['total'] += 1 |
| | MMNIAH_num[category] += 1 |
| | MMNIAH_num['total'] += 1 |
| |
|
| | for category in ['find-image', 'count-text', 'find-text', |
| | 'infer-choose', 'count-image', 'visual-reasoning', 'total']: |
| | if MMNIAH_num[category] != 0: |
| | final_score_dict[category] = MMNIAH_score[category] / MMNIAH_num[category] |
| | else: |
| | final_score_dict[category] = None |
| |
|
| | score_pth = eval_file.replace('.xlsx', '_score.json') |
| | dump(final_score_dict, score_pth) |
| | return final_score_dict |
| |
|
| | def build_prompt(self, line): |
| | msgs = super().build_prompt(line) |
| | if isinstance(line, int): |
| | line = self.data.iloc[line] |
| | totalchoice = line['multi-choice options'] |
| | totalchoice = eval(totalchoice) |
| | |
| | |
| | context = msgs[-1]['value'] |
| | context = eval(context) |
| | question = context[0] + '\n' + context[1] |
| | |
| | tgt_path = [] |
| | for i in range(len(msgs) - 1): |
| | tgt_path.append(msgs[i]['value']) |
| | choices = totalchoice[0] |
| | choices_image = totalchoice[1] |
| | if choices: |
| | for c_idx, c in enumerate(choices): |
| | question = f"{question}\n{chr(c_idx + ord('A'))}. {c}" |
| | question += "\nAnswer with the option's letter from the given choices directly." |
| | elif choices_image: |
| | for c_idx in range(len(choices_image)): |
| | question = f"{question}\n{chr(c_idx + ord('A'))}. <image>" |
| | question += "\nAnswer with the option's letter from the given choices directly." |
| | else: |
| | question += '\nAnswer the question using a single word or phrase.' |
| | question = '<start>' + question + '<end>' |
| | question = question.split('<image>') |
| | if choices_image: |
| | for i in range(len(question) - 5): |
| | question[i] = question[i] + '\n<image>' |
| | for i in range(len(question) - 5, len(question) - 1): |
| | question[i] = question[i] + '<image>' |
| | else: |
| | for i in range(len(question) - 1): |
| | question[i] = question[i] + '\n<image>' |
| | assert len(tgt_path) + 1 == len(question) |
| | context = [] |
| | for i in range(len(tgt_path)): |
| | context.append(question[i]) |
| | context.append(tgt_path[i]) |
| | context.append(question[-1]) |
| | context[0] = context[0][7:] |
| | context[-1] = context[-1][:-5] |
| | msgs = [] |
| | for i in range(len(context)): |
| | if i % 2 == 0: |
| | msgs.append(dict(type='text', value=context[i])) |
| | else: |
| | ROOT = LMUDataRoot() |
| | msgs.append(dict(type='image', value=osp.join(osp.join(ROOT, 'images', self.dataset_name), context[i]))) |
| | for element in msgs: |
| | if element['value'] == '': |
| | msgs.remove(element) |
| | return msgs |
| |
|