VLMEvalKit / vlmeval /api /lmdeploy.py
Racktic's picture
Upload folder using huggingface_hub
b5beb60 verified
# from http import HTTPStatus
import os
import requests
from ..dataset import DATASET_TYPE, DATASET_MODALITY
from vlmeval.api.base import BaseAPI
from vlmeval.smp import *
class InternVL2_PromptUtil:
def __init__(self, use_mpo_prompt=False):
self.use_mpo_prompt = use_mpo_prompt
def dump_image(self, line, dataset):
return self.dump_image_func(line)
def use_custom_prompt(self, dataset):
assert dataset is not None
assert DATASET_MODALITY(dataset) != 'VIDEO', 'not supported'
if listinstr(['MMDU', 'MME-RealWorld', 'MME-RealWorld-CN'], dataset):
# For Multi-Turn we don't have custom prompt
return False
if DATASET_MODALITY(dataset) == 'VIDEO':
# For Video benchmarks we don't have custom prompt at here
return False
else:
return True
def build_prompt(self, line, dataset=None):
assert self.use_custom_prompt(dataset)
assert dataset is None or isinstance(dataset, str)
from ..vlm.internvl.utils import (build_multi_choice_prompt,
build_mcq_cot_prompt,
build_qa_cot_prompt,
build_mpo_prompt,
reorganize_prompt)
tgt_path = self.dump_image(line, dataset)
max_num = self.get_max_num(dataset)
if dataset is not None and DATASET_TYPE(dataset) == 'Y/N':
question = line['question']
if listinstr(['MME'], dataset):
prompt = question + ' Answer the question using a single word or phrase.'
elif listinstr(['HallusionBench', 'AMBER'], dataset):
prompt = question + ' Please answer yes or no. Answer the question using a single word or phrase.'
else:
prompt = question
elif dataset is not None and DATASET_TYPE(dataset) == 'MCQ':
prompt = build_multi_choice_prompt(line, dataset)
if os.getenv('USE_COT') == '1':
prompt = build_mcq_cot_prompt(line, prompt)
elif dataset is not None and DATASET_TYPE(dataset) == 'VQA':
question = line['question']
if listinstr(['LLaVABench', 'WildVision'], dataset):
prompt = question + '\nAnswer this question in detail.'
elif listinstr(['OCRVQA', 'TextVQA', 'ChartQA', 'DocVQA', 'InfoVQA', 'OCRBench',
'DUDE', 'SLIDEVQA', 'GQA', 'MMLongBench_DOC'], dataset):
prompt = question + '\nAnswer the question using a single word or phrase.'
elif listinstr(['MathVista', 'MathVision', 'VCR', 'MTVQA', 'MMVet', 'MathVerse',
'MMDU', 'CRPE', 'MIA-Bench', 'MM-Math', 'DynaMath', 'QSpatial', 'WeMath', 'LogicVista'], dataset):
prompt = question
if os.getenv('USE_COT') == '1':
prompt = build_qa_cot_prompt(line, prompt)
else:
prompt = question + '\nAnswer the question using a single word or phrase.'
else:
# VQA_ex_prompt: OlympiadBench, VizWiz
prompt = line['question']
if os.getenv('USE_COT') == '1':
prompt = build_qa_cot_prompt(line, prompt)
message = [dict(type='text', value=prompt)]
image_num = len(tgt_path)
max_num = max(1, min(max_num, 64 // image_num))
# TODO:support upscale_flag
message.extend([dict(type='image', value=s, max_dynamic_patch=max_num) for s in tgt_path])
if self.use_mpo_prompt:
message = build_mpo_prompt(message, line, dataset)
# reorganize_prompt
prompt = reorganize_prompt(message, image_num, dataset=dataset)
prompt.replace('<image>', '<IMAGE_TOKEN>')
message[0] = dict(type='text', value=prompt)
return message
def get_max_num(self, dataset):
assert dataset is not None
res_1_datasets = ['MMBench-Video', 'Video-MME', 'MVBench', 'Video', 'WorldSense']
res_12_datasets = ['ChartQA_TEST', 'MMMU_DEV_VAL', 'MMMU_TEST', 'MME-RealWorld',
'VCR_EN', 'VCR_ZH', 'OCRVQA']
res_18_datasets = ['DocVQA_VAL', 'DocVQA_TEST', 'DUDE', 'MMLongBench_DOC', 'SLIDEVQA']
res_24_datasets = ['InfoVQA_VAL', 'InfoVQA_TEST', 'OCRBench', 'HRBench4K', 'HRBench8K']
if listinstr(res_1_datasets, dataset):
return 1
elif listinstr(res_12_datasets, dataset):
return 12
elif listinstr(res_18_datasets, dataset):
return 18
elif listinstr(res_24_datasets, dataset):
return 24
else:
return 6
class CogVLM2_PromptUtil:
def dump_image(self, line, dataset):
return self.dump_image_func(line)
def use_custom_prompt(self, dataset):
assert dataset is not None
if DATASET_TYPE(dataset) in 'MCQ':
return True
return False
def build_prompt(self, line, dataset=None):
assert dataset is None or isinstance(dataset, str)
assert self.use_custom_prompt(dataset)
tgt_path = self.dump_image(line, dataset)
if dataset is not None and DATASET_TYPE(dataset) == 'MCQ':
question = line['question']
hint = line['hint'] if ('hint' in line and not pd.isna(line['hint'])) else None
if hint is not None:
question = hint + '\n' + question
option_candidate = string.ascii_uppercase
options = {
cand: line[cand]
for cand in option_candidate
if cand in line and not pd.isna(line[cand])
}
for key, item in options.items():
question += f'\n{key}. {item}'
prompt = question
if not cn_string(prompt):
prompt = prompt + '\n' + "Answer with the option's letter from the given choices directly."
else:
prompt = prompt + '\n' + '请直接回答选项字母。'
else:
prompt = line['question']
message = [dict(type='text', value=prompt)]
message.extend([dict(type='image', value=p) for p in tgt_path])
return message
class LMDeployWrapper(BaseAPI):
is_api: bool = True
custom_prompt: str = None
prompt_map = {
'cogvlm2': CogVLM2_PromptUtil(),
'internvl2': InternVL2_PromptUtil(),
'internvl2-mpo-cot': InternVL2_PromptUtil(use_mpo_prompt=True),
}
def __init__(self,
retry: int = 5,
wait: int = 5,
key: str = 'sk-123456',
verbose: bool = True,
temperature: float = 0.0,
timeout: int = 60,
api_base: str = None,
system_prompt: str = None,
max_tokens: int = 1024,
**kwargs):
self.fail_msg = 'Failed to obtain answer via API. '
self.max_tokens = max_tokens
self.timeout = timeout
key = os.environ.get('LMDEPLOY_API_KEY', key)
api_base = os.environ.get('LMDEPLOY_API_BASE', api_base)
assert key is not None, 'Please set the environment variable LMDEPLOY_API_KEY.'
assert api_base is not None, 'Please set the environment variable LMDEPLOY_API_BASE.'
self.key = key
self.api_base = api_base
super().__init__(wait=wait, retry=retry, system_prompt=system_prompt, verbose=verbose, **kwargs)
model_url = ''.join([api_base.split('v1')[0], 'v1/models'])
resp = requests.get(model_url)
self.model = resp.json()['data'][0]['id']
self.logger.info(f'lmdeploy evaluate model: {self.model}')
self.set_prompt_pattern(self.model)
if hasattr(self, 'custom_prompt'):
self.logger.info(f'using custom prompt {self.custom_prompt}')
self.temperature = temperature
self.logger.info(f'Init temperature: {self.temperature}')
def set_dump_image(self, dump_image_func):
if self.custom_prompt in self.prompt_map:
self.prompt_map[self.custom_prompt].dump_image_func = dump_image_func
self.dump_image_func = dump_image_func
def use_custom_prompt(self, dataset):
if self.custom_prompt in self.prompt_map:
return self.prompt_map[self.custom_prompt].use_custom_prompt(dataset)
return False
def build_prompt(self, line, dataset=None):
if self.custom_prompt in self.prompt_map:
return self.prompt_map[self.custom_prompt].build_prompt(line, dataset)
raise NotImplementedError
def set_prompt_pattern(self, model_name):
if 'Phi-3.5-Vision'.lower() in model_name.lower():
self.max_tokens = 1000
self.temperature = 0.0
if 'cogvlm2-llama3-chat-19B'.lower() in model_name.lower():
self.max_tokens = 2048
self.temperature = 0.0
self.custom_prompt = 'cogvlm2'
if 'InternVL2'.lower() in model_name.lower():
self.max_tokens = 1024
self.temperature = 0.0
if 'mpo' in model_name.lower():
self.max_tokens = 4096
self.logger.info('Use custom prompt internvl2-mpo-cot')
self.custom_prompt = 'internvl2-mpo-cot'
else:
self.logger.info('Use custom prompt internvl2')
self.custom_prompt = 'internvl2'
if 'internvl2-8b-mpo-cot'.lower() in model_name.lower():
self.use_mpo_prompt = True
self.max_tokens = 1024
self.temperature = 0.0
self.logger.info('Use custom prompt internvl2-mpo-cot')
self.custom_prompt = 'internvl2-mpo-cot'
if 'qvq'.lower() in model_name.lower():
self.max_tokens = 4096
self.temperature = 0.0
self.logger.info('QVQ model detected, do not use custom prompt')
def prepare_itlist(self, inputs):
assert np.all([isinstance(x, dict) for x in inputs])
has_images = np.sum([x['type'] == 'image' for x in inputs])
if has_images:
content_list = []
for msg in inputs:
if msg['type'] == 'text':
content_list.append(dict(type='text', text=msg['value']))
elif msg['type'] == 'image':
from PIL import Image
img = Image.open(msg['value'])
b64 = encode_image_to_base64(img)
extra_args = msg.copy()
extra_args.pop('type')
extra_args.pop('value')
img_struct = dict(url=f'data:image/jpeg;base64,{b64}', **extra_args)
content_list.append(dict(type='image_url', image_url=img_struct))
else:
assert all([x['type'] == 'text' for x in inputs])
text = '\n'.join([x['value'] for x in inputs])
content_list = [dict(type='text', text=text)]
return content_list
def prepare_inputs(self, inputs):
input_msgs = []
if self.system_prompt is not None:
input_msgs.append(dict(role='system', content=self.system_prompt))
assert isinstance(inputs, list) and isinstance(inputs[0], dict)
assert np.all(['type' in x for x in inputs]) or np.all(['role' in x for x in inputs]), inputs
if 'role' in inputs[0]:
assert inputs[-1]['role'] == 'user', inputs[-1]
for item in inputs:
input_msgs.append(dict(role=item['role'], content=self.prepare_itlist(item['content'])))
else:
input_msgs.append(dict(role='user', content=self.prepare_itlist(inputs)))
return input_msgs
def generate_inner(self, inputs, **kwargs) -> str:
input_msgs = self.prepare_inputs(inputs)
temperature = kwargs.pop('temperature', self.temperature)
self.logger.info(f'Generate temperature: {temperature}')
max_tokens = kwargs.pop('max_tokens', self.max_tokens)
headers = {'Content-Type': 'application/json', 'Authorization': f'Bearer {self.key}'}
payload = dict(
model=self.model,
messages=input_msgs,
max_tokens=max_tokens,
n=1,
temperature=temperature,
**kwargs)
response = requests.post(
self.api_base,
headers=headers, data=json.dumps(payload), timeout=self.timeout * 1.1)
ret_code = response.status_code
ret_code = 0 if (200 <= int(ret_code) < 300) else ret_code
answer = self.fail_msg
try:
resp_struct = json.loads(response.text)
answer = resp_struct['choices'][0]['message']['content'].strip()
# for internvl2-8b-mpo-cot
if getattr(self, 'use_mpo_prompt', False):
from ..vlm.internvl.utils import mpo_post_processing
answer = mpo_post_processing(answer, kwargs.get('dataset'))
except:
pass
return ret_code, answer, response
class LMDeployAPI(LMDeployWrapper):
def __init__(self, **kwargs):
super().__init__(**kwargs)
def generate(self, message, dataset=None):
return super(LMDeployAPI, self).generate(message, dataset=dataset)