Racktic's picture
Upload folder using huggingface_hub
b5beb60 verified
from vlmeval.smp import *
from vlmeval.api.base import BaseAPI
from vlmeval.dataset import DATASET_TYPE, img_root_map
class TaiyiWrapper(BaseAPI):
is_api: bool = True
def __init__(self,
model: str = 'taiyi',
retry: int = 5,
wait: int = 5,
key: str = None,
verbose: bool = False,
system_prompt: str = None,
temperature: float = 0,
timeout: int = 60,
url: str = "https://taiyi.megvii.com/v1/chat/completions",
max_tokens: int = 1024,
**kwargs):
self.model = model
self.fail_msg = 'Failed to obtain answer via API. '
self.max_tokens = max_tokens
self.temperature = temperature
if key is None:
key = os.environ.get('TAIYI_API_KEY', None)
assert key is not None, ('Please set the API Key ')
self.key = key
self.timeout = timeout
super().__init__(wait=wait, retry=retry, system_prompt=system_prompt, verbose=verbose, **kwargs)
assert url is not None, ('Please set the url ')
self.url = url
self.logger.info(f'Using url: {self.url}; API Key: {self.key}')
def use_custom_prompt(self, dataset):
if DATASET_TYPE(dataset) == 'Y/N' or DATASET_TYPE(dataset) == 'MCQ' or DATASET_TYPE(dataset) == 'VQA':
return True
return False
def prepare_inputs(self, inputs):
input_msgs = []
if self.system_prompt is not None:
input_msgs.append(dict(role='system', content=self.system_prompt))
has_images = np.sum([x['type'] == 'image' for x in inputs])
if has_images:
content_list = []
for msg in inputs:
if msg['type'] == 'text':
content_list.append(dict(type='text', text=msg['value']))
elif msg['type'] == 'image':
imgbytes = open(msg['value'],'rb').read()
b64 = base64.b64encode(imgbytes).decode('ascii')
img_struct = dict(url=f'data:image/jpeg;base64,{b64}')
content_list.append(dict(type='image_url', image_url=img_struct))
input_msgs.append(dict(role='user', content=content_list))
else:
assert all([x['type'] == 'text' for x in inputs])
text = '\n'.join([x['value'] for x in inputs])
input_msgs.append(dict(role='user', content=text))
return input_msgs
def set_dump_image(self, dump_image_func):
self.dump_image_func = dump_image_func
def dump_image(self, line, dataset):
return self.dump_image_func(line)
def image_first(self, msgs):
nr_img = 0
for s in msgs:
if s['type'] == 'image':
nr_img += 1
if nr_img == 1:
new_msgs = []
img_msg = None
for s in msgs:
if s['type'] == 'text':
new_msgs.append(s)
else:
img_msg = s
new_msgs.insert(0, img_msg)
else:
new_msgs = msgs
return new_msgs
def build_multi_choice_prompt(self, line, dataset=None):
question = line['question']
hint = line['hint'] if ('hint' in line and not pd.isna(line['hint'])) else None
if hint is not None:
question = hint + '\n' + question
options = {
cand: line[cand]
for cand in string.ascii_uppercase
if cand in line and not pd.isna(line[cand])
}
for key, item in options.items():
question += f'\n{key}. {item}'
prompt = question
if len(options):
prompt += '\n请直接回答选项字母。' if cn_string(
prompt) else "\nAnswer with the option's letter from the given choices directly."
else:
prompt += '\n请直接回答问题。' if cn_string(prompt) else '\nAnswer the question directly.'
return prompt
def build_yorn_prompt(self, line, dataset=None):
if listinstr(['HallusionBench'], dataset):
pre_prompt = 'Read the following question carefully, think and solve it step by step.\n\n'
else:
pre_prompt = ''
prompt = pre_prompt + line['question'] + ' Please answer yes or no as the final answer.'
return prompt
def build_vqa_prompt(self, line, dataset=None):
if listinstr(['OCRBench'], dataset):
pre_prompt = 'Carefully identify the text in the image and answer the question.\n\n'
else:
pre_prompt = ''
if listinstr(['MMVet'], dataset):
post_prompt = '\nAnswer this question in detail.'
else:
post_prompt = ''
prompt = pre_prompt + line['question'] + post_prompt
return prompt
def build_prompt(self, line, dataset=None):
assert self.use_custom_prompt(dataset)
assert dataset is None or isinstance(dataset, str)
tgt_path = self.dump_image(line, dataset)
if DATASET_TYPE(dataset) == 'MCQ':
prompt = self.build_multi_choice_prompt(line, dataset)
elif DATASET_TYPE(dataset) == 'Y/N':
prompt = self.build_yorn_prompt(line, dataset)
elif DATASET_TYPE(dataset) == 'VQA':
prompt = self.build_vqa_prompt(line, dataset)
else:
raise RuntimeError(f'Invalid dataset type: {DATASET_TYPE(dataset)}')
message = []
message.extend([dict(type='image', value=s) for s in tgt_path])
message.extend([dict(type='text', value=prompt)])
# interleave dataset
if dataset.startswith('MMMU_'):
from .. import MMMUDataset
message = MMMUDataset.split_MMMU(message)
message = self.image_first(message)
return message
def generate_inner(self, inputs, **kwargs) -> str:
input_msgs = self.prepare_inputs(inputs)
temperature = kwargs.pop('temperature', self.temperature)
headers = {'Authorization': f'Bearer {self.key}'}
payload = dict(
model=self.model,
messages=input_msgs,
n=1,
temperature=temperature,
**kwargs)
response = requests.post(self.url, headers=headers, data=json.dumps(payload), timeout=self.timeout * 1.1)
ret_code = response.status_code
ret_code = 0 if (200 <= int(ret_code) < 300) else ret_code
answer = self.fail_msg
try:
resp_struct = json.loads(response.text)
answer = resp_struct['choices'][0]['message']['content'].strip()
except:
pass
return ret_code, answer, response
class TaiyiAPI(TaiyiWrapper):
def generate(self, message, dataset=None):
return super(TaiyiAPI, self).generate(message)