VLMEvalKit / vlmeval /api /siliconflow.py
Racktic's picture
Upload folder using huggingface_hub
b5beb60 verified
import math
from vlmeval.smp import *
from vlmeval.api.base import BaseAPI
from vlmeval.dataset import img_root_map
API_BASE = "https://api.siliconflow.cn/v1/chat/completions"
def resize_image(image: Image.Image, max_height: int, max_width: int) -> Image.Image:
width, height = image.size
if min(width, height) < 50:
scale = 50 / min(width, height)
image = image.resize((int(width * scale), int(height * scale)))
current_pixels = width * height
if current_pixels <= max_height * max_width:
return image
scale = math.sqrt(max_height * max_width / current_pixels)
new_width = int(width * scale)
new_height = int(height * scale)
return image.resize((new_width, new_height), Image.Resampling.LANCZOS)
def encode_image(path: str, max_height: int = 1024, max_width: int = 1024) -> str:
image = Image.open(path).convert("RGB")
image = resize_image(image, max_height, max_width)
width, height = image.size
if min(height, width) < 50:
scale = 50 / min(width, height)
image = image.resize((int(width * scale), int(height * scale)))
buffered = io.BytesIO()
image.save(buffered, format="PNG")
img_bytes = buffered.getvalue()
img_base64 = base64.b64encode(img_bytes).decode("utf-8")
return img_base64
class SiliconFlowAPI(BaseAPI):
is_api: bool = True
def __init__(
self,
model: str = "deepseek-ai/DeepSeek-V2.5",
retry: int = 5,
wait: int = 5,
key: str = None,
api_base: str = API_BASE,
verbose: bool = True,
system_prompt: str = None,
timeout: int = 60,
reasoning: bool = False, # If set, will return results in the format of {'content': '...', 'reasoning': '...'}
**kwargs,
):
self.model = model
self.api_base = api_base
self.reasoning = reasoning
self.timeout = timeout
default_kwargs = {
"stream": False,
"temperature": 0,
"n": 1,
"max_tokens": 1280,
}
for k, v in default_kwargs.items():
if k not in kwargs:
kwargs[k] = default_kwargs[k]
if key is not None:
self.key = key
else:
self.key = os.environ.get("SiliconFlow_API_KEY", "")
headers = {"Authorization": "Bearer {}", "Content-Type": "application/json"}
headers["Authorization"] = headers["Authorization"].format(self.key)
self.headers = headers
super().__init__(
wait=wait,
retry=retry,
system_prompt=system_prompt,
verbose=verbose,
**kwargs,
)
@staticmethod
def build_msgs(msgs_raw):
messages = []
message = {"role": "user", "content": []}
image_b64 = None
for msg in msgs_raw:
if msg["type"] == "image" and not image_b64:
image_b64 = encode_image(msg["value"])
message["content"].append(
{"image_url": {"url": image_b64}, "type": "image_url"}
)
elif msg["type"] == "text":
message["content"].append({"text": msg["value"], "type": "text"})
messages.append(message)
return messages
def generate_inner(self, inputs, **kwargs) -> str:
default_kwargs = self.default_kwargs
default_kwargs.update(kwargs)
payload = dict(
model=self.model,
messages=self.build_msgs(msgs_raw=inputs),
**default_kwargs,
)
response = requests.post(
self.api_base, headers=self.headers, data=json.dumps(payload), timeout=self.timeout * 1.1
)
ret_code = response.status_code
ret_code = 0 if (200 <= int(ret_code) < 300) else ret_code
answer = self.fail_msg
try:
resp_struct = json.loads(response.text)
msg = resp_struct["choices"][0]["message"]
if self.reasoning and 'reasoning_content' in msg:
answer = {'content': msg['content'], 'reasoning': msg['reasoning_content']}
else:
answer = resp_struct["choices"][0]["message"]["content"].strip()
except:
pass
return ret_code, answer, response
class TeleMMAPI(SiliconFlowAPI):
is_api: bool = True
def __init__(
self,
model: str = "TeleAI/TeleMM",
key: str = None,
max_height: int = 1280,
max_width: int = 784,
**kwargs,
):
super().__init__(model=model, key=key, **kwargs)
self.max_height = max_height
self.max_width = max_width
def dump_image(self, line, dataset):
"""Dump the image(s) of the input line to the corresponding dataset folder.
Args:
line (line of pd.DataFrame): The raw input line.
dataset (str): The name of the dataset.
Returns:
str | list[str]: The paths of the dumped images.
"""
ROOT = LMUDataRoot()
assert isinstance(dataset, str)
# img_root = osp.join(ROOT, 'images', img_root_map[dataset] if dataset in img_root_map else dataset)
img_root = osp.join(ROOT, "images", img_root_map(dataset))
os.makedirs(img_root, exist_ok=True)
if "image" in line:
if isinstance(line["image"], list):
tgt_path = []
assert "image_path" in line
for img, im_name in zip(line["image"], line["image_path"]):
path = osp.join(img_root, im_name)
if not read_ok(path):
decode_base64_to_image_file(img, path)
tgt_path.append(path)
else:
tgt_path = osp.join(img_root, f"{line['index']}.jpg")
if not read_ok(tgt_path):
decode_base64_to_image_file(line["image"], tgt_path)
tgt_path = [tgt_path]
else:
assert "image_path" in line
tgt_path = toliststr(line["image_path"])
return tgt_path
def _prepare_content(
self, inputs: list[dict[str, str]], dataset: str = None
) -> list[dict[str, str]]:
"""
inputs list[dict[str, str]], each dict has keys: ['type', 'value']
"""
content = []
has_image = False
for s in inputs:
if s["type"] == "image":
if not has_image:
item = {
"type": "image_url",
"image_url": {
"url": encode_image(
s["value"],
max_height=self.max_height,
max_width=self.max_width,
)
},
}
has_image = True
else:
continue
elif s["type"] == "text":
prompt = s["value"]
if len(prompt) == 0:
continue
if dataset == "HallusionBench":
prompt += " Please answer yes or no directly, without any unnecessary explanation."
elif dataset == "OCRBench":
prompt = (
prompt + "\nExtract the text from the image intactly and "
+ "answer the question concisely and clearly if possible."
)
elif (
dataset == "AI2D_TEST"
or dataset == "MMStar"
or dataset == "MMBench_TEST_EN_V11"
or dataset == "MMVet"
):
prompt = prompt.replace(
"Please select the correct answer from the options above. \n",
"Please select the correct option from the above choices based on the "
+ "input image and question. The final output should only be one option, such as 'A'",
)
elif dataset == "MMBench_TEST_CN_V11":
prompt = prompt.replace(
"Please select the correct answer from the options above. \n",
"请根据输入图像和问题从上述选项中选择正确选项,最终的输出只有一个选项,例如'A'",
)
item = {"type": "text", "text": prompt}
else:
raise ValueError(f"Invalid message type: {s['type']}, {s}")
content.append(item)
return content
def generate_inner(self, inputs, **kwargs) -> str:
default_kwargs = self.default_kwargs
default_kwargs.update(kwargs)
messages = []
messages.append(
{
"role": "user",
"content": self._prepare_content(
inputs, dataset=kwargs.get("dataset", None)
),
}
)
payload = dict(model=self.model, messages=messages, **default_kwargs)
response = requests.post(
self.api_base, headers=self.headers, data=json.dumps(payload)
)
ret_code = response.status_code
ret_code = 0 if (200 <= int(ret_code) < 300) else ret_code
answer = self.fail_msg
try:
resp_struct = json.loads(response.text)
answer = resp_struct["choices"][0]["message"]["content"].strip()
return ret_code, answer, response
except Exception as err:
import traceback
traceback.print_exc()
if self.verbose:
self.logger.error(f"{type(err)}: {err}")
self.logger.error(f"The input messages are {inputs}.")
return -1, "", ""