| from vlmeval.smp import * |
| from vlmeval.api.base import BaseAPI |
|
|
| headers = 'Content-Type: application/json' |
|
|
|
|
| class GeminiWrapper(BaseAPI): |
|
|
| is_api: bool = True |
|
|
| def __init__(self, |
| model: str = 'gemini-1.0-pro', |
| retry: int = 5, |
| key: str = None, |
| verbose: bool = True, |
| temperature: float = 0.0, |
| system_prompt: str = None, |
| max_tokens: int = 2048, |
| proxy: str = None, |
| backend='genai', |
| project_id='vlmeval', |
| thinking_budget: int = None, |
| |
| fps: int = 1, |
| media_resolution: str = None, |
| **kwargs): |
|
|
| self.model = model |
| self.fail_msg = 'Failed to obtain answer via API. ' |
| self.max_tokens = max_tokens |
| self.temperature = temperature |
| self.thinking_budget = thinking_budget |
| self.fps = fps |
| |
| |
| self.media_resolution = media_resolution |
| if self.media_resolution: |
| assert self.media_resolution in ['low', 'medium', 'high'] |
| if key is None: |
| key = os.environ.get('GOOGLE_API_KEY', None) |
| |
| be = os.environ.get('GOOGLE_API_BACKEND', None) |
| if be is not None and be in ['genai', 'vertex']: |
| backend = be |
|
|
| assert backend in ['genai', 'vertex'] |
| if backend == 'genai': |
| |
| assert key is not None |
| try: |
| from google import genai |
| from google.genai import types |
| except ImportError as e: |
| raise ImportError( |
| "Could not import 'google.genai'. Please install it with:\n" |
| " pip install --upgrade google-genai" |
| ) from e |
| self.media_resolution_dict = { |
| 'low': types.MediaResolution.MEDIA_RESOLUTION_LOW, |
| 'medium': types.MediaResolution.MEDIA_RESOLUTION_MEDIUM, |
| 'high': types.MediaResolution.MEDIA_RESOLUTION_HIGH |
| } |
| self.genai = genai |
| self.client = genai.Client(api_key=key) |
|
|
| self.backend = backend |
| self.project_id = project_id |
| self.api_key = key |
|
|
| if proxy is not None: |
| proxy_set(proxy) |
| super().__init__(retry=retry, system_prompt=system_prompt, verbose=verbose, **kwargs) |
|
|
| def upload_video_genai(self, video_path): |
| from google import genai |
| from google.genai import types |
| myfile = self.client.files.upload(file=video_path) |
|
|
| video_part = types.Part.from_uri( |
| file_uri=myfile.uri, |
| mime_type="video/mp4" |
| ) |
|
|
| video_part.video_metadata = types.VideoMetadata(fps=self.fps) |
|
|
| while True: |
| myfile = self.client.files.get(name=myfile.name) |
| if myfile.state == "ACTIVE": |
| break |
| time.sleep(2) |
|
|
| return video_part |
|
|
| def build_msgs_genai(self, inputs): |
| video_in_msg = False |
| video_parts = [] |
| text_and_images = [] if self.system_prompt is None else [self.system_prompt] |
|
|
| for inp in inputs: |
| if inp['type'] == 'text': |
| text_and_images.append(inp['value']) |
| elif inp['type'] == 'image': |
| text_and_images.append(Image.open(inp['value'])) |
| elif inp['type'] == 'video': |
| video_file = self.upload_video_genai(inp['value']) |
| video_parts.append(video_file) |
| video_in_msg = True |
|
|
| messages = video_parts + text_and_images |
| return messages, video_in_msg |
|
|
| def build_msgs_vertex(self, inputs): |
| from vertexai.generative_models import Part, Image |
| messages = [] if self.system_prompt is None else [self.system_prompt] |
| for inp in inputs: |
| if inp['type'] == 'text': |
| messages.append(inp['value']) |
| elif inp['type'] == 'image': |
| messages.append(Part.from_image(Image.load_from_file(inp['value']))) |
| return messages |
|
|
| def generate_inner(self, inputs, **kwargs) -> str: |
| if self.backend == 'genai': |
| from google.genai import types |
| assert isinstance(inputs, list) |
| model = self.model |
| messages, video_in_msg = self.build_msgs_genai(inputs) |
|
|
| |
| config_args = { |
| "temperature": self.temperature, |
| "max_output_tokens": self.max_tokens |
| } |
| |
| if self.media_resolution: |
| if video_in_msg: |
| assert self.media_resolution != 'high', "For video input, only support medium and low resolution" |
| config_args["media_resolution"] = self.media_resolution_dict[self.media_resolution] |
|
|
| |
| |
| |
| if self.thinking_budget is not None: |
| config_args["thinking_config"] = types.ThinkingConfig( |
| thinking_budget=self.thinking_budget |
| ) |
| config_args.update(kwargs) |
|
|
| try: |
| resp = self.client.models.generate_content( |
| model=model, |
| contents=messages, |
| config=types.GenerateContentConfig(**config_args) |
| ) |
| answer = resp.text |
| return 0, answer, 'Succeeded! ' |
| except Exception as err: |
| if self.verbose: |
| self.logger.error(f'{type(err)}: {err}') |
| self.logger.error(f'The input messages are {inputs}.') |
|
|
| return -1, '', '' |
| elif self.backend == 'vertex': |
| import vertexai |
| from vertexai.generative_models import GenerativeModel |
| vertexai.init(project=self.project_id, location='us-central1') |
| model_name = 'gemini-1.0-pro-vision' if self.model == 'gemini-1.0-pro' else self.model |
| model = GenerativeModel(model_name=model_name) |
| messages = self.build_msgs_vertex(inputs) |
| try: |
| resp = model.generate_content(messages) |
| answer = resp.text |
| return 0, answer, 'Succeeded! ' |
| except Exception as err: |
| if self.verbose: |
| self.logger.error(f'{type(err)}: {err}') |
| self.logger.error(f'The input messages are {inputs}.') |
|
|
| return -1, '', '' |
|
|
|
|
| class Gemini(GeminiWrapper): |
| VIDEO_LLM = True |
|
|
| def generate(self, message, dataset=None): |
| return super(Gemini, self).generate(message) |
|
|