import os import torch os.environ['CUDA_VISIBLE_DEVICES'] = '0,1,2,3' os.environ['SWIFT_DEBUG'] = '1' def _infer_model(pt_engine, system=None, messages=None, images=None, **kwargs): seed_everything(42) request_config = RequestConfig(max_tokens=128, temperature=0, repetition_penalty=1) if messages is None: messages = [] if system is not None: messages += [{'role': 'system', 'content': system}] messages += [{'role': 'user', 'content': '你好'}] resp = pt_engine.infer([{'messages': messages}], request_config=request_config) response = resp[0].choices[0].message.content messages += [{'role': 'assistant', 'content': response}, {'role': 'user', 'content': '这是什么'}] else: messages = messages.copy() if images is None: images = ['http://modelscope-open.oss-cn-hangzhou.aliyuncs.com/images/cat.png'] resp = pt_engine.infer([{'messages': messages, 'images': images, **kwargs}], request_config=request_config) response = resp[0].choices[0].message.content messages += [{'role': 'assistant', 'content': response}] logger.info(f'model: {pt_engine.model_info.model_name}, messages: {messages}') return response def test_qwen2_vl(): pt_engine = PtEngine('Qwen/Qwen2-VL-2B-Instruct') response = _infer_model(pt_engine) pt_engine.default_template.template_backend = 'jinja' response2 = _infer_model(pt_engine) assert response == response2 == '这是一只小猫的图片。它有黑白相间的毛发,眼睛大而圆,显得非常可爱。' def test_qwen2_5_vl(): pt_engine = PtEngine('Qwen/Qwen2.5-VL-7B-Instruct') messages = [{'role': 'user', 'content': 'What kind of dog is this?'}] images = ['https://qianwen-res.oss-accelerate-overseas.aliyuncs.com/Qwen2-VL/demo_small.jpg'] response = _infer_model(pt_engine, messages=messages, images=images) pt_engine.default_template.template_backend = 'jinja' response2 = _infer_model(pt_engine, messages=messages, images=images) assert response == response2 == ( 'The dog in the picture appears to be a Labrador Retriever. Labradors are known for their ' 'friendly and energetic nature, which is evident in the image where the dog seems to be ' "interacting playfully with the person. The dog's size, coat color, and build are " 'characteristic of the Labrador Retriever breed.') def test_qwen2_5_omni(): pt_engine = PtEngine('Qwen/Qwen2.5-Omni-7B') response = _infer_model(pt_engine) pt_engine.default_template.template_backend = 'jinja' response2 = _infer_model(pt_engine) assert response == response2 def test_qvq(): pt_engine = PtEngine('Qwen/QVQ-72B-Preview') response = _infer_model(pt_engine) pt_engine.default_template.template_backend = 'jinja' response2 = _infer_model(pt_engine) assert response == response2 def test_internvl2(): pt_engine = PtEngine('OpenGVLab/InternVL2-2B') response = _infer_model(pt_engine) pt_engine.default_template.template_backend = 'jinja' response2 = _infer_model(pt_engine) assert response == response2 def test_internvl2_phi3(): pt_engine = PtEngine('OpenGVLab/Mini-InternVL-Chat-4B-V1-5') _infer_model(pt_engine, system='') pt_engine.default_template.template_backend = 'jinja' _infer_model(pt_engine, system='') def test_internvl3_8b(): pt_engine = PtEngine('OpenGVLab/InternVL3-8B') response = _infer_model(pt_engine) pt_engine.default_template.template_backend = 'jinja' response2 = _infer_model(pt_engine, system='你是书生·万象,英文名是InternVL,是由上海人工智能实验室、清华大学及多家合作单位联合开发的多模态大语言模型。') assert response == response2 def test_internvl3_9b(): pt_engine = PtEngine('OpenGVLab/InternVL3-9B') response = _infer_model(pt_engine) pt_engine.default_template.template_backend = 'jinja' response2 = _infer_model(pt_engine, system='你是书生·万象,英文名是InternVL,是由上海人工智能实验室、清华大学及多家合作单位联合开发的多模态大语言模型。') assert response == response2 def test_llava(): pt_engine = PtEngine('AI-ModelScope/llava-v1.6-mistral-7b') _infer_model(pt_engine) def test_yi_vl(): pt_engine = PtEngine('01ai/Yi-VL-6B') _infer_model(pt_engine) def test_glm4v(): # There will be differences in '\n'. This is normal. pt_engine = PtEngine('ZhipuAI/glm-4v-9b') messages = [{'role': 'user', 'content': '描述这张图片'}] response = _infer_model(pt_engine, messages=messages) pt_engine.default_template.template_backend = 'jinja' response2 = _infer_model(pt_engine, messages=messages) assert response == ('这张图片是一只小猫的特写,它有着非常醒目的蓝色眼睛和混合了灰色、白色和棕色毛发的皮毛。小猫的耳朵竖立着,胡须清晰可见。它的眼神看起来既好奇又警觉,整体上显得非常可爱。') assert response2 == ('这是一张特写照片,展示了一只毛茸茸的小猫。小猫的眼睛大而圆,呈深蓝色,眼珠呈金黄色,非常明亮。它的鼻子短而小巧,' '是粉色的。小猫的嘴巴紧闭,胡须细长。它的耳朵竖立着,耳朵内侧是白色的,外侧是棕色的。小猫的毛发看起来柔软而浓密,' '主要是白色和棕色相间的花纹。背景模糊不清,但似乎是一个室内环境。') def test_cogagent(): pt_engine = PtEngine('ZhipuAI/cogagent-9b-20241220') messages = [{ 'role': 'user', 'content': """Task: I'm looking for a software to \"edit my photo with grounding\" History steps: (Platform: Mac) (Answer in Action-Operation-Sensitive format.)""" }] images = ['https://modelscope-open.oss-cn-hangzhou.aliyuncs.com/images/agent.png'] response = _infer_model(pt_engine, messages=messages, images=images) pt_engine.default_template.template_backend = 'jinja' response2 = _infer_model(pt_engine, messages=messages, images=images) assert response == response2 == ( """Action: Click on the 'Adobe Photoshop 2023' icon located in the middle of the screen to open the application. Grounded Operation: CLICK(box=[[346,574,424,710]], element_type='卡片', element_info='Adobe Photoshop 2023') <<一般操作>>""") def test_minicpmv(): pt_engine = PtEngine('OpenBMB/MiniCPM-V-2_6') _infer_model(pt_engine) pt_engine.default_template.template_backend = 'jinja' _infer_model(pt_engine) def test_minicpmo(): pt_engine = PtEngine('OpenBMB/MiniCPM-o-2_6') messages = [{ 'role': 'user', 'content': 'Compare image 1 and image 2, tell me about the differences between image 1 and image 2.' }] images = [ 'http://modelscope-open.oss-cn-hangzhou.aliyuncs.com/images/cat.png', 'http://modelscope-open.oss-cn-hangzhou.aliyuncs.com/images/animal.png' ] response = _infer_model(pt_engine, messages=messages, images=images) pt_engine.default_template.template_backend = 'jinja' response2 = _infer_model(pt_engine, messages=messages, images=images) assert response == response2 == ( 'The main difference between image 1 and image 2 is the subject matter. ' 'Image 1 features a close-up of a kitten, while image 2 depicts a cartoon illustration of four sheep ' 'standing in a grassy field. The setting, the number of subjects, and the overall style of the images ' 'are distinct from each other.') def test_got_ocr(): # https://github.com/modelscope/ms-swift/issues/2122 pt_engine = PtEngine('stepfun-ai/GOT-OCR2_0') _infer_model( pt_engine, messages=[{ 'role': 'user', 'content': 'OCR: ' }], images=['https://modelscope-open.oss-cn-hangzhou.aliyuncs.com/images/ocr.png']) def test_got_ocr_hf(): pt_engine = PtEngine('stepfun-ai/GOT-OCR-2.0-hf') response = _infer_model( pt_engine, messages=[{ 'role': 'user', 'content': 'OCR: ' }], images=['https://modelscope-open.oss-cn-hangzhou.aliyuncs.com/images/ocr.png']) assert response[:200] == ('简介 SWIFT支持250+LLM和35+MLLM(多模态大模型)的训练、推理、 评测和部署。开发者可以直接将' '我们的框架应用到自己的Research和 生产环境中,实现模型训练评测到应用的完整链路。我们除支持了 PEFT提供的轻量训练方案外' ',也提供了一个完整的Adapters库以支持 最新的训练技术,如NEFTune、LoRA+、LLaMA-PRO等,这个适配器 库可以脱离训练脚本' '直接使用在自己的') def test_llama_vision(): pt_engine = PtEngine('LLM-Research/Llama-3.2-11B-Vision-Instruct') response = _infer_model(pt_engine) pt_engine.default_template.template_backend = 'jinja' response2 = _infer_model(pt_engine) assert response == response2 def test_llava_hf(): pt_engine = PtEngine('llava-hf/llava-v1.6-mistral-7b-hf') response = _infer_model(pt_engine) pt_engine.default_template.template_backend = 'jinja' response2 = _infer_model(pt_engine) assert response == response2 def test_florence(): pt_engine = PtEngine('AI-ModelScope/Florence-2-base-ft') _infer_model(pt_engine, messages=[{'role': 'user', 'content': 'who are you?'}], images=[]) _infer_model( pt_engine, messages=[{ 'role': 'user', 'content': '' }], images=['http://modelscope-open.oss-cn-hangzhou.aliyuncs.com/images/animal.png']) def test_phi3_vision(): # pt_engine = PtEngine('LLM-Research/Phi-3-vision-128k-instruct') pt_engine = PtEngine('LLM-Research/Phi-3.5-vision-instruct') _infer_model(pt_engine) pt_engine.default_template.template_backend = 'jinja' _infer_model(pt_engine) def test_qwen_vl(): pt_engine = PtEngine('Qwen/Qwen-VL-Chat') _infer_model(pt_engine) def test_llava_onevision_hf(): pt_engine = PtEngine('llava-hf/llava-onevision-qwen2-0.5b-ov-hf') response = _infer_model(pt_engine) pt_engine.default_template.template_backend = 'jinja' response2 = _infer_model(pt_engine) assert response == response2 def test_xcomposer2_5(): pt_engine = PtEngine('Shanghai_AI_Laboratory/internlm-xcomposer2d5-ol-7b:base', torch.float16) # pt_engine = PtEngine('Shanghai_AI_Laboratory/internlm-xcomposer2d5-7b') response = _infer_model(pt_engine, system='') pt_engine.default_template.template_backend = 'jinja' response2 = _infer_model(pt_engine) assert response == response2 def test_deepseek_vl(): # pt_engine = PtEngine('deepseek-ai/deepseek-vl-1.3b-chat') pt_engine = PtEngine('deepseek-ai/Janus-1.3B') _infer_model(pt_engine) def test_deepseek_janus(): pt_engine = PtEngine('deepseek-ai/Janus-Pro-7B') messages = [{'role': 'user', 'content': '描述图片'}] response = _infer_model(pt_engine, messages=messages) assert response == ('这是一张非常可爱的猫咪图片。猫咪的毛色主要是白色,并带有灰色的条纹。它的眼睛非常大,呈现出明亮的蓝色,' '显得非常可爱和无辜。猫咪的耳朵竖立着,显得非常警觉和好奇。背景模糊,使得猫咪成为图片的焦点。' '整体画面给人一种温暖和愉悦的感觉。') def test_deepseek_vl2(): pt_engine = PtEngine('deepseek-ai/deepseek-vl2-small') response = _infer_model(pt_engine) assert response == ('这是一只可爱的小猫。它有着大大的蓝色眼睛和柔软的毛发,看起来非常天真无邪。小猫的耳朵竖立着,显得非常警觉和好奇。' '它的鼻子小巧而粉红,嘴巴微微张开,似乎在探索周围的环境。整体来看,这只小猫非常可爱,充满了活力和好奇心。') def test_mplug_owl2(): # pt_engine = PtEngine('iic/mPLUG-Owl2') pt_engine = PtEngine('iic/mPLUG-Owl2.1') _infer_model(pt_engine, messages=[{'role': 'user', 'content': '这是什么'}]) def test_mplug_owl3(): # pt_engine = PtEngine('iic/mPLUG-Owl3-7B-240728') pt_engine = PtEngine('iic/mPLUG-Owl3-7B-241101') response = _infer_model(pt_engine, system='') pt_engine.default_template.template_backend = 'jinja' response2 = _infer_model(pt_engine, system='') assert response == response2 def test_ovis1_6(): pt_engine = PtEngine('AIDC-AI/Ovis1.6-Gemma2-9B') # pt_engine = PtEngine('AIDC-AI/Ovis1.6-Gemma2-27B') response = _infer_model(pt_engine) pt_engine.default_template.template_backend = 'jinja' response2 = _infer_model(pt_engine) assert response == response2 def test_ovis1_6_llama3(): pt_engine = PtEngine('AIDC-AI/Ovis1.6-Llama3.2-3B') messages = [{'role': 'user', 'content': '这是什么'}] # llama3 response = _infer_model(pt_engine, messages=messages) pt_engine.default_template.template_backend = 'jinja' # llama3_2 _infer_model(pt_engine, messages=messages, system='You are a helpful and honest multimodal assistant.') assert response == '这是一只小猫。从图中可见的特征如大眼睛、细长的白色鼻毛和毛发的图案,表明它可能属于常见的猫种。猫的表情和毛发的质感显示出它年轻,可能是幼猫。' def test_ovis2(): pt_engine = PtEngine('AIDC-AI/Ovis2-2B') # with flash_attn response = _infer_model(pt_engine, messages=[{'role': 'user', 'content': 'Describe the image.'}]) assert response[:200] == ('The image features a close-up portrait of a young kitten with striking blue eyes. ' 'The kitten has a distinctive coat pattern with a mix of gray, black, and white fur, ' 'typical of a tabby pattern. Its ea') def test_paligemma(): pt_engine = PtEngine('AI-ModelScope/paligemma-3b-mix-224') response = _infer_model(pt_engine, messages=[{'role': 'user', 'content': 'detect cat'}]) assert response == ' cat' def test_paligemma2(): pt_engine = PtEngine('AI-ModelScope/paligemma2-3b-ft-docci-448', torch_dtype=torch.bfloat16) response = _infer_model(pt_engine, messages=[{'role': 'user', 'content': 'caption en'}]) assert response == ( 'A close up view of a white and gray kitten with black stripes on its head and face staring forward with ' 'its light blue eyes. The kitten is sitting on a white surface with a blurry background. ' "There is a light shining on the top of the kitten's head and the front of its body.") def test_pixtral(): pt_engine = PtEngine('AI-ModelScope/pixtral-12b') _infer_model(pt_engine, messages=[{'role': 'user', 'content': '这是什么'}]) def test_glm_edge_v(): pt_engine = PtEngine('ZhipuAI/glm-edge-v-2b') _infer_model(pt_engine, messages=[{'role': 'user', 'content': '这是什么'}]) def test_internvl2_5(): pt_engine = PtEngine('OpenGVLab/InternVL2_5-26B') _infer_model(pt_engine) pt_engine.default_template.template_backend = 'jinja' _infer_model(pt_engine, system='你是书生·万象,英文名是InternVL,是由上海人工智能实验室、清华大学及多家合作单位联合开发的多模态大语言模型。') def test_internvl2_5_mpo(): pt_engine = PtEngine('OpenGVLab/InternVL2_5-1B-MPO', model_type='internvl2_5') response = _infer_model(pt_engine, messages=[{'role': 'user', 'content': 'Hello, who are you?'}], images=[]) assert response == ("Hello! I'm an AI assistant whose name is InternVL, developed jointly by Shanghai AI Lab, " 'Tsinghua University and other partners.') response2 = _infer_model(pt_engine, messages=[{'role': 'user', 'content': '这是什么'}]) assert response2 == ('这是一只小猫的特写照片。照片中的小猫有大大的蓝色眼睛和毛发,看起来非常可爱。这种照片通常用于展示宠物的可爱瞬间。') def test_megrez_omni(): pt_engine = PtEngine('InfiniAI/Megrez-3B-Omni') _infer_model(pt_engine) response = _infer_model( pt_engine, messages=[{ 'role': 'user', 'content': [ { 'type': 'image' }, { 'type': 'audio', 'audio': 'weather.wav' }, ] }]) assert response == ('根据图片,无法确定确切的天气状况。然而,猫咪放松的表情和柔和的光线可能暗示着是一个晴朗或温和的日子。' '没有阴影或明亮的阳光表明这不是正午时分,也没有雨滴或雪花的迹象,这可能意味着不是下雨或下雪的日子。') def test_molmo(): # pt_engine = PtEngine('LLM-Research/Molmo-7B-O-0924') pt_engine = PtEngine('LLM-Research/Molmo-7B-D-0924') _infer_model(pt_engine) response = _infer_model(pt_engine, messages=[{'role': 'user', 'content': '这是什么'}]) assert response == ( ' This is a close-up photograph of a young kitten. ' 'The kitten has striking blue eyes and a mix of white and black fur, ' 'with distinctive black stripes on its head and face. ' "It's looking directly at the camera with an alert and curious expression. " "The kitten's fur appears soft and fluffy, and its pink nose and white whiskers are clearly visible. " 'The background is blurred, which emphasizes the kitten as the main subject of the image.') def test_molmoe(): pt_engine = PtEngine('LLM-Research/MolmoE-1B-0924') response = _infer_model(pt_engine, messages=[{'role': 'user', 'content': '这是什么'}]) assert response == (" This is a close-up photograph of a kitten's face. The kitten has striking blue eyes and " "a mix of white, black, and brown fur. It's looking directly at the camera with an adorable " "expression, its ears perked up and whiskers visible. The image captures the kitten's cute " 'features in sharp detail, while the background is blurred, creating a soft, out-of-focus ' "effect that emphasizes the young feline's charm.") def test_doc_owl2(): pt_engine = PtEngine('iic/DocOwl2', torch_dtype=torch.float16) response = _infer_model(pt_engine, messages=[{'role': 'user', 'content': '你是谁'}], images=[]) images = [ 'https://modelscope.cn/models/iic/DocOwl2/resolve/master/examples/docowl2_page0.png', 'https://modelscope.cn/models/iic/DocOwl2/resolve/master/examples/docowl2_page1.png', 'https://modelscope.cn/models/iic/DocOwl2/resolve/master/examples/docowl2_page2.png', 'https://modelscope.cn/models/iic/DocOwl2/resolve/master/examples/docowl2_page3.png', 'https://modelscope.cn/models/iic/DocOwl2/resolve/master/examples/docowl2_page4.png', 'https://modelscope.cn/models/iic/DocOwl2/resolve/master/examples/docowl2_page5.png', ] response = _infer_model( pt_engine, messages=[{ 'role': 'user', 'content': '' * len(images) + 'what is this paper about? provide detailed information.' }], images=images) assert response == ( 'This paper is about multimodal Language Models(MLMs) achieving promising OCR-free ' 'Document Understanding by performing understanding by the cost of generating thorough sands of visual ' 'tokens for a single document image, leading to excessive GPU computation time. The paper also discusses ' 'the challenges and limitations of existing multimodal OCR approaches and proposes a new framework for ' 'more efficient and accurate OCR-free document understanding.') def test_valley(): pt_engine = PtEngine('bytedance-research/Valley-Eagle-7B') _infer_model(pt_engine) def test_ui_tars(): os.environ['MAX_PIXELS'] = str(1280 * 28 * 28) pt_engine = PtEngine('bytedance-research/UI-TARS-2B-SFT') prompt = ('You are a GUI agent. You are given a task and your action history, with screenshots. ' 'You need to perform the next action to complete the task.' + r""" ## Output Format ```\nThought: ... Action: ...\n``` ## Action Space click(start_box='<|box_start|>(x1,y1)<|box_end|>') left_double(start_box='<|box_start|>(x1,y1)<|box_end|>') right_single(start_box='<|box_start|>(x1,y1)<|box_end|>') drag(start_box='<|box_start|>(x1,y1)<|box_end|>', end_box='<|box_start|>(x3,y3)<|box_end|>') hotkey(key='') type(content='') #If you want to submit your input, use \"\ \" at the end of `content`. scroll(start_box='<|box_start|>(x1,y1)<|box_end|>', direction='down or up or right or left') wait() #Sleep for 5s and take a screenshot to check for any changes. finished() call_user() # Submit the task and call the user when the task is unsolvable, or when you need the user's help. ## Note - Use Chinese in `Thought` part. - Summarize your next action (with its target element) in one sentence in `Thought` part. ## User Instruction """) instruction = "I'm looking for a software to \"edit my photo with grounding\"" messages = [ { 'role': 'user', 'content': [ { 'type': 'text', 'text': prompt + instruction }, ], }, ] images = ['https://modelscope-open.oss-cn-hangzhou.aliyuncs.com/images/agent.png'] response = _infer_model(pt_engine, messages=messages, images=images) pt_engine.default_template.template_backend = 'jinja' response2 = _infer_model(pt_engine, messages=messages, images=images) assert response == response2 def test_phi4_vision(): pt_engine = PtEngine('LLM-Research/Phi-4-multimodal-instruct') response = _infer_model(pt_engine, messages=[{'role': 'user', 'content': 'describe the image.'}]) assert response == ( "The image features a close-up of a kitten's face. The kitten has large, " 'round eyes with a bright gaze, and its fur is predominantly white with black stripes. ' "The kitten's ears are pointed and alert, and its whiskers are visible. The background is blurred, " "drawing focus to the kitten's face.") response = _infer_model( pt_engine, messages=[{ 'role': 'user', 'content': 'describe the audio.' }], images=[], audios=['http://modelscope-open.oss-cn-hangzhou.aliyuncs.com/images/weather.wav']) assert response == '今天天气真好呀' def test_gemma3_vision(): pt_engine = PtEngine('LLM-Research/gemma-3-4b-it') response = _infer_model(pt_engine, messages=[{'role': 'user', 'content': 'Describe this image in detail.'}]) pt_engine.default_template.template_backend = 'jinja' response2 = _infer_model(pt_engine, messages=[{'role': 'user', 'content': 'Describe this image in detail.'}]) assert response[:80] == response2[:80] == ( "Here's a detailed description of the image:\n\n**Overall Impression:**\n\nThe image ") def test_mistral_2503(): pt_engine = PtEngine('mistralai/Mistral-Small-3.1-24B-Instruct-2503') response = _infer_model(pt_engine, messages=[{'role': 'user', 'content': 'What is shown in this image?'}]) assert response == ( 'The image shows a close-up of a Siamese kitten. The kitten has distinctive blue almond-shaped eyes, ' 'a pink nose, and a light-colored coat with darker points on the ears, paws, tail, and face, ' 'which are characteristic features of the Siamese breed. ' 'The kitten appears to be looking directly at the viewer with a curious and endearing expression.') def test_llama4(): pt_engine = PtEngine('LLM-Research/Llama-4-Scout-17B-16E-Instruct') messages = [{'role': 'user', 'content': 'What is the difference between the two images?'}] images = [ 'http://modelscope-open.oss-cn-hangzhou.aliyuncs.com/images/cat.png', 'http://modelscope-open.oss-cn-hangzhou.aliyuncs.com/images/animal.png' ] response = _infer_model(pt_engine, messages=messages, images=images) assert response[:128] == ('The two images are distinct in their subject matter and style. The first image features ' 'a realistic depiction of a kitten, while') and len(response) == 654 def test_kimi_vl(): pt_engine = PtEngine('moonshotai/Kimi-VL-A3B-Instruct') messages = [{'role': 'user', 'content': 'What is the difference between the two images?'}] images = [ 'http://modelscope-open.oss-cn-hangzhou.aliyuncs.com/images/cat.png', 'http://modelscope-open.oss-cn-hangzhou.aliyuncs.com/images/animal.png' ] response = _infer_model(pt_engine, messages=messages, images=images) assert response == ('The first image is a close-up of a kitten with a blurred background, ' 'while the second image is a cartoon of four sheep standing in a field.') if __name__ == '__main__': from swift.llm import PtEngine, RequestConfig from swift.utils import get_logger, seed_everything logger = get_logger() # test_qwen2_vl() # test_qwen2_5_vl() # test_qwen2_5_omni() # test_internvl2() # test_internvl2_phi3() # test_llava() # test_ovis1_6() # test_ovis1_6_llama3() # test_ovis2() # test_yi_vl() # test_deepseek_vl() # test_deepseek_janus() # test_deepseek_vl2() # test_qwen_vl() # test_glm4v() # test_cogagent() # test_llava_onevision_hf() # test_minicpmv() # test_got_ocr() # test_got_ocr_hf() # test_paligemma() # test_paligemma2() # test_pixtral() # test_llama_vision() # test_llava_hf() # test_florence() # test_glm_edge_v() # test_phi3_vision() # test_phi4_vision() # test_internvl2_5() # test_internvl2_5_mpo() # test_mplug_owl3() # test_xcomposer2_5() # test_megrez_omni() # test_qvq() # test_mplug_owl2() # test_molmo() # test_molmoe() # test_doc_owl2() # test_minicpmo() # test_valley() # test_ui_tars() test_gemma3_vision() # test_mistral_2503() # test_llama4() # test_internvl3_8b() # test_internvl3_9b() # test_kimi_vl()