llava-spat / rgbd_eval.py
chouss's picture
Uploading folder contents
754ac61 verified
from llava.mm_utils import get_model_name_from_path
from llava.eval.run_llava_backup import eval_model
# from llava.eval.run_llava import eval_model
model_path = "/home/aiops/wangzh/llava-spat/checkpoints/llava-v1.5-7b-final-neg-lora-new"
# prompt = "Describe the orientation and position relationship between two giraffes in the picture."
# image_file = "/home/aiops/wangzh/data/RGBD-benchmark/out_doors/pic_all/000000023744.jpg"
# prompt = "Describe the orientation position relationship between two bus in the picture."
# image_file = "/home/aiops/wangzh/data/RGBD-benchmark/out_doors/pic_all/000000017286.jpg"
# prompt = "Describe the spatial relationship between the three stop signs in the picture."
# image_file = "/home/aiops/wangzh/data/RGBD-benchmark/out_doors/pic_all/000000010700.jpg"
# prompt = "Describe the position of the objects in the picture."
# image_file = "/home/aiops/wangzh/data/RGBD-benchmark/out_doors/pic_all/000000016983.jpg"
base_path = "/home/aiops/wangzh/data/RGBD-benchmark"
args = type('Args', (), {
"model_path": model_path,
"model_base": '/home/aiops/wangzh/llava/vicuna-7b-v1.5',
# "model_base": 'lmsys/vicuna-13b-v1.5',
"model_name": get_model_name_from_path(model_path),
# "query": prompt,
"conv_mode": None,
# "image_file": image_file,
"base_path": base_path,
"sep": ",",
"temperature": 0.3,
"top_p": 0.7,
"num_beams": 3,
"max_new_tokens": 512
})()
eval_model(args)