FruitBench / script /CogVLM2-0-shot.py
TJIET's picture
Upload 18 files
3baa3fd verified
import os
import torch
from PIL import Image
from modelscope import AutoModelForCausalLM, AutoTokenizer
MODEL_PATH = "./cogvlm2-llama3-chat-19B-int4"
DEVICE = 'cuda' if torch.cuda.is_available() else 'cpu'
TORCH_TYPE = torch.bfloat16 if torch.cuda.is_available() and torch.cuda.get_device_capability()[0] >= 8 else torch.float16
tokenizer = AutoTokenizer.from_pretrained(MODEL_PATH, trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained(
MODEL_PATH,
torch_dtype=TORCH_TYPE,
trust_remote_code=True,
low_cpu_mem_usage=True,
).eval()
question = (
''' "1. Identify the type of fruit or crop shown in the image. "
"2. Determine its current growth stage. (Options: unripe, mature, pest-damaged, rotten) "
"3. Recommend the farmer’s next action. (Options: keep for further growth, try to recover it, picking it, discard it) "
"4. Evaluate the consumer’s willingness to consume this fruit, from 1 (very unlikely) to 100 (very likely). "
"Please respond in the following format:\n"
"Type: [Fruit/Crop Name] Growth Stage: [unripe / mature / pest-damaged / rotten] "
"Recommendation: [keep for further growth / try to recover it / picking it / discard it] Consumer Score: [1-100]'''
)
root_folder = "../data"
output_root = "result"
os.makedirs(output_root, exist_ok=True)
for fruit in os.listdir(root_folder):
fruit_path = os.path.join(root_folder, fruit)
if not os.path.isdir(fruit_path):
continue
for subfolder in os.listdir(fruit_path):
subfolder_path = os.path.join(fruit_path, subfolder)
if not os.path.isdir(subfolder_path):
continue
image_files = [f for f in os.listdir(subfolder_path) if f.lower().endswith(('.jpg', '.jpeg', '.png', '.bmp', '.webp'))]
if not image_files:
continue
output_file = os.path.join(output_root, f"{fruit}_{subfolder}.txt")
with open(output_file, "w", encoding="utf-8") as out_file:
for filename in image_files:
image_path = os.path.join(subfolder_path, filename)
try:
print(f"🖼️ Processing {filename}...")
image = Image.open(image_path).convert('RGB')
history = []
input_by_model = model.build_conversation_input_ids(
tokenizer,
query=question,
history=history,
images=[image],
template_version='chat'
)
inputs = {
'input_ids': input_by_model['input_ids'].unsqueeze(0).to(DEVICE),
'token_type_ids': input_by_model['token_type_ids'].unsqueeze(0).to(DEVICE),
'attention_mask': input_by_model['attention_mask'].unsqueeze(0).to(DEVICE),
'images': [[input_by_model['images'][0].to(DEVICE).to(TORCH_TYPE)]],
}
gen_kwargs = {
"max_new_tokens": 2048,
"pad_token_id": 128002,
}
with torch.no_grad():
outputs = model.generate(**inputs, **gen_kwargs)
outputs = outputs[:, inputs['input_ids'].shape[1]:]
response = tokenizer.decode(outputs[0])
response = response.split("<|end_of_text|>")[0].strip()
print(f"✅ Response: {response}\n")
out_file.write(f"{'='*25} IMAGE START {'='*25}\n")
out_file.write(f"🖼️ Image Name: {filename}\n")
out_file.write(f"📝 Answer:\n{response}\n")
out_file.write(f"{'='*25} IMAGE END {'='*25}\n\n")
except Exception as e:
print(f"[ERROR] {fruit}/{subfolder}/{filename}: {e}")
out_file.write(f"{'='*25} IMAGE START {'='*25}\n")
out_file.write(f"🖼️ Image Name: {filename}\n")
out_file.write(f"❌ ERROR: {e}\n")
out_file.write(f"{'='*25} IMAGE END {'='*25}\n\n")