| import json |
| import subprocess |
|
|
| from llava.train.llava_trainer import LLaVATrainer |
|
|
|
|
| class LLaVAEvalTrainer(LLaVATrainer): |
| def evaluate(self, evaluate_args): |
| cmd = f"accelerate launch --num_processes {evaluate_args.eval_num_processes} -m lmms_eval \ |
| --model {evaluate_args.model} \ |
| --model_args {evaluate_args.model_args} \ |
| --tasks {evaluate_args.task_names} \ |
| --batch_size {evaluate_args.batch_size} \ |
| --log_samples_suffix {evaluate_args.log_samples_suffix} \ |
| --output_path {evaluate_args.output_path}" |
| if evaluate_args.limit: |
| cmd += f" --limit {evaluate_args.limit}" |
| if evaluate_args.num_fewshot: |
| cmd += f" --num_fewshot {evaluate_args.num_fewshot}" |
| if evaluate_args.gen_kwargs != "": |
| cmd += f" --gen_kwargs {evaluate_args.gen_kwargs}" |
| if evaluate_args.log_samples: |
| cmd += f" --log_samples" |
| else: |
| assert False, "Please log samples so that the result can be parsed" |
| results = subprocess.run([cmd], shell=True, capture_output=True, text=True) |
| try: |
| result_file_index_start = results.stdout.index("Saved samples to ") |
| result_file_index_end = results.stdout.index(f".json") |
| result_file_index_start += len("Saved samples to ") |
| file = results.stdout[result_file_index_start:result_file_index_end] |
| except: |
| result_file_index_start = results.stderr.index("Saved samples to ") |
| result_file_index_end = results.stderr.index(f".json") |
| result_file_index_start += len("Saved samples to ") |
| file = results.stderr[result_file_index_start:result_file_index_end] |
| file = file.split("/")[:-1] |
| file = "/".join(file) + "/results.json" |
| with open(file, "r") as f: |
| lmms_eval_results = json.load(f) |
| result_dict = {} |
| tasks_list = evaluate_args.task_names.split(",") |
| for task in tasks_list: |
| task_results = lmms_eval_results["results"][task] |
| for k, v in task_results.items(): |
| if k != "alias" and "stderr" not in k: |
| metric = k.split(",")[0] |
| result_dict[f"{task}_{metric}"] = v |
| return result_dict |
|
|
| """def evaluate(self, evaluate_args): |
| initialize_tasks() |
| tasks_list = evaluate_args.task_names.split(",") |
| result_dict = {} |
| results = evaluator.simple_evaluate( |
| model=evaluate_args.model, |
| model_args=evaluate_args.model_args, |
| tasks=tasks_list, |
| num_fewshot=evaluate_args.num_fewshot, |
| batch_size=evaluate_args.batch_size, |
| device=evaluate_args.device, |
| limit=evaluate_args.limit, |
| check_integrity=evaluate_args.check_integrity, |
| show_task_to_terminal=evaluate_args.show_task_to_terminal, |
| log_samples=evaluate_args.log_samples, |
| gen_kwargs=evaluate_args.gen_kwargs, |
| cli_args=evaluate_args, |
| ) |
| for task in tasks_list: |
| task_results = results["results"][task] |
| for k,v in task_results.items(): |
| if k != "alias" and "stderr" not in k: |
| metric = k.split(",")[0] |
| result_dict[f"{task}_{metric}"] = v |
| |
| return result_dict""" |
|
|