|
|
import os |
|
|
import json |
|
|
import time |
|
|
import sys |
|
|
from abc import abstractmethod |
|
|
from tabulate import tabulate |
|
|
|
|
|
|
|
|
def pick_response_text(json_path): |
|
|
""" |
|
|
""" |
|
|
try: |
|
|
with open(json_path, "r") as f: |
|
|
json_data = json.load(f) |
|
|
except Exception as e: |
|
|
print("--> file error: msg: {}, path: {}".format(e, json_path)) |
|
|
return None |
|
|
|
|
|
for required_key in ["model_name", "response"]: |
|
|
if required_key not in json_data: |
|
|
print("--> required key not exists, name: {}, path: {}".format(required_key, json_path)) |
|
|
return None |
|
|
|
|
|
model_name = json_data["model_name"] |
|
|
model_response = json_data["response"] |
|
|
|
|
|
response_text = None |
|
|
if model_name.startswith("gpt") or model_name.startswith("o1"): |
|
|
response_text = model_response.get("data", {}).get("response", {}).get("choices", [{}])[0].get("message", {}).get("content", None) |
|
|
elif model_name.startswith("local_"): |
|
|
response_text = model_response |
|
|
else: |
|
|
if model_name.startswith("claude"): |
|
|
content_list = model_response.get("content", None) |
|
|
elif model_name.startswith("gemini"): |
|
|
content_list = model_response.get("candidates", [{}])[0].get("content", {}).get("parts", None) |
|
|
elif model_name.startswith("qwen"): |
|
|
content_list = model_response.get("output", {}).get("choices", [{}])[0].get("message", {}).get("content", None) |
|
|
else: |
|
|
raise NotImplementedError("The pick_response_text NOT implemented for model: {}".format(model_name)) |
|
|
|
|
|
if isinstance(content_list, list) and len(content_list) > 0: |
|
|
response_text = content_list[0].get("text", None) |
|
|
|
|
|
if response_text is None: |
|
|
print("--> [error][{}] text pick error, path: {}".format(model_name, json_path)) |
|
|
return response_text |
|
|
|
|
|
|
|
|
def load_response_from_dir(res_dir): |
|
|
""" |
|
|
""" |
|
|
response_info = {} |
|
|
for file_name in os.listdir(res_dir): |
|
|
file_path = os.path.abspath(os.path.join(res_dir, file_name)) |
|
|
if not file_name.endswith(".json"): |
|
|
print("--> skip: result file should be a json: but got: {}".format(file_path)) |
|
|
continue |
|
|
|
|
|
response_text = pick_response_text(file_path) |
|
|
if response_text is None: |
|
|
continue |
|
|
|
|
|
file_name_wo_ext, ext = os.path.splitext(file_name) |
|
|
response_info[file_name_wo_ext] = response_text |
|
|
return response_info |
|
|
|
|
|
|
|
|
class BaseMetric(object): |
|
|
""" BaseMetric """ |
|
|
""" OCRMetric """ |
|
|
def __init__(self, group_name, **kwargs): |
|
|
self.group_name = group_name |
|
|
self.kwargs = kwargs |
|
|
|
|
|
def response_post_func(self, response_text, **kwargs): |
|
|
return response_text |
|
|
|
|
|
@abstractmethod |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
def evaluate(self, response_info, gt_info, normalize_func=None, **kwargs): |
|
|
pass |
|
|
|
|
|
def __call__(self, pdt_res_dir, gt_info, with_response_ratio=True, **kwargs): |
|
|
if isinstance(pdt_res_dir, dict): |
|
|
raw_response_info = pdt_res_dir |
|
|
elif os.path.exists(pdt_res_dir) and os.path.isdir(pdt_res_dir): |
|
|
raw_response_info = load_response_from_dir(pdt_res_dir) |
|
|
else: |
|
|
return ValueError("invalid input: response dict or folder are required, but got {}".format(pdt_res_dir)) |
|
|
|
|
|
post_error_list, response_info = [], {} |
|
|
response_error_list = list(gt_info.keys() - raw_response_info.keys()) |
|
|
for file_name, single_pdt_str in raw_response_info.items(): |
|
|
single_pdt_str = self.response_post_func(single_pdt_str, **kwargs) |
|
|
if single_pdt_str is None: |
|
|
post_error_list.append(file_name) |
|
|
continue |
|
|
response_info[file_name] = single_pdt_str |
|
|
|
|
|
meta_info = { |
|
|
"gt_total_num": len(gt_info), "pdt_total_num": len(response_info), |
|
|
"post_error_list": post_error_list, "response_error_list": response_error_list, |
|
|
} |
|
|
eval_info = self.evaluate(response_info, gt_info, **kwargs) |
|
|
|
|
|
|
|
|
if "summary" in eval_info and with_response_ratio: |
|
|
success_ratio = (len(response_info) + len(post_error_list)) / (len(gt_info) + 1e-9) |
|
|
eval_info["summary"].update({"response_success_ratio": success_ratio}) |
|
|
return meta_info, eval_info |
|
|
|
|
|
|
|
|
def summary(index_path, exp_dir_base, is_weighted_sum=False): |
|
|
""" |
|
|
""" |
|
|
with open(index_path, "r") as f: |
|
|
data_list = json.load(f) |
|
|
|
|
|
all_data_info = {} |
|
|
for data_info_item in data_list: |
|
|
data_name = data_info_item["dataset"] |
|
|
if not data_info_item.get("release", True): |
|
|
continue |
|
|
all_data_info[data_name] = data_info_item |
|
|
dataset_list = list(all_data_info.keys()) |
|
|
summary_path = summary_multi_exp(exp_dir_base, dataset_list, is_weighted_sum=is_weighted_sum) |
|
|
return summary_path |
|
|
|
|
|
|
|
|
def summary_multi_exp(exp_dir_base, dataset_list=None, is_weighted_sum=False): |
|
|
""" |
|
|
""" |
|
|
if dataset_list is None: |
|
|
all_dataset_name = [] |
|
|
for exp_name in os.listdir(exp_dir_base): |
|
|
dir_status_path = os.path.join(exp_dir_base, exp_name, "status.json") |
|
|
if not os.path.exists(dir_status_path): |
|
|
continue |
|
|
with open(dir_status_path, "r") as f: |
|
|
data_status_info = json.load(f) |
|
|
all_dataset_name.extend(data_status_info.keys()) |
|
|
dataset_list = sorted(set(all_dataset_name)) |
|
|
|
|
|
|
|
|
all_evaluate_info, _ = {}, 0 |
|
|
for exp_name in os.listdir(exp_dir_base): |
|
|
dir_status_path = os.path.join(exp_dir_base, exp_name, "status.json") |
|
|
if not os.path.exists(dir_status_path): |
|
|
print("--> skip: status.json not exist: {}".format(dir_status_path)) |
|
|
continue |
|
|
|
|
|
with open(dir_status_path, "r") as f: |
|
|
all_status_info = json.load(f) |
|
|
|
|
|
for data_name in dataset_list: |
|
|
total_num = all_status_info.get(data_name, {}).get("config", {}).get("num", "-1") |
|
|
summary_info = all_status_info.get(data_name, {}).get("evaluation", {}).get("summary", {}) |
|
|
for metric_name, metric_value in summary_info.items(): |
|
|
if metric_name not in all_evaluate_info: |
|
|
all_evaluate_info[metric_name] = {} |
|
|
if exp_name not in all_evaluate_info[metric_name]: |
|
|
all_evaluate_info[metric_name][exp_name] = {} |
|
|
all_evaluate_info[metric_name][exp_name][data_name] = (metric_value, total_num) |
|
|
|
|
|
all_table_md = [] |
|
|
for metric_name, metric_info in all_evaluate_info.items(): |
|
|
formatted_time = time.strftime("%Y-%m-%d %H:%M", time.localtime(time.time())) |
|
|
summary_line_list = [] |
|
|
summary_key_name = "summary(weighted)" if is_weighted_sum else "summary" |
|
|
summary_head = [f"exp_name({metric_name}_{formatted_time})"] + dataset_list + [summary_key_name] |
|
|
for exp_name, data_eval_info in metric_info.items(): |
|
|
summary_line = [exp_name, ] |
|
|
|
|
|
all_metric_value = 0 |
|
|
is_summary_valid, all_total_num, all_weighted_metric = True, 0, 0 |
|
|
for data_name in dataset_list: |
|
|
metric_value, total_num = data_eval_info.get(data_name, ("-1", "-1")) |
|
|
summary_line.append("{:.2f}".format(float(metric_value) * 100)) |
|
|
if str(metric_value) == "-1" or str(metric_value) == "-1": |
|
|
is_summary_valid = False |
|
|
continue |
|
|
|
|
|
all_total_num += float(total_num) |
|
|
all_weighted_metric += float(total_num) * float(metric_value) |
|
|
all_metric_value += float(metric_value) |
|
|
|
|
|
summary_value_valid = ((all_weighted_metric / (all_total_num + 1e-9)) * 100) if is_weighted_sum \ |
|
|
else (all_metric_value / (len(dataset_list) + 1e-9) * 100) |
|
|
summary_value = "-" if not is_summary_valid else "{:.2f}".format(summary_value_valid) |
|
|
summary_line.append(summary_value) |
|
|
summary_line_list.append(summary_line) |
|
|
|
|
|
md_table_info = tabulate(summary_line_list, headers=summary_head, tablefmt='pipe') |
|
|
all_table_md.append(md_table_info) |
|
|
|
|
|
print("\n\n".join(all_table_md)) |
|
|
summary_path = os.path.abspath(os.path.join(exp_dir_base, "summary.md")) |
|
|
with open(summary_path, "w") as f: |
|
|
f.write("\n\n".join(all_table_md)) |
|
|
return summary_path |
|
|
|
|
|
|
|
|
if __name__ == '__main__': |
|
|
if len(sys.argv) != 2: |
|
|
print("Usage: python {} exp_base_dir".format(__file__)) |
|
|
exit(-1) |
|
|
else: |
|
|
print('--> info: {}'.format(sys.argv)) |
|
|
exp_base_dir = sys.argv[1] |
|
|
|
|
|
summary_path = summary_multi_exp(exp_base_dir, dataset_list=None, is_weighted_sum=False) |
|
|
print("--> info: summary saved at : {}".format(summary_path)) |
|
|
print("happy coding.") |
|
|
|