import json import sys import re import pandas as pd import numpy as np def parse_jsonl(file_name): with open(file_name, 'r', encoding='utf-8') as f: lines = f.readlines() return [json.loads(line) for line in lines] def unwrap_answer(data): if isinstance(data, dict) and "answer" in data: return data["answer"] return data def extract_json_content(text): if not text: return None text = text.strip() start = text.find('{') end = text.rfind('}') + 1 if start == -1 or end <= start: return None try: return json.loads(text[start:end]) except json.JSONDecodeError: return None def normalize_val(val): if val is None: return "" s = re.sub(r"[<>]", "", str(val)).strip() if not s: return "" num_candidate = s.replace(",", "") try: num = float(num_candidate) except ValueError: return s return str(int(num)) if num.is_integer() else str(num) def flatten_json(data, prefix=""): items = set() if isinstance(data, dict): for k, v in data.items(): items.update(flatten_json(v, f"{prefix}.{k}" if prefix else k)) elif isinstance(data, list): for v in data: if isinstance(v, list): row_tuple = tuple(normalize_val(sub_item) for sub_item in v) items.add((prefix, row_tuple)) else: items.update(flatten_json(v, prefix)) else: s = normalize_val(data) parts = re.split(r'[,,]', s) for part in parts: part = part.strip() if part: items.add((prefix, part)) return items def calculate_metrics(pred_json, gt_json, doc_id=None): def filter_present_categories(obj): return {"present_categories": obj.get("present_categories", [])} if isinstance(obj, dict) else {} if doc_id and doc_id.startswith("DTR_003"): pred_json = filter_present_categories(pred_json) gt_json = filter_present_categories(gt_json) if doc_id and (doc_id.startswith("NC_001") or doc_id.startswith("NC_002")): if pred_json is None: return None try: return 1.0 if abs(float(pred_json) - float(gt_json)) <= 2 else 0.0 except Exception: return 0.0 pred_set = flatten_json(pred_json) if pred_json is not None else set() gt_set = flatten_json(gt_json) if not gt_set: return 1.0 if not pred_set else 0.0 tp = len(pred_set & gt_set) fp = len(pred_set) - tp fn = len(gt_set) - tp precision = tp / (tp + fp) if (tp + fp) else 0.0 recall = tp / (tp + fn) if (tp + fn) else 0.0 f1 = (2 * precision * recall) / (precision + recall) if (precision + recall) else 0.0 return f1 def evaluate_samples(pred_file, gt_file): predictions = parse_jsonl(pred_file) gts = parse_jsonl(gt_file) gt_map = {gt['id']: gt for gt in gts} raw_scores = [] for pred_data in predictions: doc_id = pred_data['id'] # task NC_005 is currently under construction if doc_id.startswith('NC_005'): continue pred_json = extract_json_content(pred_data['response']) gt_json = json.loads(gt_map[doc_id]['response']) pred_json = unwrap_answer(pred_json) gt_json = unwrap_answer(gt_json) f1 = calculate_metrics(pred_json, gt_json, doc_id) raw_score = { "id": doc_id, "task": gt_map[doc_id]['task'], "subtask": gt_map[doc_id]['subtask'], "robustness": gt_map[doc_id]['robustness'], "f1": f1 } raw_scores.append(raw_score) return raw_scores def evaluate_tasks(raw_scores): df = pd.DataFrame(raw_scores) subtask_mean = ( df.groupby(["task", "subtask"], as_index=False) .agg(subtask_f1_mean=("f1", "mean")) ) task_mean = ( subtask_mean.groupby("task", as_index=False) .agg(task_f1=("subtask_f1_mean", "mean")) ) order = ["DTR", "KIE", "IQE", "CC", "VC", "NC", "RR"] task_mean_sorted = ( task_mean.assign(task=pd.Categorical(task_mean["task"], categories=order, ordered=True)) .sort_values("task") .reset_index(drop=True) ) print("Performance by Task:\n", task_mean_sorted) def evaluate_robustness(raw_scores): df = pd.DataFrame(raw_scores) subtask_mean = ( df.loc[df["task"].ne("IQE")] .groupby(["robustness", "task", "subtask"], as_index=False) .agg(subtask_f1_mean=("f1", "mean")) ) normal_sub_mean = ( subtask_mean[subtask_mean["robustness"].eq("Normal Captures")] ) normal_task_mean = ( normal_sub_mean.groupby("task", as_index=False) .agg(task_f1=("subtask_f1_mean", "mean")) ) normal_overall_mean = normal_task_mean["task_f1"].mean() sec_keys = ( subtask_mean.loc[subtask_mean["robustness"].eq("Secondary Captures"), ["task", "subtask"]] .drop_duplicates() ) sec_normal_sub_mean = ( subtask_mean[subtask_mean["robustness"].eq("Normal Captures")] .merge(sec_keys, on=["task", "subtask"], how="inner") ) sec_normal_task_mean = ( sec_normal_sub_mean.groupby("task", as_index=False) .agg(task_f1=("subtask_f1_mean", "mean")) ) sec_normal_overall_mean = sec_normal_task_mean["task_f1"].mean() multi_subtask_mean = ( df.loc[df["subtask"].isin([ "DTR_001_001", "DTR_003_001", "KIE_002_001", ])] .groupby(["robustness", "task", "subtask"], as_index=False) .agg(subtask_f1_mean=("f1", "mean")) ) multi_normal_sub_mean = ( multi_subtask_mean[multi_subtask_mean["robustness"].eq("Normal Captures")] ) multi_normal_task_mean = ( multi_normal_sub_mean.groupby("task", as_index=False) .agg(task_f1=("subtask_f1_mean", "mean")) ) multi_normal_overall_mean = multi_normal_task_mean["task_f1"].mean() task_mean = ( subtask_mean.groupby(["robustness", "task"], as_index=False) .agg(task_f1=("subtask_f1_mean", "mean")) ) robustness_scores = ( task_mean.groupby("robustness", as_index=False) .agg(robustness_macro_f1=("task_f1", "mean")) .sort_values("robustness_macro_f1", ascending=False) .reset_index(drop=True) ) use_special_denoms = { "Secondary Captures": sec_normal_overall_mean, "Cluttered Background": sec_normal_overall_mean, "Multi-doc Images": multi_normal_overall_mean} robustness_scores = robustness_scores.copy() robustness_scores["normal_denom"] = robustness_scores["robustness"].apply( lambda r: use_special_denoms.get(r, normal_overall_mean) ) robustness_scores["relative_to_normal"] = ( robustness_scores["robustness_macro_f1"] / robustness_scores["normal_denom"].replace(0, np.nan) ) robustness_scores = robustness_scores.drop(columns=["normal_denom", "robustness_macro_f1"]) robustness_scores = robustness_scores.sort_values('relative_to_normal', ascending=False) print("Performance by Robustness:\n", robustness_scores) if __name__ == "__main__": p_file = sys.argv[1] g_file = sys.argv[2] raw_scores = evaluate_samples(p_file, g_file) evaluate_tasks(raw_scores) evaluate_robustness(raw_scores)