cmdyoooppp commited on
Commit
dbc6c3e
·
verified ·
1 Parent(s): 5d215fe

Upload evaluation.py

Browse files
Files changed (1) hide show
  1. vision_language/evaluation.py +230 -0
vision_language/evaluation.py ADDED
@@ -0,0 +1,230 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import json
2
+ import sys
3
+ import re
4
+ import pandas as pd
5
+ import numpy as np
6
+
7
+
8
+ def parse_jsonl(file_name):
9
+ with open(file_name, 'r', encoding='utf-8') as f:
10
+ lines = f.readlines()
11
+ return [json.loads(line) for line in lines]
12
+
13
+
14
+ def unwrap_answer(data):
15
+ if isinstance(data, dict) and "answer" in data:
16
+ return data["answer"]
17
+ return data
18
+
19
+
20
+ def extract_json_content(text):
21
+ if not text:
22
+ return None
23
+
24
+ text = text.strip()
25
+ start = text.find('{')
26
+ end = text.rfind('}') + 1
27
+ if start == -1 or end <= start:
28
+ return None
29
+
30
+ try:
31
+ return json.loads(text[start:end])
32
+ except json.JSONDecodeError:
33
+ return None
34
+
35
+
36
+ def normalize_val(val):
37
+ if val is None:
38
+ return ""
39
+ s = re.sub(r"[<>]", "", str(val)).strip()
40
+ if not s:
41
+ return ""
42
+ num_candidate = s.replace(",", "")
43
+ try:
44
+ num = float(num_candidate)
45
+ except ValueError:
46
+ return s
47
+ return str(int(num)) if num.is_integer() else str(num)
48
+
49
+
50
+ def flatten_json(data, prefix=""):
51
+ items = set()
52
+ if isinstance(data, dict):
53
+ for k, v in data.items():
54
+ items.update(flatten_json(v, f"{prefix}.{k}" if prefix else k))
55
+ elif isinstance(data, list):
56
+ for v in data:
57
+ if isinstance(v, list):
58
+ row_tuple = tuple(normalize_val(sub_item) for sub_item in v)
59
+ items.add((prefix, row_tuple))
60
+ else:
61
+ items.update(flatten_json(v, prefix))
62
+ else:
63
+ s = normalize_val(data)
64
+ parts = re.split(r'[,,]', s)
65
+ for part in parts:
66
+ part = part.strip()
67
+ if part:
68
+ items.add((prefix, part))
69
+ return items
70
+
71
+
72
+ def calculate_metrics(pred_json, gt_json, doc_id=None):
73
+ def filter_present_categories(obj):
74
+ return {"present_categories": obj.get("present_categories", [])} if isinstance(obj, dict) else {}
75
+
76
+ if doc_id and doc_id.startswith("DTR_003"):
77
+ pred_json = filter_present_categories(pred_json)
78
+ gt_json = filter_present_categories(gt_json)
79
+
80
+ if doc_id and (doc_id.startswith("NC_001") or doc_id.startswith("NC_002")):
81
+ if pred_json is None:
82
+ return None
83
+ try:
84
+ return 1.0 if abs(float(pred_json) - float(gt_json)) <= 2 else 0.0
85
+ except Exception:
86
+ return 0.0
87
+
88
+ pred_set = flatten_json(pred_json) if pred_json is not None else set()
89
+ gt_set = flatten_json(gt_json)
90
+
91
+ if not gt_set:
92
+ return 1.0 if not pred_set else 0.0
93
+
94
+ tp = len(pred_set & gt_set)
95
+ fp = len(pred_set) - tp
96
+ fn = len(gt_set) - tp
97
+
98
+ precision = tp / (tp + fp) if (tp + fp) else 0.0
99
+ recall = tp / (tp + fn) if (tp + fn) else 0.0
100
+ f1 = (2 * precision * recall) / (precision + recall) if (precision + recall) else 0.0
101
+ return f1
102
+
103
+
104
+ def evaluate_samples(pred_file, gt_file):
105
+ predictions = parse_jsonl(pred_file)
106
+ gts = parse_jsonl(gt_file)
107
+ gt_map = {gt['id']: gt for gt in gts}
108
+ raw_scores = []
109
+ for pred_data in predictions:
110
+ doc_id = pred_data['id']
111
+ # task NC_005 is currently under construction
112
+ if doc_id.startswith('NC_005'):
113
+ continue
114
+ pred_json = extract_json_content(pred_data['response'])
115
+ gt_json = json.loads(gt_map[doc_id]['response'])
116
+ pred_json = unwrap_answer(pred_json)
117
+ gt_json = unwrap_answer(gt_json)
118
+
119
+ f1 = calculate_metrics(pred_json, gt_json, doc_id)
120
+ raw_score = {
121
+ "id": doc_id,
122
+ "task": gt_map[doc_id]['task'],
123
+ "subtask": gt_map[doc_id]['subtask'],
124
+ "robustness": gt_map[doc_id]['robustness'],
125
+ "f1": f1
126
+ }
127
+ raw_scores.append(raw_score)
128
+ return raw_scores
129
+
130
+
131
+ def evaluate_tasks(raw_scores):
132
+ df = pd.DataFrame(raw_scores)
133
+ subtask_mean = (
134
+ df.groupby(["task", "subtask"], as_index=False)
135
+ .agg(subtask_f1_mean=("f1", "mean"))
136
+ )
137
+ task_mean = (
138
+ subtask_mean.groupby("task", as_index=False)
139
+ .agg(task_f1=("subtask_f1_mean", "mean"))
140
+ )
141
+ order = ["DTR", "KIE", "IQE", "CC", "VC", "NC", "RR"]
142
+ task_mean_sorted = (
143
+ task_mean.assign(task=pd.Categorical(task_mean["task"], categories=order, ordered=True))
144
+ .sort_values("task")
145
+ .reset_index(drop=True)
146
+ )
147
+ print("Performance by Task:\n", task_mean_sorted)
148
+
149
+
150
+ def evaluate_robustness(raw_scores):
151
+ df = pd.DataFrame(raw_scores)
152
+ subtask_mean = (
153
+ df.loc[df["task"].ne("IQE")]
154
+ .groupby(["robustness", "task", "subtask"], as_index=False)
155
+ .agg(subtask_f1_mean=("f1", "mean"))
156
+ )
157
+ normal_sub_mean = (
158
+ subtask_mean[subtask_mean["robustness"].eq("Normal Captures")]
159
+ )
160
+ normal_task_mean = (
161
+ normal_sub_mean.groupby("task", as_index=False)
162
+ .agg(task_f1=("subtask_f1_mean", "mean"))
163
+ )
164
+ normal_overall_mean = normal_task_mean["task_f1"].mean()
165
+
166
+ sec_keys = (
167
+ subtask_mean.loc[subtask_mean["robustness"].eq("Secondary Captures"), ["task", "subtask"]]
168
+ .drop_duplicates()
169
+ )
170
+ sec_normal_sub_mean = (
171
+ subtask_mean[subtask_mean["robustness"].eq("Normal Captures")]
172
+ .merge(sec_keys, on=["task", "subtask"], how="inner")
173
+ )
174
+ sec_normal_task_mean = (
175
+ sec_normal_sub_mean.groupby("task", as_index=False)
176
+ .agg(task_f1=("subtask_f1_mean", "mean"))
177
+ )
178
+ sec_normal_overall_mean = sec_normal_task_mean["task_f1"].mean()
179
+
180
+ multi_subtask_mean = (
181
+ df.loc[df["subtask"].isin([
182
+ "DTR_001_001",
183
+ "DTR_003_001",
184
+ "KIE_002_001",
185
+ ])]
186
+ .groupby(["robustness", "task", "subtask"], as_index=False)
187
+ .agg(subtask_f1_mean=("f1", "mean"))
188
+ )
189
+
190
+ multi_normal_sub_mean = (
191
+ multi_subtask_mean[multi_subtask_mean["robustness"].eq("Normal Captures")]
192
+ )
193
+ multi_normal_task_mean = (
194
+ multi_normal_sub_mean.groupby("task", as_index=False)
195
+ .agg(task_f1=("subtask_f1_mean", "mean"))
196
+ )
197
+ multi_normal_overall_mean = multi_normal_task_mean["task_f1"].mean()
198
+
199
+ task_mean = (
200
+ subtask_mean.groupby(["robustness", "task"], as_index=False)
201
+ .agg(task_f1=("subtask_f1_mean", "mean"))
202
+ )
203
+ robustness_scores = (
204
+ task_mean.groupby("robustness", as_index=False)
205
+ .agg(robustness_macro_f1=("task_f1", "mean"))
206
+ .sort_values("robustness_macro_f1", ascending=False)
207
+ .reset_index(drop=True)
208
+ )
209
+ use_special_denoms = {
210
+ "Secondary Captures": sec_normal_overall_mean,
211
+ "Cluttered Background": sec_normal_overall_mean,
212
+ "Multi-doc Images": multi_normal_overall_mean}
213
+ robustness_scores = robustness_scores.copy()
214
+ robustness_scores["normal_denom"] = robustness_scores["robustness"].apply(
215
+ lambda r: use_special_denoms.get(r, normal_overall_mean)
216
+ )
217
+ robustness_scores["relative_to_normal"] = (
218
+ robustness_scores["robustness_macro_f1"] / robustness_scores["normal_denom"].replace(0, np.nan)
219
+ )
220
+ robustness_scores = robustness_scores.drop(columns=["normal_denom", "robustness_macro_f1"])
221
+ robustness_scores = robustness_scores.sort_values('relative_to_normal', ascending=False)
222
+ print("Performance by Robustness:\n", robustness_scores)
223
+
224
+
225
+ if __name__ == "__main__":
226
+ p_file = sys.argv[1]
227
+ g_file = sys.argv[2]
228
+ raw_scores = evaluate_samples(p_file, g_file)
229
+ evaluate_tasks(raw_scores)
230
+ evaluate_robustness(raw_scores)