| | import json |
| | from pathlib import Path |
| |
|
| | import numpy as np |
| | import fire |
| | import h5py |
| | from sklearn.metrics import accuracy_score, mean_absolute_error, f1_score |
| |
|
| | from infer_eval_utils import read_time_series_data, concat_base_path, non_zero_rel_mae, DATASET_TO_TASK |
| |
|
| |
|
| | class Runner(object): |
| |
|
| | def multitask_classification(self, |
| | infer_path: str = "", |
| | gts: list = [], |
| | preds: list = []): |
| | tasks = gts[0].keys() |
| | output_fpath = Path( |
| | infer_path).parent / f"results/{Path(infer_path).stem}.json" |
| | output_fpath.parent.mkdir(parents=True, exist_ok=True) |
| | res_dict = {} |
| | success = 0 |
| | fail = 0 |
| | for task in tasks: |
| | correct_count = 0 |
| | for gt, pred in zip(gts, preds): |
| | try: |
| | if pred[task] == gt[task][0] or pred[task].lower( |
| | ) == gt[task][0]: |
| | correct_count += 1 |
| | success += 1 |
| | except: |
| | fail += 1 |
| | acc = correct_count / len(gts) |
| | print(f"Accuracy for {task}: {acc}") |
| | class_f1s = [] |
| | task_gts = [gt[task][0] for gt in gts] |
| | labels = list(set(task_gts)) |
| | for label in labels: |
| | |
| | true_indices = [ |
| | i for i, gt in enumerate(task_gts) if gt == label |
| | ] |
| | if len(true_indices) == 0: |
| | recall = 0.0 |
| | else: |
| | correct_predictions = 0 |
| | for idx in true_indices: |
| | pred = preds[idx][task] |
| | gt = task_gts[idx] |
| | if pred == gt or pred.lower() == gt: |
| | correct_predictions += 1 |
| |
|
| | recall = correct_predictions / len(true_indices) |
| |
|
| | pred_indices = [ |
| | i for i, pred in enumerate(preds) |
| | if pred[task].lower() == label or pred[task] == label |
| | ] |
| | if len(pred_indices) == 0: |
| | precision = 0.0 |
| | else: |
| | correct_predictions = 0 |
| | for idx in pred_indices: |
| | pred = preds[idx][task] |
| | gt = task_gts[idx] |
| | if pred == gt or pred.lower() == gt: |
| | correct_predictions += 1 |
| | precision = correct_predictions / len(pred_indices) |
| |
|
| | f1 = (2 * recall * precision) / (recall + precision + 1e-6) |
| | class_f1s.append(f1) |
| |
|
| | res_dict[task] = { |
| | "acc": acc, |
| | "f1": f1, |
| | "success": success, |
| | "fail": fail, |
| | "success_rate": success / (success + fail) |
| | } |
| |
|
| | res_dict["overall"] = { |
| | "f1": np.mean(class_f1s), |
| | "acc": np.mean([r["acc"] for r in res_dict.values()]), |
| | } |
| |
|
| | with open(output_fpath, "w") as writer: |
| | json.dump(res_dict, writer, indent=4) |
| | writer.write("\n") |
| |
|
| | def multichoice_classification(self, |
| | infer_path: str = "", |
| | gts: list = [], |
| | preds: list = []): |
| | all_labels = set() |
| | for gt in gts: |
| | if isinstance(gt, list): |
| | all_labels.update(gt) |
| | else: |
| | all_labels.add(gt) |
| |
|
| | all_labels = sorted(list(all_labels)) |
| |
|
| | |
| | y_true_multilabel = [] |
| | y_pred_multilabel = [] |
| | success = 0 |
| | fail = 0 |
| | for gt, pred in zip(gts, preds): |
| | |
| | if isinstance(gt, list): |
| | gt_labels = gt |
| | else: |
| | gt_labels = [gt] |
| |
|
| | |
| | while '\n\n' in pred: |
| | pred = pred.replace('\n\n', '\n') |
| | pred_labels = [x.strip() for x in pred.split("\n")] |
| |
|
| | |
| | gt_binary = [ |
| | 1 if label in gt_labels else 0 for label in all_labels |
| | ] |
| | pred_binary = [] |
| | for label in all_labels: |
| | is_found = False |
| | for pred_label in pred_labels: |
| | if pred_label == label or pred_label.lower() == label: |
| | is_found = True |
| | break |
| | if is_found: |
| | pred_binary.append(1) |
| | else: |
| | pred_binary.append(0) |
| |
|
| | y_true_multilabel.append(gt_binary) |
| | y_pred_multilabel.append(pred_binary) |
| |
|
| | y_true_multilabel = np.array(y_true_multilabel) |
| | y_pred_multilabel = np.array(y_pred_multilabel) |
| |
|
| | |
| | f1_scores = [] |
| | for i, label in enumerate(all_labels): |
| | f1 = f1_score(y_true_multilabel[:, i], |
| | y_pred_multilabel[:, i], |
| | zero_division=0) |
| | f1_scores.append(f1) |
| | print(f"F1 score for class {label}: {f1:.4f}") |
| |
|
| | |
| | macro_f1 = np.mean(f1_scores) |
| | print(f"Macro F1 score (mean of all classes): {macro_f1:.4f}") |
| |
|
| | |
| | output_fpath = Path( |
| | infer_path).parent / f"results/{Path(infer_path).stem}.json" |
| | output_fpath.parent.mkdir(parents=True, exist_ok=True) |
| |
|
| | results = { |
| | "macro_f1": macro_f1, |
| | "per_class_f1": dict(zip(all_labels, f1_scores)) |
| | } |
| |
|
| | with open(output_fpath, "w") as writer: |
| | json.dump(results, writer, indent=4) |
| | writer.write("\n") |
| |
|
| | def classification(self, infer_path: str = ""): |
| | gts, preds = [], [] |
| | with open(infer_path, "r") as f: |
| | for line in f: |
| | item = json.loads(line) |
| | if "id" not in item: |
| | continue |
| | gts.append(item["ground_truth"]) |
| | preds.append(item["output"]) |
| |
|
| | if any(isinstance(gt, list) for gt in gts): |
| | return self.multichoice_classification(infer_path, gts, preds) |
| |
|
| | if isinstance(preds[0], dict): |
| | return self.multitask_classification(infer_path, gts, preds) |
| | |
| | |
| | correct_count = 0 |
| | for gt, pred in zip(gts, preds): |
| | if pred == gt or pred.lower() == gt: |
| | correct_count += 1 |
| |
|
| | acc = correct_count / len(gts) |
| | print(f"Accuracy: {acc}") |
| |
|
| | |
| | labels = list(set(gts)) |
| |
|
| | |
| | class_recalls = [] |
| | class_precisions = [] |
| | class_f1s = [] |
| | for label in labels: |
| | |
| | true_indices = [i for i, gt in enumerate(gts) if gt == label] |
| | if len(true_indices) == 0: |
| | recall = 0.0 |
| | class_recalls.append(0.0) |
| | else: |
| | |
| | correct_predictions = 0 |
| | for idx in true_indices: |
| | pred = preds[idx] |
| | gt = gts[idx] |
| | if pred == gt or pred.lower() == gt: |
| | correct_predictions += 1 |
| |
|
| | recall = correct_predictions / len(true_indices) |
| | class_recalls.append(recall) |
| |
|
| | pred_indices = [ |
| | i for i, pred in enumerate(preds) |
| | if pred.lower() == label or pred == label |
| | ] |
| | if len(pred_indices) == 0: |
| | precision = 0.0 |
| | class_precisions.append(0.0) |
| | else: |
| | correct_predictions = 0 |
| | for idx in pred_indices: |
| | pred = preds[idx] |
| | gt = gts[idx] |
| | if pred == gt or pred.lower() == gt: |
| | correct_predictions += 1 |
| |
|
| | precision = correct_predictions / len(pred_indices) |
| | class_precisions.append(precision) |
| |
|
| | f1 = (2 * recall * precision) / (recall + precision + 1e-6) |
| | class_f1s.append(f1) |
| |
|
| | output_fpath = Path( |
| | infer_path).parent / f"results/{Path(infer_path).stem}.json" |
| | output_fpath.parent.mkdir(parents=True, exist_ok=True) |
| | with open(output_fpath, "w") as writer: |
| | json.dump({ |
| | "acc": acc, |
| | "f1": np.mean(class_f1s), |
| | }, |
| | writer, |
| | indent=4) |
| | writer.write("\n") |
| |
|
| | def mcq(self, infer_path: str = ""): |
| | gts, preds = [], [] |
| | with open(infer_path, "r") as f: |
| | for line in f: |
| | item = json.loads(line) |
| | if "id" not in item: |
| | continue |
| | gts.append(item["ground_truth"]) |
| | preds.append(item["output"]) |
| |
|
| | |
| | |
| | correct_count = 0 |
| | for gt, pred in zip(gts, preds): |
| | if pred == gt or pred.lower() == gt: |
| | correct_count += 1 |
| |
|
| | acc = correct_count / len(gts) |
| | print(f"Accuracy: {acc}") |
| |
|
| | |
| | |
| | try: |
| | labels = list(set(gts)) |
| |
|
| | |
| | class_recalls = [] |
| | for label in labels: |
| | |
| | true_indices = [i for i, gt in enumerate(gts) if gt == label] |
| | if len(true_indices) == 0: |
| | class_recalls.append(0.0) |
| | continue |
| |
|
| | |
| | correct_predictions = 0 |
| | for idx in true_indices: |
| | pred = preds[idx] |
| | gt = gts[idx] |
| | if pred == gt or pred.lower() == gt: |
| | correct_predictions += 1 |
| |
|
| | recall = correct_predictions / len(true_indices) |
| | class_recalls.append(recall) |
| |
|
| | |
| | weighted_recall = np.mean(class_recalls) |
| | print(f"Weighted Recall: {weighted_recall}") |
| |
|
| | output_fpath = Path( |
| | infer_path).parent / f"results/{Path(infer_path).stem}.json" |
| | output_fpath.parent.mkdir(parents=True, exist_ok=True) |
| | with open(output_fpath, "w") as writer: |
| | json.dump({ |
| | "acc": acc, |
| | "uar": weighted_recall |
| | }, |
| | writer, |
| | indent=4) |
| | writer.write("\n") |
| |
|
| | except Exception as e: |
| | print(f"Error calculating Weighted Recall: {e}") |
| | print( |
| | "Possible reasons: labels are not numeric or contain non-numeric labels" |
| | ) |
| |
|
| | def anomaly_detection(self, infer_path: str = ""): |
| | gts, preds = [], [] |
| | success, fail = 0, 0 |
| | with open(infer_path, "r") as f: |
| | for line in f: |
| | item = json.loads(line) |
| | if "id" not in item: |
| | continue |
| |
|
| | if item["output"].lower() == "yes": |
| | preds.append(True) |
| | elif item["output"].lower() == "no": |
| | preds.append(False) |
| | else: |
| | fail += 1 |
| | continue |
| | success += 1 |
| | gts.append(item["ground_truth"]) |
| |
|
| | correct_count = 0 |
| | for gt, pred in zip(gts, preds): |
| | if pred == gt: |
| | correct_count += 1 |
| | print(f"Success: {success}, Fail: {fail}") |
| | acc = correct_count / len(gts) |
| | print(f"Accuracy: {acc}") |
| |
|
| | f1 = f1_score(gts, preds) |
| | print(f"F1 Score: {f1}") |
| | output_fpath = Path( |
| | infer_path).parent / f"results/{Path(infer_path).stem}.json" |
| | output_fpath.parent.mkdir(parents=True, exist_ok=True) |
| | with open(output_fpath, "w") as writer: |
| | json.dump({ |
| | "acc": acc, |
| | "f1": f1, |
| | "success": success, |
| | "fail": fail |
| | }, |
| | writer, |
| | indent=4) |
| | writer.write("\n") |
| |
|
| | def forecasting(self, infer_path: str = ""): |
| | gt_arrs = [] |
| | pred_arrs = [] |
| | success = 0 |
| | fail = 0 |
| | with h5py.File(infer_path, "r") as f: |
| | base_path = Path(f["base_path"][()].decode("utf-8")) |
| | for id in f.keys(): |
| | try: |
| | if id not in [ |
| | "base_path", "dataset_name", "domain", "task", |
| | "scene" |
| | ]: |
| | gt_path = concat_base_path( |
| | base_path, |
| | f[id]["gt_path"][()].decode("utf-8").strip("/")) |
| | gt_data = read_time_series_data(gt_path) |
| | gt_data = np.array(gt_data, dtype=np.float32) |
| | pred = f[id]["pred_result"][()] |
| |
|
| | if pred.shape != gt_data.shape: |
| | raise ValueError( |
| | f"Pred shape {pred.shape} does not match gt shape {gt_data.shape}" |
| | ) |
| | gt_arrs.append(gt_data.reshape(-1)) |
| | pred_arrs.append(pred.reshape(-1)) |
| | success += 1 |
| | except Exception as e: |
| | print(f"Error processing {id}: {e}") |
| | fail += 1 |
| |
|
| | if len(gt_arrs) == 0: |
| | mae = "N/A" |
| | rel_mae = "N/A" |
| | else: |
| | gt_arrs = np.concatenate(gt_arrs) |
| | pred_arrs = np.concatenate(pred_arrs) |
| |
|
| | |
| | mae = mean_absolute_error(gt_arrs, pred_arrs) |
| | rel_mae = non_zero_rel_mae(gt_arrs, pred_arrs) |
| | print( |
| | f"MAE: {mae}, REL_MAE: {rel_mae}, Success: {success}, Fail: {fail}" |
| | ) |
| | output_fpath = Path( |
| | infer_path).parent / f"results/{Path(infer_path).stem}.json" |
| | output_fpath.parent.mkdir(parents=True, exist_ok=True) |
| | with open(output_fpath, "w") as writer: |
| | json.dump( |
| | { |
| | "rel_mae": rel_mae, |
| | "mae": mae, |
| | "success": success, |
| | "fail": fail, |
| | "success_rate": success / (success + fail) |
| | }, |
| | writer, |
| | indent=4) |
| | writer.write("\n") |
| |
|
| | def synthesize(self, infer_path: str = ""): |
| | return self.forecasting(infer_path) |
| |
|
| | def imputation(self, infer_path: str = ""): |
| | gt_arrs = [] |
| | pred_arrs = [] |
| | success = 0 |
| | fail = 0 |
| | with h5py.File(infer_path, "r") as f: |
| | base_path = Path(f["base_path"][()].decode("utf-8")) |
| | for id in f.keys(): |
| | try: |
| | if id not in [ |
| | "base_path", "dataset_name", "domain", "task", |
| | "scene" |
| | ]: |
| | |
| | |
| | gt_path = concat_base_path( |
| | base_path, |
| | f[id]["gt_path"][()].decode("utf-8").strip("/")) |
| | gt_data = read_time_series_data(gt_path) |
| |
|
| | |
| | |
| | input_path = concat_base_path( |
| | base_path, f[id]["input_ts_path"][( |
| | )].decode("utf-8").strip("/")) |
| | input_data = read_time_series_data(input_path) |
| |
|
| | pred_indices = np.where(input_data == "X")[0] |
| | pred = f[id]["pred_result"][()] |
| |
|
| | pred = pred[pred_indices] |
| | gt_data = gt_data[pred_indices] |
| | if len(pred) != len(gt_data): |
| | length_mismatch += 1 |
| | else: |
| | success += 1 |
| | if len(pred) < len(gt_data): |
| | pred = pred[:len(gt_data)] |
| | if len(pred) > len(gt_data): |
| | gt_data = gt_data[:len(pred)] |
| | gt_arrs.append(gt_data) |
| | pred_arrs.append(pred) |
| | |
| | except Exception as e: |
| | print(f"Error processing {id}: {e}") |
| | fail += 1 |
| |
|
| | gt_arrs = np.concatenate(gt_arrs) |
| | pred_arrs = np.concatenate(pred_arrs) |
| | |
| | rel_mae = non_zero_rel_mae(gt_arrs, pred_arrs) |
| | mae = mean_absolute_error(gt_arrs, pred_arrs) |
| |
|
| | print( |
| | f"REL_MAE: {rel_mae}, MAE: {mae}, Success: {success}, Fail: {fail}" |
| | ) |
| | output_fpath = Path( |
| | infer_path).parent / f"results/{Path(infer_path).stem}.json" |
| | output_fpath.parent.mkdir(parents=True, exist_ok=True) |
| | with open(output_fpath, "w") as writer: |
| | json.dump( |
| | { |
| | "rel_mae": rel_mae, |
| | "mae": mae, |
| | "success": success, |
| | "fail": fail, |
| | "success_rate": success / (success + fail) |
| | }, |
| | writer, |
| | indent=4) |
| | writer.write("\n") |
| |
|
| | def event_detection(self, infer_path: str = ""): |
| | event_gts, event_preds = [], [] |
| | seq_length = None |
| | success = 0 |
| | total = 0 |
| | timestamp_gts, timestamp_preds = [], [] |
| | with open(infer_path, "r") as f: |
| | for line in f: |
| | item = json.loads(line) |
| | if "id" not in item: |
| | seq_length = item["seq_length"] |
| | continue |
| | event_gt = item["ground_truth"]["contain"] |
| | event_gts.append(1 if event_gt else 0) |
| |
|
| | if "\n" in item["output"]: |
| | while '\n\n' in item["output"]: |
| | item["output"] = item["output"].replace('\n\n', '\n') |
| | event_pred, *timestamps = item["output"].split("\n") |
| | else: |
| | event_pred = item["output"] |
| | timestamps = None |
| | event_preds.append(1 if event_pred.lower() == "yes" else 0) |
| |
|
| | if event_gt: |
| | if "start_time" in item["ground_truth"]: |
| | gt_timestamps = [item["ground_truth"]["start_time"]] |
| | elif "start_time_p" in item["ground_truth"]: |
| | gt_timestamps = [ |
| | item["ground_truth"]["start_time_p"], |
| | item["ground_truth"]["start_time_s"] |
| | ] |
| |
|
| | if timestamps is None: |
| | pass |
| | else: |
| | try: |
| | assert len(timestamps) == len(gt_timestamps) |
| | for pred_timestamp, gt_timestamp in zip( |
| | timestamps, gt_timestamps): |
| | pred_timestamp = eval(pred_timestamp) |
| | timestamp_preds.append(pred_timestamp) |
| | timestamp_gts.append(gt_timestamp) |
| | success += 1 |
| | except Exception as e: |
| | print(str(e)) |
| | total += 1 |
| |
|
| | event_acc = accuracy_score(event_gts, event_preds) |
| | event_f1 = f1_score(event_gts, event_preds) |
| | timestamp_gts = np.array(timestamp_gts) |
| | timestamp_preds = np.array(timestamp_preds) |
| | mape = non_zero_rel_mae(timestamp_gts, timestamp_preds) |
| | output_fpath = Path( |
| | infer_path).parent / f"results/{Path(infer_path).stem}.json" |
| | output_fpath.parent.mkdir(parents=True, exist_ok=True) |
| | with open(output_fpath, "w") as writer: |
| | json.dump( |
| | { |
| | "acc": event_acc, |
| | "f1": event_f1, |
| | "mape": mape, |
| | "success_rate": success / total |
| | }, |
| | writer, |
| | indent=4) |
| | writer.write("\n") |
| | print({ |
| | "acc": event_acc, |
| | "f1": event_f1, |
| | "mape": mape, |
| | "success_rate": success / total |
| | }) |
| |
|
| | def evaluate(self, infer_dir: str): |
| | for infer_path in Path(infer_dir).glob("*"): |
| | if infer_path.is_dir(): |
| | continue |
| | dataset_id = infer_path.stem |
| | task = DATASET_TO_TASK[dataset_id] |
| | print(f"evaluating {dataset_id} ...") |
| | getattr(self, task)(infer_path) |
| |
|
| |
|
| | if __name__ == "__main__": |
| | fire.Fire(Runner) |
| |
|