| import json |
| import numpy as np |
| from omegaconf import open_dict |
| from fvcore.common.registry import Registry |
|
|
| from common.misc import gather_dict |
|
|
| EVALUATOR_REGISTRY = Registry("EVALUATOR") |
|
|
|
|
| class BaseEvaluator(): |
| def __init__(self, cfg, accelerator): |
| self.accelerator = accelerator |
| self.best_result = -np.inf |
| self.save = cfg.eval.save |
| self.save_dir.mkdir(parents=True, exist_ok=True) |
| self.reset() |
|
|
| def reset(self): |
| self.eval_results = [] |
| self.eval_dict = {} |
|
|
| def batch_metrics(self, data_dict, include_count=False): |
| raise NotImplementedError("Per batch metrics calculation is required for evaluation") |
|
|
| def update(self, data_dict): |
| metrics = self.batch_metrics(data_dict, include_count=True) |
| for key in metrics.keys(): |
| if key not in self.eval_dict: |
| self.eval_dict[key] = [] |
| self.eval_dict[key].append(metrics[key]) |
|
|
| def record(self): |
| self.eval_dict = gather_dict(self.accelerator, self.eval_dict) |
| for k, metrics in self.eval_dict.items(): |
| if not isinstance(metrics, list): |
| continue |
| |
| total_value = sum(x[0] for x in metrics) |
| total_count = sum(x[1] for x in metrics) |
| self.eval_dict[k] = total_value / max(total_count, 1) |
| print(k, total_value, total_count) |
| |
| if self.save and self.accelerator.is_main_process: |
| with (self.save_dir / "results.json").open("w") as f: |
| json.dump(self.eval_results, f) |
| |
| self.eval_dict['target_metric'] = self.eval_dict[self.target_metric] |
| if self.eval_dict["target_metric"] > self.best_result: |
| is_best = True |
| self.best_result = self.eval_dict["target_metric"] |
| else: |
| is_best = False |
| self.eval_dict['best_result'] = self.best_result |
| return is_best, self.eval_dict |
|
|
|
|
| def get_eval(name, cfg, accelerator, **kwargs): |
| """Get an evaluator or a list of evaluators.""" |
| if isinstance(name, str): |
| eval = EVALUATOR_REGISTRY.get(name)(cfg, accelerator, **kwargs) |
| else: |
| eval = [EVALUATOR_REGISTRY.get(i)(cfg, accelerator, **kwargs) for i in name] |
| return eval |
|
|
| def build_eval(cfg, accelerator, **kwargs): |
| if cfg.eval.get("train", None) is not None: |
| train_eval = get_eval(cfg.eval.train.name, cfg, accelerator, **kwargs) |
| val_eval = get_eval(cfg.eval.val.name, cfg, accelerator, **kwargs) |
| return {"train": train_eval, "val": val_eval} |
| elif cfg.eval.get("name", None) is not None: |
| return get_eval(cfg.eval.name, cfg, accelerator, **kwargs) |
| else: |
| with open_dict(cfg): |
| cfg.eval.name = [cfg.data.get(dataset).evaluator for dataset in cfg.data.val] |
| return get_eval(cfg.eval.name, cfg, accelerator, **kwargs) |