|
|
import json |
|
|
import numpy as np |
|
|
from omegaconf import open_dict |
|
|
from fvcore.common.registry import Registry |
|
|
|
|
|
from common.misc import gather_dict |
|
|
|
|
|
EVALUATOR_REGISTRY = Registry("EVALUATOR") |
|
|
|
|
|
|
|
|
class BaseEvaluator(): |
|
|
def __init__(self, cfg, accelerator): |
|
|
self.accelerator = accelerator |
|
|
self.best_result = -np.inf |
|
|
self.save = cfg.eval.save |
|
|
self.save_dir.mkdir(parents=True, exist_ok=True) |
|
|
self.reset() |
|
|
|
|
|
def reset(self): |
|
|
self.eval_results = [] |
|
|
self.eval_dict = {} |
|
|
|
|
|
def batch_metrics(self, data_dict, include_count=False): |
|
|
raise NotImplementedError("Per batch metrics calculation is required for evaluation") |
|
|
|
|
|
def update(self, data_dict): |
|
|
metrics = self.batch_metrics(data_dict, include_count=True) |
|
|
for key in metrics.keys(): |
|
|
if key not in self.eval_dict: |
|
|
self.eval_dict[key] = [] |
|
|
self.eval_dict[key].append(metrics[key]) |
|
|
|
|
|
def record(self): |
|
|
self.eval_dict = gather_dict(self.accelerator, self.eval_dict) |
|
|
for k, metrics in self.eval_dict.items(): |
|
|
if not isinstance(metrics, list): |
|
|
continue |
|
|
|
|
|
total_value = sum(x[0] for x in metrics) |
|
|
total_count = sum(x[1] for x in metrics) |
|
|
self.eval_dict[k] = total_value / max(total_count, 1) |
|
|
print(k, total_value, total_count) |
|
|
|
|
|
if self.save and self.accelerator.is_main_process: |
|
|
with (self.save_dir / "results.json").open("w") as f: |
|
|
json.dump(self.eval_results, f) |
|
|
|
|
|
self.eval_dict['target_metric'] = self.eval_dict[self.target_metric] |
|
|
if self.eval_dict["target_metric"] > self.best_result: |
|
|
is_best = True |
|
|
self.best_result = self.eval_dict["target_metric"] |
|
|
else: |
|
|
is_best = False |
|
|
self.eval_dict['best_result'] = self.best_result |
|
|
return is_best, self.eval_dict |
|
|
|
|
|
|
|
|
def get_eval(name, cfg, accelerator, **kwargs): |
|
|
"""Get an evaluator or a list of evaluators.""" |
|
|
if isinstance(name, str): |
|
|
eval = EVALUATOR_REGISTRY.get(name)(cfg, accelerator, **kwargs) |
|
|
else: |
|
|
eval = [EVALUATOR_REGISTRY.get(i)(cfg, accelerator, **kwargs) for i in name] |
|
|
return eval |
|
|
|
|
|
def build_eval(cfg, accelerator, **kwargs): |
|
|
if cfg.eval.get("train", None) is not None: |
|
|
train_eval = get_eval(cfg.eval.train.name, cfg, accelerator, **kwargs) |
|
|
val_eval = get_eval(cfg.eval.val.name, cfg, accelerator, **kwargs) |
|
|
return {"train": train_eval, "val": val_eval} |
|
|
elif cfg.eval.get("name", None) is not None: |
|
|
return get_eval(cfg.eval.name, cfg, accelerator, **kwargs) |
|
|
else: |
|
|
with open_dict(cfg): |
|
|
cfg.eval.name = [cfg.data.get(dataset).evaluator for dataset in cfg.data.val] |
|
|
return get_eval(cfg.eval.name, cfg, accelerator, **kwargs) |