backup / evaluator /objcls_eval.py
MatchLab's picture
Upload folder using huggingface_hub
c94c8c9 verified
import torch
from pathlib import Path
from evaluator.build import EVALUATOR_REGISTRY, BaseEvaluator
@EVALUATOR_REGISTRY.register()
class PretrainObjEval(BaseEvaluator):
def __init__(self, cfg, accelerator, **kwargs):
self.target_metric = "accuracy"
self.save_dir = Path(cfg.exp_dir) / "eval_results" / self.__class__.__name__
super().__init__(cfg, accelerator, **kwargs)
def batch_metrics(self, data_dict, include_count=False):
metrics = {}
logits = data_dict["obj_logits"][data_dict["obj_masks"]].view(-1, data_dict["obj_logits"].shape[-1])
labels = data_dict["obj_labels"][data_dict["obj_masks"]].view(-1)
_, pred = torch.max(logits, 1)
metrics["accuracy"] = ((pred == labels.view(-1)).sum().item(), labels.shape[0])
if not include_count:
for key, v in metrics.items():
metrics[key] = v[0] / max(v[1], 1)
return metrics