|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
""" |
|
|
This script is almost same with https://github.com/MIC-DKFZ/nnDetection/blob/main/nndet/evaluator/detection/coco.py |
|
|
The changes include 1) code reformatting, 2) docstrings. |
|
|
""" |
|
|
|
|
|
from __future__ import annotations |
|
|
|
|
|
import logging as logger |
|
|
import time |
|
|
from collections.abc import Sequence |
|
|
from typing import Any |
|
|
|
|
|
import numpy as np |
|
|
|
|
|
|
|
|
class COCOMetric: |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
classes: Sequence[str], |
|
|
iou_list: Sequence[float] = (0.1, 0.5, 0.75), |
|
|
iou_range: Sequence[float] = (0.1, 0.5, 0.05), |
|
|
max_detection: Sequence[int] = (1, 5, 100), |
|
|
per_class: bool = True, |
|
|
verbose: bool = True, |
|
|
): |
|
|
""" |
|
|
Class to compute COCO metrics |
|
|
Metrics computed includes, |
|
|
|
|
|
- mAP over the IoU range specified by `iou_range` at last value of `max_detection` |
|
|
- AP values at IoU thresholds specified by `iou_list` at last value of `max_detection` |
|
|
- AR over max detections thresholds defined by `max_detection` (over iou range) |
|
|
|
|
|
Args: |
|
|
classes (Sequence[str]): name of each class (index needs to correspond to predicted class indices!) |
|
|
iou_list (Sequence[float]): specific thresholds where ap is evaluated and saved |
|
|
iou_range (Sequence[float]): (start, stop, step) for mAP iou thresholds |
|
|
max_detection (Sequence[int]): maximum number of detections per image |
|
|
verbose (bool): log time needed for evaluation |
|
|
|
|
|
Example: |
|
|
|
|
|
.. code-block:: python |
|
|
|
|
|
from monai.data.box_utils import box_iou |
|
|
from monai.apps.detection.metrics.coco import COCOMetric |
|
|
from monai.apps.detection.metrics.matching import matching_batch |
|
|
# 3D example outputs of one image from detector |
|
|
val_outputs_all = [ |
|
|
{"boxes": torch.tensor([[1,1,1,3,4,5]],dtype=torch.float16), |
|
|
"labels": torch.randint(3,(1,)), |
|
|
"scores": torch.randn((1,)).absolute()}, |
|
|
] |
|
|
val_targets_all = [ |
|
|
{"boxes": torch.tensor([[1,1,1,2,6,4]],dtype=torch.float16), |
|
|
"labels": torch.randint(3,(1,))}, |
|
|
] |
|
|
|
|
|
coco_metric = COCOMetric( |
|
|
classes=['c0','c1','c2'], iou_list=[0.1], max_detection=[10] |
|
|
) |
|
|
results_metric = matching_batch( |
|
|
iou_fn=box_iou, |
|
|
iou_thresholds=coco_metric.iou_thresholds, |
|
|
pred_boxes=[val_data_i["boxes"].numpy() for val_data_i in val_outputs_all], |
|
|
pred_classes=[val_data_i["labels"].numpy() for val_data_i in val_outputs_all], |
|
|
pred_scores=[val_data_i["scores"].numpy() for val_data_i in val_outputs_all], |
|
|
gt_boxes=[val_data_i["boxes"].numpy() for val_data_i in val_targets_all], |
|
|
gt_classes=[val_data_i["labels"].numpy() for val_data_i in val_targets_all], |
|
|
) |
|
|
val_metric_dict = coco_metric(results_metric) |
|
|
print(val_metric_dict) |
|
|
""" |
|
|
self.verbose = verbose |
|
|
self.classes = classes |
|
|
self.per_class = per_class |
|
|
|
|
|
iou_list_np = np.array(iou_list) |
|
|
_iou_range = np.linspace( |
|
|
iou_range[0], iou_range[1], int(np.round((iou_range[1] - iou_range[0]) / iou_range[2])) + 1, endpoint=True |
|
|
) |
|
|
self.iou_thresholds = np.union1d(iou_list_np, _iou_range) |
|
|
self.iou_range = iou_range |
|
|
|
|
|
|
|
|
self.iou_list_idx = np.nonzero(iou_list_np[:, np.newaxis] == self.iou_thresholds[np.newaxis])[1] |
|
|
self.iou_range_idx = np.nonzero(_iou_range[:, np.newaxis] == self.iou_thresholds[np.newaxis])[1] |
|
|
|
|
|
if ( |
|
|
not (self.iou_thresholds[self.iou_list_idx] == iou_list_np).all() |
|
|
or not (self.iou_thresholds[self.iou_range_idx] == _iou_range).all() |
|
|
): |
|
|
raise ValueError( |
|
|
"Require self.iou_thresholds[self.iou_list_idx] == iou_list_np and " |
|
|
"self.iou_thresholds[self.iou_range_idx] == _iou_range." |
|
|
) |
|
|
|
|
|
self.recall_thresholds = np.linspace(0.0, 1.00, int(np.round((1.00 - 0.0) / 0.01)) + 1, endpoint=True) |
|
|
self.max_detections = max_detection |
|
|
|
|
|
def __call__(self, *args: Any, **kwargs: Any) -> tuple[dict[str, float], dict[str, np.ndarray] | None]: |
|
|
""" |
|
|
Compute metric. See :func:`compute` for more information. |
|
|
|
|
|
Args: |
|
|
*args: positional arguments passed to :func:`compute` |
|
|
**kwargs: keyword arguments passed to :func:`compute` |
|
|
|
|
|
Returns: |
|
|
dict[str, float]: dictionary with scalar values for evaluation |
|
|
dict[str, np.ndarray]: dictionary with arrays, e.g. for visualization of graphs |
|
|
""" |
|
|
return self.compute(*args, **kwargs) |
|
|
|
|
|
def check_number_of_iou(self, *args: np.ndarray) -> None: |
|
|
""" |
|
|
Check if shape of input in first dimension is consistent with expected IoU values |
|
|
(assumes IoU dimension is the first dimension) |
|
|
|
|
|
Args: |
|
|
args: array like inputs with shape function |
|
|
""" |
|
|
num_ious = len(self.get_iou_thresholds()) |
|
|
for arg in args: |
|
|
if arg.shape[0] != num_ious: |
|
|
raise ValueError( |
|
|
f"Require arg.shape[0] == len(self.get_iou_thresholds()). Got arg.shape[0]={arg.shape[0]}, " |
|
|
f"self.get_iou_thresholds()={self.get_iou_thresholds()}." |
|
|
) |
|
|
|
|
|
def get_iou_thresholds(self) -> Sequence[float]: |
|
|
""" |
|
|
Return IoU thresholds needed for this metric in an numpy array |
|
|
|
|
|
Returns: |
|
|
Sequence[float]: IoU thresholds [M], M is the number of thresholds |
|
|
""" |
|
|
return list(self.iou_thresholds) |
|
|
|
|
|
def compute(self, results_list: list[dict[int, dict[str, np.ndarray]]]) -> tuple[dict[str, float], None]: |
|
|
""" |
|
|
Compute COCO metrics |
|
|
|
|
|
Args: |
|
|
results_list (list[dict[int, dict[str, np.ndarray]]]): list with results per image (in list) |
|
|
per category (dict). Inner dict contains multiple results obtained by :func:`box_matching_batch`. |
|
|
|
|
|
- `dtMatches`: matched detections [T, D], where T = number of |
|
|
thresholds, D = number of detections |
|
|
- `gtMatches`: matched ground truth boxes [T, G], where T = number |
|
|
of thresholds, G = number of ground truth |
|
|
- `dtScores`: prediction scores [D] detection scores |
|
|
- `gtIgnore`: ground truth boxes which should be ignored |
|
|
[G] indicate whether ground truth should be ignored |
|
|
- `dtIgnore`: detections which should be ignored [T, D], |
|
|
indicate which detections should be ignored |
|
|
|
|
|
Returns: |
|
|
dict[str, float], dictionary with coco metrics |
|
|
""" |
|
|
if self.verbose: |
|
|
logger.info("Start COCO metric computation...") |
|
|
tic = time.time() |
|
|
|
|
|
dataset_statistics = self._compute_statistics(results_list=results_list) |
|
|
|
|
|
if self.verbose: |
|
|
toc = time.time() |
|
|
logger.info(f"Statistics for COCO metrics finished (t={(toc - tic):0.2f}s).") |
|
|
|
|
|
results = {} |
|
|
results.update(self._compute_ap(dataset_statistics)) |
|
|
results.update(self._compute_ar(dataset_statistics)) |
|
|
|
|
|
if self.verbose: |
|
|
toc = time.time() |
|
|
logger.info(f"COCO metrics computed in t={(toc - tic):0.2f}s.") |
|
|
return results, None |
|
|
|
|
|
def _compute_ap(self, dataset_statistics: dict[str, np.ndarray | list]) -> dict[str, float]: |
|
|
""" |
|
|
Compute AP metrics |
|
|
|
|
|
Args: |
|
|
dataset_statistics (list[dict[int, dict[str, np.ndarray]]]): list with result s per image (in list) |
|
|
per category (dict). Inner dict contains multiple results obtained by :func:`box_matching_batch`. |
|
|
|
|
|
- `dtMatches`: matched detections [T, D], where T = number of |
|
|
thresholds, D = number of detections |
|
|
- `gtMatches`: matched ground truth boxes [T, G], where T = number |
|
|
of thresholds, G = number of ground truth |
|
|
- `dtScores`: prediction scores [D] detection scores |
|
|
- `gtIgnore`: ground truth boxes which should be ignored |
|
|
[G] indicate whether ground truth should be ignored |
|
|
- `dtIgnore`: detections which should be ignored [T, D], |
|
|
indicate which detections should be ignored |
|
|
""" |
|
|
results = {} |
|
|
if self.iou_range: |
|
|
key = ( |
|
|
f"mAP_IoU_{self.iou_range[0]:.2f}_{self.iou_range[1]:.2f}_{self.iou_range[2]:.2f}_" |
|
|
f"MaxDet_{self.max_detections[-1]}" |
|
|
) |
|
|
results[key] = self._select_ap(dataset_statistics, iou_idx=self.iou_range_idx, max_det_idx=-1) |
|
|
|
|
|
if self.per_class: |
|
|
for cls_idx, cls_str in enumerate(self.classes): |
|
|
key = ( |
|
|
f"{cls_str}_" |
|
|
f"mAP_IoU_{self.iou_range[0]:.2f}_{self.iou_range[1]:.2f}_{self.iou_range[2]:.2f}_" |
|
|
f"MaxDet_{self.max_detections[-1]}" |
|
|
) |
|
|
results[key] = self._select_ap( |
|
|
dataset_statistics, iou_idx=self.iou_range_idx, cls_idx=cls_idx, max_det_idx=-1 |
|
|
) |
|
|
|
|
|
for idx in self.iou_list_idx: |
|
|
key = f"AP_IoU_{self.iou_thresholds[idx]:.2f}_MaxDet_{self.max_detections[-1]}" |
|
|
results[key] = self._select_ap(dataset_statistics, iou_idx=[idx], max_det_idx=-1) |
|
|
|
|
|
if self.per_class: |
|
|
for cls_idx, cls_str in enumerate(self.classes): |
|
|
key = f"{cls_str}_" f"AP_IoU_{self.iou_thresholds[idx]:.2f}_" f"MaxDet_{self.max_detections[-1]}" |
|
|
results[key] = self._select_ap(dataset_statistics, iou_idx=[idx], cls_idx=cls_idx, max_det_idx=-1) |
|
|
return results |
|
|
|
|
|
def _compute_ar(self, dataset_statistics: dict[str, np.ndarray | list]) -> dict[str, float]: |
|
|
""" |
|
|
Compute AR metrics |
|
|
|
|
|
Args: |
|
|
dataset_statistics (list[dict[int, dict[str, np.ndarray]]]): list with result s per image (in list) |
|
|
per category (dict). Inner dict contains multiple results obtained by :func:`box_matching_batch`. |
|
|
|
|
|
- `dtMatches`: matched detections [T, D], where T = number of |
|
|
thresholds, D = number of detections |
|
|
- `gtMatches`: matched ground truth boxes [T, G], where T = number |
|
|
of thresholds, G = number of ground truth |
|
|
- `dtScores`: prediction scores [D] detection scores |
|
|
- `gtIgnore`: ground truth boxes which should be ignored |
|
|
[G] indicate whether ground truth should be ignored |
|
|
- `dtIgnore`: detections which should be ignored [T, D], |
|
|
indicate which detections should be ignored |
|
|
""" |
|
|
results = {} |
|
|
for max_det_idx, max_det in enumerate(self.max_detections): |
|
|
key = f"mAR_IoU_{self.iou_range[0]:.2f}_{self.iou_range[1]:.2f}_{self.iou_range[2]:.2f}_MaxDet_{max_det}" |
|
|
results[key] = self._select_ar(dataset_statistics, max_det_idx=max_det_idx) |
|
|
|
|
|
if self.per_class: |
|
|
for cls_idx, cls_str in enumerate(self.classes): |
|
|
key = ( |
|
|
f"{cls_str}_" |
|
|
f"mAR_IoU_{self.iou_range[0]:.2f}_{self.iou_range[1]:.2f}_{self.iou_range[2]:.2f}_" |
|
|
f"MaxDet_{max_det}" |
|
|
) |
|
|
results[key] = self._select_ar(dataset_statistics, cls_idx=cls_idx, max_det_idx=max_det_idx) |
|
|
|
|
|
for idx in self.iou_list_idx: |
|
|
key = f"AR_IoU_{self.iou_thresholds[idx]:.2f}_MaxDet_{self.max_detections[-1]}" |
|
|
results[key] = self._select_ar(dataset_statistics, iou_idx=idx, max_det_idx=-1) |
|
|
|
|
|
if self.per_class: |
|
|
for cls_idx, cls_str in enumerate(self.classes): |
|
|
key = f"{cls_str}_" f"AR_IoU_{self.iou_thresholds[idx]:.2f}_" f"MaxDet_{self.max_detections[-1]}" |
|
|
results[key] = self._select_ar(dataset_statistics, iou_idx=idx, cls_idx=cls_idx, max_det_idx=-1) |
|
|
return results |
|
|
|
|
|
@staticmethod |
|
|
def _select_ap( |
|
|
dataset_statistics: dict, |
|
|
iou_idx: int | list[int] | np.ndarray | None = None, |
|
|
cls_idx: int | Sequence[int] | None = None, |
|
|
max_det_idx: int = -1, |
|
|
) -> float: |
|
|
""" |
|
|
Compute average precision |
|
|
|
|
|
Args: |
|
|
dataset_statistics (dict): computed statistics over dataset |
|
|
|
|
|
- `counts`: Number of thresholds, Number recall thresholds, Number of classes, Number of max |
|
|
detection thresholds |
|
|
- `recall`: Computed recall values [num_iou_th, num_classes, num_max_detections] |
|
|
- `precision`: Precision values at specified recall thresholds |
|
|
[num_iou_th, num_recall_th, num_classes, num_max_detections] |
|
|
- `scores`: Scores corresponding to specified recall thresholds |
|
|
[num_iou_th, num_recall_th, num_classes, num_max_detections] |
|
|
iou_idx: index of IoU values to select for evaluation(if None, all values are used) |
|
|
cls_idx: class indices to select, if None all classes will be selected |
|
|
max_det_idx (int): index to select max detection threshold from data |
|
|
|
|
|
Returns: |
|
|
np.ndarray: AP value |
|
|
""" |
|
|
prec = dataset_statistics["precision"] |
|
|
if iou_idx is not None: |
|
|
prec = prec[iou_idx] |
|
|
if cls_idx is not None: |
|
|
prec = prec[..., cls_idx, :] |
|
|
prec = prec[..., max_det_idx] |
|
|
return float(np.mean(prec)) |
|
|
|
|
|
@staticmethod |
|
|
def _select_ar( |
|
|
dataset_statistics: dict, |
|
|
iou_idx: int | Sequence[int] | None = None, |
|
|
cls_idx: int | Sequence[int] | None = None, |
|
|
max_det_idx: int = -1, |
|
|
) -> float: |
|
|
""" |
|
|
Compute average recall |
|
|
|
|
|
Args: |
|
|
dataset_statistics (dict): computed statistics over dataset |
|
|
|
|
|
- `counts`: Number of thresholds, Number recall thresholds, Number of classes, Number of max |
|
|
detection thresholds |
|
|
- `recall`: Computed recall values [num_iou_th, num_classes, num_max_detections] |
|
|
- `precision`: Precision values at specified recall thresholds |
|
|
[num_iou_th, num_recall_th, num_classes, num_max_detections] |
|
|
- `scores`: Scores corresponding to specified recall thresholds |
|
|
[num_iou_th, num_recall_th, num_classes, num_max_detections] |
|
|
iou_idx: index of IoU values to select for evaluation(if None, all values are used) |
|
|
cls_idx: class indices to select, if None all classes will be selected |
|
|
max_det_idx (int): index to select max detection threshold from data |
|
|
|
|
|
Returns: |
|
|
np.ndarray: recall value |
|
|
""" |
|
|
rec = dataset_statistics["recall"] |
|
|
if iou_idx is not None: |
|
|
rec = rec[iou_idx] |
|
|
if cls_idx is not None: |
|
|
rec = rec[..., cls_idx, :] |
|
|
rec = rec[..., max_det_idx] |
|
|
|
|
|
if len(rec[rec > -1]) == 0: |
|
|
return -1.0 |
|
|
|
|
|
return float(np.mean(rec[rec > -1])) |
|
|
|
|
|
def _compute_statistics(self, results_list: list[dict[int, dict[str, np.ndarray]]]) -> dict[str, np.ndarray | list]: |
|
|
""" |
|
|
Compute statistics needed for COCO metrics (mAP, AP of individual classes, mAP@IoU_Thresholds, AR) |
|
|
Adapted from https://github.com/cocodataset/cocoapi/blob/master/PythonAPI/pycocotools/cocoeval.py |
|
|
|
|
|
Args: |
|
|
results_list (list[dict[int, dict[str, np.ndarray]]]): list with result s per image (in list) |
|
|
per category (dict). Inner dict contains multiple results obtained by :func:`box_matching_batch`. |
|
|
|
|
|
- `dtMatches`: matched detections [T, D], where T = number of |
|
|
thresholds, D = number of detections |
|
|
- `gtMatches`: matched ground truth boxes [T, G], where T = number |
|
|
of thresholds, G = number of ground truth |
|
|
- `dtScores`: prediction scores [D] detection scores |
|
|
- `gtIgnore`: ground truth boxes which should be ignored |
|
|
[G] indicate whether ground truth should be ignored |
|
|
- `dtIgnore`: detections which should be ignored [T, D], |
|
|
indicate which detections should be ignored |
|
|
|
|
|
Returns: |
|
|
dict: computed statistics over dataset |
|
|
- `counts`: Number of thresholds, Number recall thresholds, Number of classes, Number of max |
|
|
detection thresholds |
|
|
- `recall`: Computed recall values [num_iou_th, num_classes, num_max_detections] |
|
|
- `precision`: Precision values at specified recall thresholds |
|
|
[num_iou_th, num_recall_th, num_classes, num_max_detections] |
|
|
- `scores`: Scores corresponding to specified recall thresholds |
|
|
[num_iou_th, num_recall_th, num_classes, num_max_detections] |
|
|
""" |
|
|
num_iou_th = len(self.iou_thresholds) |
|
|
num_recall_th = len(self.recall_thresholds) |
|
|
num_classes = len(self.classes) |
|
|
num_max_detections = len(self.max_detections) |
|
|
|
|
|
|
|
|
precision = -np.ones((num_iou_th, num_recall_th, num_classes, num_max_detections)) |
|
|
recall = -np.ones((num_iou_th, num_classes, num_max_detections)) |
|
|
scores = -np.ones((num_iou_th, num_recall_th, num_classes, num_max_detections)) |
|
|
|
|
|
for cls_idx, cls_i in enumerate(self.classes): |
|
|
for max_det_idx, max_det in enumerate(self.max_detections): |
|
|
results = [r[cls_idx] for r in results_list if cls_idx in r] |
|
|
|
|
|
if len(results) == 0: |
|
|
logger.warning(f"WARNING, no results found for coco metric for class {cls_i}") |
|
|
continue |
|
|
|
|
|
dt_scores = np.concatenate([r["dtScores"][0:max_det] for r in results]) |
|
|
|
|
|
|
|
|
inds = np.argsort(-dt_scores, kind="mergesort") |
|
|
dt_scores_sorted = dt_scores[inds] |
|
|
|
|
|
|
|
|
dt_matches = np.concatenate([r["dtMatches"][:, 0:max_det] for r in results], axis=1)[:, inds] |
|
|
dt_ignores = np.concatenate([r["dtIgnore"][:, 0:max_det] for r in results], axis=1)[:, inds] |
|
|
self.check_number_of_iou(dt_matches, dt_ignores) |
|
|
gt_ignore = np.concatenate([r["gtIgnore"] for r in results]) |
|
|
num_gt = np.count_nonzero(gt_ignore == 0) |
|
|
if num_gt == 0: |
|
|
logger.warning(f"WARNING, no gt found for coco metric for class {cls_i}") |
|
|
continue |
|
|
|
|
|
|
|
|
tps = np.logical_and(dt_matches, np.logical_not(dt_ignores)) |
|
|
fps = np.logical_and(np.logical_not(dt_matches), np.logical_not(dt_ignores)) |
|
|
|
|
|
tp_sum = np.cumsum(tps, axis=1).astype(dtype=np.float32) |
|
|
fp_sum = np.cumsum(fps, axis=1).astype(dtype=np.float32) |
|
|
|
|
|
for th_ind, (tp, fp) in enumerate(zip(tp_sum, fp_sum)): |
|
|
tp, fp = np.array(tp), np.array(fp) |
|
|
r, p, s = _compute_stats_single_threshold(tp, fp, dt_scores_sorted, self.recall_thresholds, num_gt) |
|
|
recall[th_ind, cls_idx, max_det_idx] = r |
|
|
precision[th_ind, :, cls_idx, max_det_idx] = p |
|
|
|
|
|
scores[th_ind, :, cls_idx, max_det_idx] = s |
|
|
|
|
|
return { |
|
|
"counts": [num_iou_th, num_recall_th, num_classes, num_max_detections], |
|
|
"recall": recall, |
|
|
"precision": precision, |
|
|
"scores": scores, |
|
|
} |
|
|
|
|
|
|
|
|
def _compute_stats_single_threshold( |
|
|
tp: np.ndarray, |
|
|
fp: np.ndarray, |
|
|
dt_scores_sorted: np.ndarray, |
|
|
recall_thresholds: np.ndarray | Sequence[float], |
|
|
num_gt: int, |
|
|
) -> tuple[float, np.ndarray, np.ndarray]: |
|
|
""" |
|
|
Compute recall value, precision curve and scores thresholds |
|
|
Adapted from https://github.com/cocodataset/cocoapi/blob/master/PythonAPI/pycocotools/cocoeval.py |
|
|
|
|
|
Args: |
|
|
tp (np.ndarray): cumsum over true positives [R], R is the number of detections |
|
|
fp (np.ndarray): cumsum over false positives [R], R is the number of detections |
|
|
dt_scores_sorted (np.ndarray): sorted (descending) scores [R], R is the number of detections |
|
|
recall_thresholds (Sequence[float]): recall thresholds which should be evaluated |
|
|
num_gt (int): number of ground truth bounding boxes (excluding boxes which are ignored) |
|
|
|
|
|
Returns: |
|
|
- float, overall recall for given IoU value |
|
|
- np.ndarray, precision values at defined recall values |
|
|
[RTH], where RTH is the number of recall thresholds |
|
|
- np.ndarray, prediction scores corresponding to recall values |
|
|
[RTH], where RTH is the number of recall thresholds |
|
|
""" |
|
|
num_recall_th = len(recall_thresholds) |
|
|
|
|
|
rc = tp / num_gt |
|
|
|
|
|
pr = tp / (fp + tp + np.spacing(1)) |
|
|
|
|
|
if len(tp): |
|
|
recall = rc[-1] |
|
|
else: |
|
|
|
|
|
recall = 0 |
|
|
|
|
|
|
|
|
precision = np.zeros((num_recall_th,)) |
|
|
|
|
|
th_scores = np.zeros((num_recall_th,)) |
|
|
|
|
|
|
|
|
pr = pr.tolist() |
|
|
precision = precision.tolist() |
|
|
|
|
|
|
|
|
for i in range(len(tp) - 1, 0, -1): |
|
|
if pr[i] > pr[i - 1]: |
|
|
pr[i - 1] = pr[i] |
|
|
|
|
|
|
|
|
inds = np.searchsorted(rc, recall_thresholds, side="left") |
|
|
try: |
|
|
for save_idx, array_index in enumerate(inds): |
|
|
precision[save_idx] = pr[array_index] |
|
|
th_scores[save_idx] = dt_scores_sorted[array_index] |
|
|
except BaseException: |
|
|
pass |
|
|
|
|
|
return recall, np.array(precision), np.array(th_scores) |
|
|
|