|
|
from typing import Callable |
|
|
|
|
|
import torch |
|
|
from .ddp_allgather import AllGatherGrad |
|
|
from .tensor_utilities import sum_tensor |
|
|
from torch import nn |
|
|
|
|
|
|
|
|
class SoftDiceLoss(nn.Module): |
|
|
def __init__(self, apply_nonlin: Callable = None, batch_dice: bool = False, do_bg: bool = True, smooth: float = 1., |
|
|
ddp: bool = True, clip_tp: float = None): |
|
|
""" |
|
|
""" |
|
|
super(SoftDiceLoss, self).__init__() |
|
|
|
|
|
self.do_bg = do_bg |
|
|
self.batch_dice = batch_dice |
|
|
self.apply_nonlin = apply_nonlin |
|
|
self.smooth = smooth |
|
|
self.clip_tp = clip_tp |
|
|
self.ddp = ddp |
|
|
|
|
|
def forward(self, x, y, loss_mask=None): |
|
|
shp_x = x.shape |
|
|
|
|
|
if self.batch_dice: |
|
|
axes = [0] + list(range(2, len(shp_x))) |
|
|
else: |
|
|
axes = list(range(2, len(shp_x))) |
|
|
|
|
|
if self.apply_nonlin is not None: |
|
|
x = self.apply_nonlin(x) |
|
|
|
|
|
tp, fp, fn, _ = get_tp_fp_fn_tn(x, y, axes, loss_mask, False) |
|
|
|
|
|
if self.ddp and self.batch_dice: |
|
|
tp = AllGatherGrad.apply(tp).sum(0) |
|
|
fp = AllGatherGrad.apply(fp).sum(0) |
|
|
fn = AllGatherGrad.apply(fn).sum(0) |
|
|
|
|
|
if self.clip_tp is not None: |
|
|
tp = torch.clip(tp, min=self.clip_tp , max=None) |
|
|
|
|
|
nominator = 2 * tp |
|
|
denominator = 2 * tp + fp + fn |
|
|
|
|
|
dc = (nominator + self.smooth) / (torch.clip(denominator + self.smooth, 1e-8)) |
|
|
|
|
|
if not self.do_bg: |
|
|
if self.batch_dice: |
|
|
dc = dc[1:] |
|
|
else: |
|
|
dc = dc[:, 1:] |
|
|
dc = dc.mean() |
|
|
|
|
|
return -dc |
|
|
|
|
|
class MemoryEfficientSoftDiceLoss(nn.Module): |
|
|
def __init__(self, apply_nonlin: Callable = None, batch_dice: bool = False, do_bg: bool = True, smooth: float = 1., |
|
|
ddp: bool = True): |
|
|
""" |
|
|
saves 1.6 GB on Dataset017 3d_lowres |
|
|
""" |
|
|
super(MemoryEfficientSoftDiceLoss, self).__init__() |
|
|
|
|
|
self.do_bg = do_bg |
|
|
self.batch_dice = batch_dice |
|
|
self.apply_nonlin = apply_nonlin |
|
|
self.smooth = smooth |
|
|
self.ddp = ddp |
|
|
|
|
|
def forward(self, x, y, loss_mask=None): |
|
|
shp_x, shp_y = x.shape, y.shape |
|
|
|
|
|
if self.apply_nonlin is not None: |
|
|
x = self.apply_nonlin(x) |
|
|
|
|
|
if not self.do_bg: |
|
|
x = x[:, 1:] |
|
|
|
|
|
|
|
|
axes = list(range(2, len(shp_x))) |
|
|
|
|
|
with torch.no_grad(): |
|
|
if len(shp_x) != len(shp_y): |
|
|
y = y.view((shp_y[0], 1, *shp_y[1:])) |
|
|
|
|
|
if all([i == j for i, j in zip(shp_x, shp_y)]): |
|
|
|
|
|
y_onehot = y |
|
|
else: |
|
|
gt = y.long() |
|
|
y_onehot = torch.zeros(shp_x, device=x.device, dtype=torch.bool) |
|
|
y_onehot.scatter_(1, gt, 1) |
|
|
|
|
|
if not self.do_bg: |
|
|
y_onehot = y_onehot[:, 1:] |
|
|
sum_gt = y_onehot.sum(axes) if loss_mask is None else (y_onehot * loss_mask).sum(axes) |
|
|
|
|
|
intersect = (x * y_onehot).sum(axes) if loss_mask is None else (x * y_onehot * loss_mask).sum(axes) |
|
|
sum_pred = x.sum(axes) if loss_mask is None else (x * loss_mask).sum(axes) |
|
|
|
|
|
if self.ddp and self.batch_dice: |
|
|
intersect = AllGatherGrad.apply(intersect).sum(0) |
|
|
sum_pred = AllGatherGrad.apply(sum_pred).sum(0) |
|
|
sum_gt = AllGatherGrad.apply(sum_gt).sum(0) |
|
|
|
|
|
if self.batch_dice: |
|
|
intersect = intersect.sum(0) |
|
|
sum_pred = sum_pred.sum(0) |
|
|
sum_gt = sum_gt.sum(0) |
|
|
|
|
|
dc = (2 * intersect + self.smooth) / (torch.clip(sum_gt + sum_pred + self.smooth, 1e-8)) |
|
|
|
|
|
dc = dc.mean() |
|
|
return -dc |
|
|
|
|
|
def get_tp_fp_fn_tn(net_output, gt, axes=None, mask=None, square=False): |
|
|
""" |
|
|
net_output must be (b, c, x, y(, z))) |
|
|
gt must be a label map (shape (b, 1, x, y(, z)) OR shape (b, x, y(, z))) or one hot encoding (b, c, x, y(, z)) |
|
|
if mask is provided it must have shape (b, 1, x, y(, z))) |
|
|
:param net_output: |
|
|
:param gt: |
|
|
:param axes: can be (, ) = no summation |
|
|
:param mask: mask must be 1 for valid pixels and 0 for invalid pixels |
|
|
:param square: if True then fp, tp and fn will be squared before summation |
|
|
:return: |
|
|
""" |
|
|
if axes is None: |
|
|
axes = tuple(range(2, len(net_output.size()))) |
|
|
|
|
|
shp_x = net_output.shape |
|
|
shp_y = gt.shape |
|
|
|
|
|
with torch.no_grad(): |
|
|
if len(shp_x) != len(shp_y): |
|
|
gt = gt.view((shp_y[0], 1, *shp_y[1:])) |
|
|
|
|
|
if all([i == j for i, j in zip(net_output.shape, gt.shape)]): |
|
|
|
|
|
y_onehot = gt |
|
|
else: |
|
|
gt = gt.long() |
|
|
y_onehot = torch.zeros(shp_x, device=net_output.device) |
|
|
y_onehot.scatter_(1, gt, 1) |
|
|
|
|
|
tp = net_output * y_onehot |
|
|
fp = net_output * (1 - y_onehot) |
|
|
fn = (1 - net_output) * y_onehot |
|
|
tn = (1 - net_output) * (1 - y_onehot) |
|
|
|
|
|
if mask is not None: |
|
|
with torch.no_grad(): |
|
|
mask_here = torch.tile(mask, (1, tp.shape[1], *[1 for i in range(2, len(tp.shape))])) |
|
|
tp *= mask_here |
|
|
fp *= mask_here |
|
|
fn *= mask_here |
|
|
tn *= mask_here |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
if square: |
|
|
tp = tp ** 2 |
|
|
fp = fp ** 2 |
|
|
fn = fn ** 2 |
|
|
tn = tn ** 2 |
|
|
|
|
|
if len(axes) > 0: |
|
|
tp = sum_tensor(tp, axes, keepdim=False) |
|
|
fp = sum_tensor(fp, axes, keepdim=False) |
|
|
fn = sum_tensor(fn, axes, keepdim=False) |
|
|
tn = sum_tensor(tn, axes, keepdim=False) |
|
|
|
|
|
return tp, fp, fn, tn |
|
|
|
|
|
|
|
|
if __name__ == '__main__': |
|
|
from nnunetv2.utilities.helpers import softmax_helper_dim1 |
|
|
pred = torch.rand((2, 3, 32, 32, 32)) |
|
|
ref = torch.randint(0, 3, (2, 32, 32, 32)) |
|
|
|
|
|
dl_old = SoftDiceLoss(apply_nonlin=softmax_helper_dim1, batch_dice=True, do_bg=False, smooth=0, ddp=False) |
|
|
dl_new = MemoryEfficientSoftDiceLoss(apply_nonlin=softmax_helper_dim1, batch_dice=True, do_bg=False, smooth=0, ddp=False) |
|
|
res_old = dl_old(pred, ref) |
|
|
res_new = dl_new(pred, ref) |
|
|
print(res_old, res_new) |
|
|
|