|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
from __future__ import annotations |
|
|
|
|
|
from functools import partial |
|
|
from typing import Any, Callable |
|
|
|
|
|
import torch |
|
|
|
|
|
from monai.networks.utils import replace_modules_temp |
|
|
from monai.utils.module import optional_import |
|
|
from monai.visualize.class_activation_maps import ModelWithHooks |
|
|
|
|
|
trange, has_trange = optional_import("tqdm", name="trange") |
|
|
|
|
|
__all__ = ["VanillaGrad", "SmoothGrad", "GuidedBackpropGrad", "GuidedBackpropSmoothGrad"] |
|
|
|
|
|
|
|
|
class _AutoGradReLU(torch.autograd.Function): |
|
|
|
|
|
@staticmethod |
|
|
def forward(ctx, x): |
|
|
pos_mask = (x > 0).type_as(x) |
|
|
output = torch.mul(x, pos_mask) |
|
|
ctx.save_for_backward(x, output) |
|
|
return output |
|
|
|
|
|
@staticmethod |
|
|
def backward(ctx, grad_output): |
|
|
x, _ = ctx.saved_tensors |
|
|
pos_mask_1 = (x > 0).type_as(grad_output) |
|
|
pos_mask_2 = (grad_output > 0).type_as(grad_output) |
|
|
y = torch.mul(grad_output, pos_mask_1) |
|
|
grad_input = torch.mul(y, pos_mask_2) |
|
|
return grad_input |
|
|
|
|
|
|
|
|
class _GradReLU(torch.nn.Module): |
|
|
""" |
|
|
A customized ReLU with the backward pass imputed for guided backpropagation (https://arxiv.org/abs/1412.6806). |
|
|
""" |
|
|
|
|
|
def forward(self, x: torch.Tensor) -> torch.Tensor: |
|
|
out: torch.Tensor = _AutoGradReLU.apply(x) |
|
|
return out |
|
|
|
|
|
|
|
|
class VanillaGrad: |
|
|
""" |
|
|
Given an input image ``x``, calling this class will perform the forward pass, then set to zero |
|
|
all activations except one (defined by ``index``) and propagate back to the image to achieve a gradient-based |
|
|
saliency map. |
|
|
|
|
|
If ``index`` is None, argmax of the output logits will be used. |
|
|
|
|
|
See also: |
|
|
|
|
|
- Simonyan et al. Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps |
|
|
(https://arxiv.org/abs/1312.6034) |
|
|
""" |
|
|
|
|
|
def __init__(self, model: torch.nn.Module) -> None: |
|
|
if not isinstance(model, ModelWithHooks): |
|
|
self._model = ModelWithHooks(model, target_layer_names=(), register_backward=True) |
|
|
else: |
|
|
self._model = model |
|
|
|
|
|
@property |
|
|
def model(self): |
|
|
return self._model.model |
|
|
|
|
|
@model.setter |
|
|
def model(self, m): |
|
|
if not isinstance(m, ModelWithHooks): |
|
|
self._model.model = m |
|
|
else: |
|
|
self._model = m |
|
|
|
|
|
def get_grad( |
|
|
self, x: torch.Tensor, index: torch.Tensor | int | None, retain_graph: bool = True, **kwargs: Any |
|
|
) -> torch.Tensor: |
|
|
if x.shape[0] != 1: |
|
|
raise ValueError("expect batch size of 1") |
|
|
x.requires_grad = True |
|
|
|
|
|
self._model(x, class_idx=index, retain_graph=retain_graph, **kwargs) |
|
|
grad: torch.Tensor = x.grad.detach() |
|
|
return grad |
|
|
|
|
|
def __call__(self, x: torch.Tensor, index: torch.Tensor | int | None = None, **kwargs: Any) -> torch.Tensor: |
|
|
return self.get_grad(x, index, **kwargs) |
|
|
|
|
|
|
|
|
class SmoothGrad(VanillaGrad): |
|
|
""" |
|
|
Compute averaged sensitivity map based on ``n_samples`` (Gaussian additive) of noisy versions |
|
|
of the input image ``x``. |
|
|
|
|
|
See also: |
|
|
|
|
|
- Smilkov et al. SmoothGrad: removing noise by adding noise https://arxiv.org/abs/1706.03825 |
|
|
""" |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
model: torch.nn.Module, |
|
|
stdev_spread: float = 0.15, |
|
|
n_samples: int = 25, |
|
|
magnitude: bool = True, |
|
|
verbose: bool = True, |
|
|
) -> None: |
|
|
super().__init__(model) |
|
|
self.stdev_spread = stdev_spread |
|
|
self.n_samples = n_samples |
|
|
self.magnitude = magnitude |
|
|
self.range: Callable |
|
|
if verbose and has_trange: |
|
|
self.range = partial(trange, desc=f"Computing {self.__class__.__name__}") |
|
|
else: |
|
|
self.range = range |
|
|
|
|
|
def __call__(self, x: torch.Tensor, index: torch.Tensor | int | None = None, **kwargs: Any) -> torch.Tensor: |
|
|
stdev = (self.stdev_spread * (x.max() - x.min())).item() |
|
|
total_gradients = torch.zeros_like(x) |
|
|
for _ in self.range(self.n_samples): |
|
|
|
|
|
noise = torch.normal(0, stdev, size=x.shape, dtype=torch.float32, device=x.device) |
|
|
x_plus_noise = x + noise |
|
|
x_plus_noise = x_plus_noise.detach() |
|
|
|
|
|
|
|
|
grad = self.get_grad(x_plus_noise, index, **kwargs) |
|
|
total_gradients += (grad * grad) if self.magnitude else grad |
|
|
|
|
|
|
|
|
if self.magnitude: |
|
|
total_gradients = total_gradients**0.5 |
|
|
|
|
|
return total_gradients / self.n_samples |
|
|
|
|
|
|
|
|
class GuidedBackpropGrad(VanillaGrad): |
|
|
""" |
|
|
Based on Springenberg and Dosovitskiy et al. https://arxiv.org/abs/1412.6806, |
|
|
compute gradient-based saliency maps by backpropagating positive gradients and inputs (see ``_AutoGradReLU``). |
|
|
|
|
|
See also: |
|
|
|
|
|
- Springenberg and Dosovitskiy et al. Striving for Simplicity: The All Convolutional Net |
|
|
(https://arxiv.org/abs/1412.6806) |
|
|
""" |
|
|
|
|
|
def __call__(self, x: torch.Tensor, index: torch.Tensor | int | None = None, **kwargs: Any) -> torch.Tensor: |
|
|
with replace_modules_temp(self.model, "relu", _GradReLU(), strict_match=False): |
|
|
return super().__call__(x, index, **kwargs) |
|
|
|
|
|
|
|
|
class GuidedBackpropSmoothGrad(SmoothGrad): |
|
|
""" |
|
|
Compute gradient-based saliency maps based on both ``GuidedBackpropGrad`` and ``SmoothGrad``. |
|
|
""" |
|
|
|
|
|
def __call__(self, x: torch.Tensor, index: torch.Tensor | int | None = None, **kwargs: Any) -> torch.Tensor: |
|
|
with replace_modules_temp(self.model, "relu", _GradReLU(), strict_match=False): |
|
|
return super().__call__(x, index, **kwargs) |
|
|
|