|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
""" |
|
|
A collection of dictionary-based wrappers around the "vanilla" transforms for model output tensors |
|
|
defined in :py:class:`monai.transforms.utility.array`. |
|
|
|
|
|
Class names are ended with 'd' to denote dictionary-based transforms. |
|
|
""" |
|
|
|
|
|
from __future__ import annotations |
|
|
|
|
|
import warnings |
|
|
from collections.abc import Callable, Hashable, Iterable, Mapping, Sequence |
|
|
from copy import deepcopy |
|
|
from typing import Any |
|
|
|
|
|
import numpy as np |
|
|
import torch |
|
|
|
|
|
from monai import config |
|
|
from monai.config.type_definitions import KeysCollection, NdarrayOrTensor, PathLike |
|
|
from monai.data.csv_saver import CSVSaver |
|
|
from monai.data.meta_tensor import MetaTensor |
|
|
from monai.transforms.inverse import InvertibleTransform |
|
|
from monai.transforms.post.array import ( |
|
|
Activations, |
|
|
AsDiscrete, |
|
|
DistanceTransformEDT, |
|
|
FillHoles, |
|
|
KeepLargestConnectedComponent, |
|
|
LabelFilter, |
|
|
LabelToContour, |
|
|
MeanEnsemble, |
|
|
ProbNMS, |
|
|
RemoveSmallObjects, |
|
|
SobelGradients, |
|
|
VoteEnsemble, |
|
|
) |
|
|
from monai.transforms.transform import MapTransform |
|
|
from monai.transforms.utility.array import ToTensor |
|
|
from monai.transforms.utils import allow_missing_keys_mode, convert_applied_interp_mode |
|
|
from monai.utils import PostFix, convert_to_tensor, ensure_tuple, ensure_tuple_rep |
|
|
|
|
|
__all__ = [ |
|
|
"ActivationsD", |
|
|
"ActivationsDict", |
|
|
"Activationsd", |
|
|
"AsDiscreteD", |
|
|
"AsDiscreteDict", |
|
|
"AsDiscreted", |
|
|
"Ensembled", |
|
|
"EnsembleD", |
|
|
"EnsembleDict", |
|
|
"FillHolesD", |
|
|
"FillHolesDict", |
|
|
"FillHolesd", |
|
|
"InvertD", |
|
|
"InvertDict", |
|
|
"Invertd", |
|
|
"KeepLargestConnectedComponentD", |
|
|
"KeepLargestConnectedComponentDict", |
|
|
"KeepLargestConnectedComponentd", |
|
|
"RemoveSmallObjectsD", |
|
|
"RemoveSmallObjectsDict", |
|
|
"RemoveSmallObjectsd", |
|
|
"LabelFilterD", |
|
|
"LabelFilterDict", |
|
|
"LabelFilterd", |
|
|
"LabelToContourD", |
|
|
"LabelToContourDict", |
|
|
"LabelToContourd", |
|
|
"MeanEnsembleD", |
|
|
"MeanEnsembleDict", |
|
|
"MeanEnsembled", |
|
|
"ProbNMSD", |
|
|
"ProbNMSDict", |
|
|
"ProbNMSd", |
|
|
"SaveClassificationD", |
|
|
"SaveClassificationDict", |
|
|
"SaveClassificationd", |
|
|
"SobelGradientsD", |
|
|
"SobelGradientsDict", |
|
|
"SobelGradientsd", |
|
|
"VoteEnsembleD", |
|
|
"VoteEnsembleDict", |
|
|
"VoteEnsembled", |
|
|
"DistanceTransformEDTd", |
|
|
"DistanceTransformEDTD", |
|
|
"DistanceTransformEDTDict", |
|
|
] |
|
|
|
|
|
DEFAULT_POST_FIX = PostFix.meta() |
|
|
|
|
|
|
|
|
class Activationsd(MapTransform): |
|
|
""" |
|
|
Dictionary-based wrapper of :py:class:`monai.transforms.AddActivations`. |
|
|
Add activation layers to the input data specified by `keys`. |
|
|
""" |
|
|
|
|
|
backend = Activations.backend |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
keys: KeysCollection, |
|
|
sigmoid: Sequence[bool] | bool = False, |
|
|
softmax: Sequence[bool] | bool = False, |
|
|
other: Sequence[Callable] | Callable | None = None, |
|
|
allow_missing_keys: bool = False, |
|
|
**kwargs, |
|
|
) -> None: |
|
|
""" |
|
|
Args: |
|
|
keys: keys of the corresponding items to model output and label. |
|
|
See also: :py:class:`monai.transforms.compose.MapTransform` |
|
|
sigmoid: whether to execute sigmoid function on model output before transform. |
|
|
it also can be a sequence of bool, each element corresponds to a key in ``keys``. |
|
|
softmax: whether to execute softmax function on model output before transform. |
|
|
it also can be a sequence of bool, each element corresponds to a key in ``keys``. |
|
|
other: callable function to execute other activation layers, |
|
|
for example: `other = torch.tanh`. it also can be a sequence of Callable, each |
|
|
element corresponds to a key in ``keys``. |
|
|
allow_missing_keys: don't raise exception if key is missing. |
|
|
kwargs: additional parameters to `torch.softmax` (used when ``softmax=True``). |
|
|
Defaults to ``dim=0``, unrecognized parameters will be ignored. |
|
|
|
|
|
""" |
|
|
super().__init__(keys, allow_missing_keys) |
|
|
self.sigmoid = ensure_tuple_rep(sigmoid, len(self.keys)) |
|
|
self.softmax = ensure_tuple_rep(softmax, len(self.keys)) |
|
|
self.other = ensure_tuple_rep(other, len(self.keys)) |
|
|
self.converter = Activations() |
|
|
self.converter.kwargs = kwargs |
|
|
|
|
|
def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
|
|
d = dict(data) |
|
|
for key, sigmoid, softmax, other in self.key_iterator(d, self.sigmoid, self.softmax, self.other): |
|
|
d[key] = self.converter(d[key], sigmoid, softmax, other) |
|
|
return d |
|
|
|
|
|
|
|
|
class AsDiscreted(MapTransform): |
|
|
""" |
|
|
Dictionary-based wrapper of :py:class:`monai.transforms.AsDiscrete`. |
|
|
""" |
|
|
|
|
|
backend = AsDiscrete.backend |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
keys: KeysCollection, |
|
|
argmax: Sequence[bool] | bool = False, |
|
|
to_onehot: Sequence[int | None] | int | None = None, |
|
|
threshold: Sequence[float | None] | float | None = None, |
|
|
rounding: Sequence[str | None] | str | None = None, |
|
|
allow_missing_keys: bool = False, |
|
|
**kwargs, |
|
|
) -> None: |
|
|
""" |
|
|
Args: |
|
|
keys: keys of the corresponding items to model output and label. |
|
|
See also: :py:class:`monai.transforms.compose.MapTransform` |
|
|
argmax: whether to execute argmax function on input data before transform. |
|
|
it also can be a sequence of bool, each element corresponds to a key in ``keys``. |
|
|
to_onehot: if not None, convert input data into the one-hot format with specified number of classes. |
|
|
defaults to ``None``. it also can be a sequence, each element corresponds to a key in ``keys``. |
|
|
threshold: if not None, threshold the float values to int number 0 or 1 with specified threshold value. |
|
|
defaults to ``None``. it also can be a sequence, each element corresponds to a key in ``keys``. |
|
|
rounding: if not None, round the data according to the specified option, |
|
|
available options: ["torchrounding"]. it also can be a sequence of str or None, |
|
|
each element corresponds to a key in ``keys``. |
|
|
allow_missing_keys: don't raise exception if key is missing. |
|
|
kwargs: additional parameters to ``AsDiscrete``. |
|
|
``dim``, ``keepdim``, ``dtype`` are supported, unrecognized parameters will be ignored. |
|
|
These default to ``0``, ``True``, ``torch.float`` respectively. |
|
|
|
|
|
""" |
|
|
super().__init__(keys, allow_missing_keys) |
|
|
self.argmax = ensure_tuple_rep(argmax, len(self.keys)) |
|
|
self.to_onehot = [] |
|
|
for flag in ensure_tuple_rep(to_onehot, len(self.keys)): |
|
|
if isinstance(flag, bool): |
|
|
raise ValueError("`to_onehot=True/False` is deprecated, please use `to_onehot=num_classes` instead.") |
|
|
self.to_onehot.append(flag) |
|
|
|
|
|
self.threshold = [] |
|
|
for flag in ensure_tuple_rep(threshold, len(self.keys)): |
|
|
if isinstance(flag, bool): |
|
|
raise ValueError("`threshold_values=True/False` is deprecated, please use `threshold=value` instead.") |
|
|
self.threshold.append(flag) |
|
|
|
|
|
self.rounding = ensure_tuple_rep(rounding, len(self.keys)) |
|
|
self.converter = AsDiscrete() |
|
|
self.converter.kwargs = kwargs |
|
|
|
|
|
def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
|
|
d = dict(data) |
|
|
for key, argmax, to_onehot, threshold, rounding in self.key_iterator( |
|
|
d, self.argmax, self.to_onehot, self.threshold, self.rounding |
|
|
): |
|
|
d[key] = self.converter(d[key], argmax, to_onehot, threshold, rounding) |
|
|
return d |
|
|
|
|
|
|
|
|
class KeepLargestConnectedComponentd(MapTransform): |
|
|
""" |
|
|
Dictionary-based wrapper of :py:class:`monai.transforms.KeepLargestConnectedComponent`. |
|
|
""" |
|
|
|
|
|
backend = KeepLargestConnectedComponent.backend |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
keys: KeysCollection, |
|
|
applied_labels: Sequence[int] | int | None = None, |
|
|
is_onehot: bool | None = None, |
|
|
independent: bool = True, |
|
|
connectivity: int | None = None, |
|
|
num_components: int = 1, |
|
|
allow_missing_keys: bool = False, |
|
|
) -> None: |
|
|
""" |
|
|
Args: |
|
|
keys: keys of the corresponding items to be transformed. |
|
|
See also: :py:class:`monai.transforms.compose.MapTransform` |
|
|
applied_labels: Labels for applying the connected component analysis on. |
|
|
If given, voxels whose value is in this list will be analyzed. |
|
|
If `None`, all non-zero values will be analyzed. |
|
|
is_onehot: if `True`, treat the input data as OneHot format data, otherwise, not OneHot format data. |
|
|
default to None, which treats multi-channel data as OneHot and single channel data as not OneHot. |
|
|
independent: whether to treat ``applied_labels`` as a union of foreground labels. |
|
|
If ``True``, the connected component analysis will be performed on each foreground label independently |
|
|
and return the intersection of the largest components. |
|
|
If ``False``, the analysis will be performed on the union of foreground labels. |
|
|
default is `True`. |
|
|
connectivity: Maximum number of orthogonal hops to consider a pixel/voxel as a neighbor. |
|
|
Accepted values are ranging from 1 to input.ndim. If ``None``, a full |
|
|
connectivity of ``input.ndim`` is used. for more details: |
|
|
https://scikit-image.org/docs/dev/api/skimage.measure.html#skimage.measure.label. |
|
|
num_components: The number of largest components to preserve. |
|
|
allow_missing_keys: don't raise exception if key is missing. |
|
|
|
|
|
""" |
|
|
super().__init__(keys, allow_missing_keys) |
|
|
self.converter = KeepLargestConnectedComponent( |
|
|
applied_labels=applied_labels, |
|
|
is_onehot=is_onehot, |
|
|
independent=independent, |
|
|
connectivity=connectivity, |
|
|
num_components=num_components, |
|
|
) |
|
|
|
|
|
def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
|
|
d = dict(data) |
|
|
for key in self.key_iterator(d): |
|
|
d[key] = self.converter(d[key]) |
|
|
return d |
|
|
|
|
|
|
|
|
class RemoveSmallObjectsd(MapTransform): |
|
|
""" |
|
|
Dictionary-based wrapper of :py:class:`monai.transforms.RemoveSmallObjectsd`. |
|
|
|
|
|
Args: |
|
|
min_size: objects smaller than this size (in number of voxels; or surface area/volume value |
|
|
in whatever units your image is if by_measure is True) are removed. |
|
|
connectivity: Maximum number of orthogonal hops to consider a pixel/voxel as a neighbor. |
|
|
Accepted values are ranging from 1 to input.ndim. If ``None``, a full |
|
|
connectivity of ``input.ndim`` is used. For more details refer to linked scikit-image |
|
|
documentation. |
|
|
independent_channels: Whether or not to consider channels as independent. If true, then |
|
|
conjoining islands from different labels will be removed if they are below the threshold. |
|
|
If false, the overall size islands made from all non-background voxels will be used. |
|
|
by_measure: Whether the specified min_size is in number of voxels. if this is True then min_size |
|
|
represents a surface area or volume value of whatever units your image is in (mm^3, cm^2, etc.) |
|
|
default is False. e.g. if min_size is 3, by_measure is True and the units of your data is mm, |
|
|
objects smaller than 3mm^3 are removed. |
|
|
pixdim: the pixdim of the input image. if a single number, this is used for all axes. |
|
|
If a sequence of numbers, the length of the sequence must be equal to the image dimensions. |
|
|
""" |
|
|
|
|
|
backend = RemoveSmallObjects.backend |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
keys: KeysCollection, |
|
|
min_size: int = 64, |
|
|
connectivity: int = 1, |
|
|
independent_channels: bool = True, |
|
|
by_measure: bool = False, |
|
|
pixdim: Sequence[float] | float | np.ndarray | None = None, |
|
|
allow_missing_keys: bool = False, |
|
|
) -> None: |
|
|
super().__init__(keys, allow_missing_keys) |
|
|
self.converter = RemoveSmallObjects(min_size, connectivity, independent_channels, by_measure, pixdim) |
|
|
|
|
|
def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
|
|
d = dict(data) |
|
|
for key in self.key_iterator(d): |
|
|
d[key] = self.converter(d[key]) |
|
|
return d |
|
|
|
|
|
|
|
|
class LabelFilterd(MapTransform): |
|
|
""" |
|
|
Dictionary-based wrapper of :py:class:`monai.transforms.LabelFilter`. |
|
|
""" |
|
|
|
|
|
backend = LabelFilter.backend |
|
|
|
|
|
def __init__( |
|
|
self, keys: KeysCollection, applied_labels: Sequence[int] | int, allow_missing_keys: bool = False |
|
|
) -> None: |
|
|
""" |
|
|
Args: |
|
|
keys: keys of the corresponding items to be transformed. |
|
|
See also: :py:class:`monai.transforms.compose.MapTransform` |
|
|
applied_labels: Label(s) to filter on. |
|
|
allow_missing_keys: don't raise exception if key is missing. |
|
|
|
|
|
""" |
|
|
super().__init__(keys, allow_missing_keys) |
|
|
self.converter = LabelFilter(applied_labels) |
|
|
|
|
|
def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
|
|
d = dict(data) |
|
|
for key in self.key_iterator(d): |
|
|
d[key] = self.converter(d[key]) |
|
|
return d |
|
|
|
|
|
|
|
|
class FillHolesd(MapTransform): |
|
|
""" |
|
|
Dictionary-based wrapper of :py:class:`monai.transforms.FillHoles`. |
|
|
""" |
|
|
|
|
|
backend = FillHoles.backend |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
keys: KeysCollection, |
|
|
applied_labels: Iterable[int] | int | None = None, |
|
|
connectivity: int | None = None, |
|
|
allow_missing_keys: bool = False, |
|
|
) -> None: |
|
|
""" |
|
|
Initialize the connectivity and limit the labels for which holes are filled. |
|
|
|
|
|
Args: |
|
|
keys: keys of the corresponding items to be transformed. |
|
|
See also: :py:class:`monai.transforms.compose.MapTransform` |
|
|
applied_labels (Optional[Union[Iterable[int], int]], optional): Labels for which to fill holes. Defaults to None, |
|
|
that is filling holes for all labels. |
|
|
connectivity (int, optional): Maximum number of orthogonal hops to consider a pixel/voxel as a neighbor. |
|
|
Accepted values are ranging from 1 to input.ndim. Defaults to a full |
|
|
connectivity of ``input.ndim``. |
|
|
allow_missing_keys: don't raise exception if key is missing. |
|
|
""" |
|
|
super().__init__(keys, allow_missing_keys) |
|
|
self.converter = FillHoles(applied_labels=applied_labels, connectivity=connectivity) |
|
|
|
|
|
def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
|
|
d = dict(data) |
|
|
for key in self.key_iterator(d): |
|
|
d[key] = self.converter(d[key]) |
|
|
return d |
|
|
|
|
|
|
|
|
class LabelToContourd(MapTransform): |
|
|
""" |
|
|
Dictionary-based wrapper of :py:class:`monai.transforms.LabelToContour`. |
|
|
""" |
|
|
|
|
|
backend = LabelToContour.backend |
|
|
|
|
|
def __init__(self, keys: KeysCollection, kernel_type: str = "Laplace", allow_missing_keys: bool = False) -> None: |
|
|
""" |
|
|
Args: |
|
|
keys: keys of the corresponding items to be transformed. |
|
|
See also: :py:class:`monai.transforms.compose.MapTransform` |
|
|
kernel_type: the method applied to do edge detection, default is "Laplace". |
|
|
allow_missing_keys: don't raise exception if key is missing. |
|
|
|
|
|
""" |
|
|
super().__init__(keys, allow_missing_keys) |
|
|
self.converter = LabelToContour(kernel_type=kernel_type) |
|
|
|
|
|
def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
|
|
d = dict(data) |
|
|
for key in self.key_iterator(d): |
|
|
d[key] = self.converter(d[key]) |
|
|
return d |
|
|
|
|
|
|
|
|
class Ensembled(MapTransform): |
|
|
""" |
|
|
Base class of dictionary-based ensemble transforms. |
|
|
|
|
|
""" |
|
|
|
|
|
backend = list(set(VoteEnsemble.backend) & set(MeanEnsemble.backend)) |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
keys: KeysCollection, |
|
|
ensemble: Callable[[Sequence[NdarrayOrTensor] | NdarrayOrTensor], NdarrayOrTensor], |
|
|
output_key: str | None = None, |
|
|
allow_missing_keys: bool = False, |
|
|
) -> None: |
|
|
""" |
|
|
Args: |
|
|
keys: keys of the corresponding items to be stack and execute ensemble. |
|
|
if only 1 key provided, suppose it's a PyTorch Tensor with data stacked on dimension `E`. |
|
|
output_key: the key to store ensemble result in the dictionary. |
|
|
ensemble: callable method to execute ensemble on specified data. |
|
|
if only 1 key provided in `keys`, `output_key` can be None and use `keys` as default. |
|
|
allow_missing_keys: don't raise exception if key is missing. |
|
|
|
|
|
Raises: |
|
|
TypeError: When ``ensemble`` is not ``callable``. |
|
|
ValueError: When ``len(keys) > 1`` and ``output_key=None``. Incompatible values. |
|
|
|
|
|
""" |
|
|
super().__init__(keys, allow_missing_keys) |
|
|
if not callable(ensemble): |
|
|
raise TypeError(f"ensemble must be callable but is {type(ensemble).__name__}.") |
|
|
self.ensemble = ensemble |
|
|
if len(self.keys) > 1 and output_key is None: |
|
|
raise ValueError("Incompatible values: len(self.keys) > 1 and output_key=None.") |
|
|
self.output_key = output_key if output_key is not None else self.keys[0] |
|
|
|
|
|
def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
|
|
d = dict(data) |
|
|
items: list[NdarrayOrTensor] | NdarrayOrTensor |
|
|
if len(self.keys) == 1 and self.keys[0] in d: |
|
|
items = d[self.keys[0]] |
|
|
else: |
|
|
items = [d[key] for key in self.key_iterator(d)] |
|
|
|
|
|
if len(items) > 0: |
|
|
d[self.output_key] = self.ensemble(items) |
|
|
|
|
|
return d |
|
|
|
|
|
|
|
|
class MeanEnsembled(Ensembled): |
|
|
""" |
|
|
Dictionary-based wrapper of :py:class:`monai.transforms.MeanEnsemble`. |
|
|
""" |
|
|
|
|
|
backend = MeanEnsemble.backend |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
keys: KeysCollection, |
|
|
output_key: str | None = None, |
|
|
weights: Sequence[float] | NdarrayOrTensor | None = None, |
|
|
) -> None: |
|
|
""" |
|
|
Args: |
|
|
keys: keys of the corresponding items to be stack and execute ensemble. |
|
|
if only 1 key provided, suppose it's a PyTorch Tensor with data stacked on dimension `E`. |
|
|
output_key: the key to store ensemble result in the dictionary. |
|
|
if only 1 key provided in `keys`, `output_key` can be None and use `keys` as default. |
|
|
weights: can be a list or tuple of numbers for input data with shape: [E, C, H, W[, D]]. |
|
|
or a Numpy ndarray or a PyTorch Tensor data. |
|
|
the `weights` will be added to input data from highest dimension, for example: |
|
|
1. if the `weights` only has 1 dimension, it will be added to the `E` dimension of input data. |
|
|
2. if the `weights` has 2 dimensions, it will be added to `E` and `C` dimensions. |
|
|
it's a typical practice to add weights for different classes: |
|
|
to ensemble 3 segmentation model outputs, every output has 4 channels(classes), |
|
|
so the input data shape can be: [3, 4, H, W, D]. |
|
|
and add different `weights` for different classes, so the `weights` shape can be: [3, 4]. |
|
|
for example: `weights = [[1, 2, 3, 4], [4, 3, 2, 1], [1, 1, 1, 1]]`. |
|
|
|
|
|
""" |
|
|
ensemble = MeanEnsemble(weights=weights) |
|
|
super().__init__(keys, ensemble, output_key) |
|
|
|
|
|
|
|
|
class VoteEnsembled(Ensembled): |
|
|
""" |
|
|
Dictionary-based wrapper of :py:class:`monai.transforms.VoteEnsemble`. |
|
|
""" |
|
|
|
|
|
backend = VoteEnsemble.backend |
|
|
|
|
|
def __init__(self, keys: KeysCollection, output_key: str | None = None, num_classes: int | None = None) -> None: |
|
|
""" |
|
|
Args: |
|
|
keys: keys of the corresponding items to be stack and execute ensemble. |
|
|
if only 1 key provided, suppose it's a PyTorch Tensor with data stacked on dimension `E`. |
|
|
output_key: the key to store ensemble result in the dictionary. |
|
|
if only 1 key provided in `keys`, `output_key` can be None and use `keys` as default. |
|
|
num_classes: if the input is single channel data instead of One-Hot, we can't get class number |
|
|
from channel, need to explicitly specify the number of classes to vote. |
|
|
|
|
|
""" |
|
|
ensemble = VoteEnsemble(num_classes=num_classes) |
|
|
super().__init__(keys, ensemble, output_key) |
|
|
|
|
|
|
|
|
class ProbNMSd(MapTransform): |
|
|
""" |
|
|
Performs probability based non-maximum suppression (NMS) on the probabilities map via |
|
|
iteratively selecting the coordinate with highest probability and then move it as well |
|
|
as its surrounding values. The remove range is determined by the parameter `box_size`. |
|
|
If multiple coordinates have the same highest probability, only one of them will be |
|
|
selected. |
|
|
|
|
|
Args: |
|
|
spatial_dims: number of spatial dimensions of the input probabilities map. |
|
|
Defaults to 2. |
|
|
sigma: the standard deviation for gaussian filter. |
|
|
It could be a single value, or `spatial_dims` number of values. Defaults to 0.0. |
|
|
prob_threshold: the probability threshold, the function will stop searching if |
|
|
the highest probability is no larger than the threshold. The value should be |
|
|
no less than 0.0. Defaults to 0.5. |
|
|
box_size: the box size (in pixel) to be removed around the pixel with the maximum probability. |
|
|
It can be an integer that defines the size of a square or cube, |
|
|
or a list containing different values for each dimensions. Defaults to 48. |
|
|
|
|
|
Return: |
|
|
a list of selected lists, where inner lists contain probability and coordinates. |
|
|
For example, for 3D input, the inner lists are in the form of [probability, x, y, z]. |
|
|
|
|
|
Raises: |
|
|
ValueError: When ``prob_threshold`` is less than 0.0. |
|
|
ValueError: When ``box_size`` is a list or tuple, and its length is not equal to `spatial_dims`. |
|
|
ValueError: When ``box_size`` has a less than 1 value. |
|
|
|
|
|
""" |
|
|
|
|
|
backend = ProbNMS.backend |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
keys: KeysCollection, |
|
|
spatial_dims: int = 2, |
|
|
sigma: Sequence[float] | float | Sequence[torch.Tensor] | torch.Tensor = 0.0, |
|
|
prob_threshold: float = 0.5, |
|
|
box_size: int | Sequence[int] = 48, |
|
|
allow_missing_keys: bool = False, |
|
|
) -> None: |
|
|
super().__init__(keys, allow_missing_keys) |
|
|
self.prob_nms = ProbNMS( |
|
|
spatial_dims=spatial_dims, sigma=sigma, prob_threshold=prob_threshold, box_size=box_size |
|
|
) |
|
|
|
|
|
def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]): |
|
|
d = dict(data) |
|
|
for key in self.key_iterator(d): |
|
|
d[key] = self.prob_nms(d[key]) |
|
|
return d |
|
|
|
|
|
|
|
|
class Invertd(MapTransform): |
|
|
""" |
|
|
Utility transform to invert the previously applied transforms. |
|
|
|
|
|
Taking the ``transform`` previously applied on ``orig_keys``, this ``Invertd`` will apply the inverse of it |
|
|
to the data stored at ``keys``. |
|
|
|
|
|
``Invertd``'s output will also include a copy of the metadata |
|
|
dictionary (originally from ``orig_meta_keys`` or the metadata of ``orig_keys``), |
|
|
with the relevant fields inverted and stored at ``meta_keys``. |
|
|
|
|
|
A typical usage is to apply the inverse of the preprocessing (``transform=preprocessings``) on |
|
|
input ``orig_keys=image`` to the model predictions ``keys=pred``. |
|
|
|
|
|
A detailed usage example is available in the tutorial: |
|
|
https://github.com/Project-MONAI/tutorials/blob/master/3d_segmentation/torch/unet_inference_dict.py |
|
|
|
|
|
Note: |
|
|
|
|
|
- The output of the inverted data and metadata will be stored at ``keys`` and ``meta_keys`` respectively. |
|
|
- To correctly invert the transforms, the information of the previously applied transforms should be |
|
|
available at ``{orig_keys}_transforms``, and the original metadata at ``orig_meta_keys``. |
|
|
(``meta_key_postfix`` is an optional string to conveniently construct "meta_keys" and/or "orig_meta_keys".) |
|
|
see also: :py:class:`monai.transforms.TraceableTransform`. |
|
|
- The transform will not change the content in ``orig_keys`` and ``orig_meta_key``. |
|
|
These keys are only used to represent the data status of ``key`` before inverting. |
|
|
|
|
|
""" |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
keys: KeysCollection, |
|
|
transform: InvertibleTransform, |
|
|
orig_keys: KeysCollection | None = None, |
|
|
meta_keys: KeysCollection | None = None, |
|
|
orig_meta_keys: KeysCollection | None = None, |
|
|
meta_key_postfix: str = DEFAULT_POST_FIX, |
|
|
nearest_interp: bool | Sequence[bool] = True, |
|
|
to_tensor: bool | Sequence[bool] = True, |
|
|
device: str | torch.device | Sequence[str | torch.device] | None = None, |
|
|
post_func: Callable | Sequence[Callable] | None = None, |
|
|
allow_missing_keys: bool = False, |
|
|
) -> None: |
|
|
""" |
|
|
Args: |
|
|
keys: the key of expected data in the dict, the inverse of ``transforms`` will be applied on it in-place. |
|
|
It also can be a list of keys, will apply the inverse transform respectively. |
|
|
transform: the transform applied to ``orig_key``, its inverse will be applied on ``key``. |
|
|
orig_keys: the key of the original input data in the dict. These keys default to `self.keys` if not set. |
|
|
the transform trace information of ``transforms`` should be stored at ``{orig_keys}_transforms``. |
|
|
It can also be a list of keys, each matches the ``keys``. |
|
|
meta_keys: The key to output the inverted metadata dictionary. |
|
|
The metadata is a dictionary optionally containing: filename, original_shape. |
|
|
It can be a sequence of strings, maps to ``keys``. |
|
|
If None, will try to create a metadata dict with the default key: `{key}_{meta_key_postfix}`. |
|
|
orig_meta_keys: the key of the metadata of original input data. |
|
|
The metadata is a dictionary optionally containing: filename, original_shape. |
|
|
It can be a sequence of strings, maps to the `keys`. |
|
|
If None, will try to create a metadata dict with the default key: `{orig_key}_{meta_key_postfix}`. |
|
|
This metadata dict will also be included in the inverted dict, stored in `meta_keys`. |
|
|
meta_key_postfix: if `orig_meta_keys` is None, use `{orig_key}_{meta_key_postfix}` to fetch the |
|
|
metadata from dict, if `meta_keys` is None, use `{key}_{meta_key_postfix}`. Default: ``"meta_dict"``. |
|
|
nearest_interp: whether to use `nearest` interpolation mode when inverting the spatial transforms, |
|
|
default to `True`. If `False`, use the same interpolation mode as the original transform. |
|
|
It also can be a list of bool, each matches to the `keys` data. |
|
|
to_tensor: whether to convert the inverted data into PyTorch Tensor first, default to `True`. |
|
|
It also can be a list of bool, each matches to the `keys` data. |
|
|
device: if converted to Tensor, move the inverted results to target device before `post_func`, |
|
|
default to None, it also can be a list of string or `torch.device`, each matches to the `keys` data. |
|
|
post_func: post processing for the inverted data, should be a callable function. |
|
|
It also can be a list of callable, each matches to the `keys` data. |
|
|
allow_missing_keys: don't raise exception if key is missing. |
|
|
|
|
|
""" |
|
|
super().__init__(keys, allow_missing_keys) |
|
|
if not isinstance(transform, InvertibleTransform): |
|
|
raise ValueError("transform is not invertible, can't invert transform for the data.") |
|
|
self.transform = transform |
|
|
self.orig_keys = ensure_tuple_rep(orig_keys, len(self.keys)) if orig_keys is not None else self.keys |
|
|
self.meta_keys = ensure_tuple_rep(None, len(self.keys)) if meta_keys is None else ensure_tuple(meta_keys) |
|
|
if len(self.keys) != len(self.meta_keys): |
|
|
raise ValueError("meta_keys should have the same length as keys.") |
|
|
self.orig_meta_keys = ensure_tuple_rep(orig_meta_keys, len(self.keys)) |
|
|
self.meta_key_postfix = ensure_tuple_rep(meta_key_postfix, len(self.keys)) |
|
|
self.nearest_interp = ensure_tuple_rep(nearest_interp, len(self.keys)) |
|
|
self.to_tensor = ensure_tuple_rep(to_tensor, len(self.keys)) |
|
|
self.device = ensure_tuple_rep(device, len(self.keys)) |
|
|
self.post_func = ensure_tuple_rep(post_func, len(self.keys)) |
|
|
self._totensor = ToTensor() |
|
|
|
|
|
def __call__(self, data: Mapping[Hashable, Any]) -> dict[Hashable, Any]: |
|
|
d = dict(data) |
|
|
for ( |
|
|
key, |
|
|
orig_key, |
|
|
meta_key, |
|
|
orig_meta_key, |
|
|
meta_key_postfix, |
|
|
nearest_interp, |
|
|
to_tensor, |
|
|
device, |
|
|
post_func, |
|
|
) in self.key_iterator( |
|
|
d, |
|
|
self.orig_keys, |
|
|
self.meta_keys, |
|
|
self.orig_meta_keys, |
|
|
self.meta_key_postfix, |
|
|
self.nearest_interp, |
|
|
self.to_tensor, |
|
|
self.device, |
|
|
self.post_func, |
|
|
): |
|
|
if isinstance(d[key], MetaTensor): |
|
|
if orig_key not in d: |
|
|
warnings.warn(f"transform info of `{orig_key}` is not available in MetaTensor {key}.") |
|
|
continue |
|
|
else: |
|
|
transform_key = InvertibleTransform.trace_key(orig_key) |
|
|
if transform_key not in d: |
|
|
warnings.warn(f"transform info of `{orig_key}` is not available or no InvertibleTransform applied.") |
|
|
continue |
|
|
|
|
|
orig_meta_key = orig_meta_key or f"{orig_key}_{meta_key_postfix}" |
|
|
if orig_key in d and isinstance(d[orig_key], MetaTensor): |
|
|
transform_info = d[orig_key].applied_operations |
|
|
meta_info = d[orig_key].meta |
|
|
else: |
|
|
transform_info = d[InvertibleTransform.trace_key(orig_key)] |
|
|
meta_info = d.get(orig_meta_key, {}) |
|
|
if nearest_interp: |
|
|
transform_info = convert_applied_interp_mode( |
|
|
trans_info=transform_info, mode="nearest", align_corners=None |
|
|
) |
|
|
|
|
|
inputs = d[key] |
|
|
if isinstance(inputs, torch.Tensor): |
|
|
inputs = inputs.detach() |
|
|
|
|
|
if not isinstance(inputs, MetaTensor): |
|
|
inputs = convert_to_tensor(inputs, track_meta=True) |
|
|
inputs.applied_operations = deepcopy(transform_info) |
|
|
inputs.meta = deepcopy(meta_info) |
|
|
|
|
|
|
|
|
input_dict = {orig_key: inputs} |
|
|
if config.USE_META_DICT: |
|
|
input_dict[InvertibleTransform.trace_key(orig_key)] = transform_info |
|
|
input_dict[PostFix.meta(orig_key)] = meta_info |
|
|
with allow_missing_keys_mode(self.transform): |
|
|
inverted = self.transform.inverse(input_dict) |
|
|
|
|
|
|
|
|
inverted_data = inverted[orig_key] |
|
|
if to_tensor and not isinstance(inverted_data, MetaTensor): |
|
|
inverted_data = self._totensor(inverted_data) |
|
|
if isinstance(inverted_data, np.ndarray) and device is not None and torch.device(device).type != "cpu": |
|
|
raise ValueError(f"Inverted data with type of 'numpy.ndarray' support device='cpu', got {device}.") |
|
|
if isinstance(inverted_data, torch.Tensor): |
|
|
inverted_data = inverted_data.to(device=device) |
|
|
d[key] = post_func(inverted_data) if callable(post_func) else inverted_data |
|
|
|
|
|
if InvertibleTransform.trace_key(orig_key) in d: |
|
|
d[InvertibleTransform.trace_key(orig_key)] = inverted_data.applied_operations |
|
|
|
|
|
if orig_meta_key in d: |
|
|
meta_key = meta_key or f"{key}_{meta_key_postfix}" |
|
|
d[meta_key] = inverted.get(orig_meta_key) |
|
|
return d |
|
|
|
|
|
|
|
|
class SaveClassificationd(MapTransform): |
|
|
""" |
|
|
Save the classification results and metadata into CSV file or other storage. |
|
|
|
|
|
""" |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
keys: KeysCollection, |
|
|
meta_keys: KeysCollection | None = None, |
|
|
meta_key_postfix: str = DEFAULT_POST_FIX, |
|
|
saver: CSVSaver | None = None, |
|
|
output_dir: PathLike = "./", |
|
|
filename: str = "predictions.csv", |
|
|
delimiter: str = ",", |
|
|
overwrite: bool = True, |
|
|
flush: bool = True, |
|
|
allow_missing_keys: bool = False, |
|
|
) -> None: |
|
|
""" |
|
|
Args: |
|
|
keys: keys of the corresponding items to model output, this transform only supports 1 key. |
|
|
See also: :py:class:`monai.transforms.compose.MapTransform` |
|
|
meta_keys: explicitly indicate the key of the corresponding metadata dictionary. |
|
|
for example, for data with key `image`, the metadata by default is in `image_meta_dict`. |
|
|
the metadata is a dictionary object which contains: filename, original_shape, etc. |
|
|
it can be a sequence of string, map to the `keys`. |
|
|
if None, will try to construct meta_keys by `key_{meta_key_postfix}`. |
|
|
will extract the filename of input image to save classification results. |
|
|
meta_key_postfix: `key_{postfix}` was used to store the metadata in `LoadImaged`. |
|
|
so need the key to extract the metadata of input image, like filename, etc. default is `meta_dict`. |
|
|
for example, for data with key `image`, the metadata by default is in `image_meta_dict`. |
|
|
the metadata is a dictionary object which contains: filename, original_shape, etc. |
|
|
this arg only works when `meta_keys=None`. if no corresponding metadata, set to `None`. |
|
|
saver: the saver instance to save classification results, if None, create a CSVSaver internally. |
|
|
the saver must provide `save(data, meta_data)` and `finalize()` APIs. |
|
|
output_dir: if `saver=None`, specify the directory to save the CSV file. |
|
|
filename: if `saver=None`, specify the name of the saved CSV file. |
|
|
delimiter: the delimiter character in the saved file, default to "," as the default output type is `csv`. |
|
|
to be consistent with: https://docs.python.org/3/library/csv.html#csv.Dialect.delimiter. |
|
|
overwrite: if `saver=None`, indicate whether to overwriting existing CSV file content, if True, |
|
|
will clear the file before saving. otherwise, will append new content to the CSV file. |
|
|
flush: if `saver=None`, indicate whether to write the cache data to CSV file immediately |
|
|
in this transform and clear the cache. default to True. |
|
|
If False, may need user to call `saver.finalize()` manually or use `ClassificationSaver` handler. |
|
|
allow_missing_keys: don't raise exception if key is missing. |
|
|
|
|
|
""" |
|
|
super().__init__(keys, allow_missing_keys) |
|
|
if len(self.keys) != 1: |
|
|
raise ValueError("only 1 key is allowed when saving the classification result.") |
|
|
self.saver = saver or CSVSaver( |
|
|
output_dir=output_dir, filename=filename, overwrite=overwrite, flush=flush, delimiter=delimiter |
|
|
) |
|
|
self.flush = flush |
|
|
self.meta_keys = ensure_tuple_rep(meta_keys, len(self.keys)) |
|
|
self.meta_key_postfix = ensure_tuple_rep(meta_key_postfix, len(self.keys)) |
|
|
|
|
|
def __call__(self, data): |
|
|
d = dict(data) |
|
|
for key, meta_key, meta_key_postfix in self.key_iterator(d, self.meta_keys, self.meta_key_postfix): |
|
|
if meta_key is None and meta_key_postfix is not None: |
|
|
meta_key = f"{key}_{meta_key_postfix}" |
|
|
meta_data = d[meta_key] if meta_key is not None else None |
|
|
self.saver.save(data=d[key], meta_data=meta_data) |
|
|
if self.flush: |
|
|
self.saver.finalize() |
|
|
|
|
|
return d |
|
|
|
|
|
def get_saver(self): |
|
|
""" |
|
|
If want to write content into file, may need to call `finalize` of saver when epoch completed. |
|
|
Or users can also get the cache content from `saver` instead of writing into file. |
|
|
|
|
|
""" |
|
|
return self.saver |
|
|
|
|
|
|
|
|
class SobelGradientsd(MapTransform): |
|
|
"""Calculate Sobel horizontal and vertical gradients of a grayscale image. |
|
|
|
|
|
Args: |
|
|
keys: keys of the corresponding items to model output. |
|
|
kernel_size: the size of the Sobel kernel. Defaults to 3. |
|
|
spatial_axes: the axes that define the direction of the gradient to be calculated. It calculate the gradient |
|
|
along each of the provide axis. By default it calculate the gradient for all spatial axes. |
|
|
normalize_kernels: if normalize the Sobel kernel to provide proper gradients. Defaults to True. |
|
|
normalize_gradients: if normalize the output gradient to 0 and 1. Defaults to False. |
|
|
padding_mode: the padding mode of the image when convolving with Sobel kernels. Defaults to `"reflect"`. |
|
|
Acceptable values are ``'zeros'``, ``'reflect'``, ``'replicate'`` or ``'circular'``. |
|
|
See ``torch.nn.Conv1d()`` for more information. |
|
|
dtype: kernel data type (torch.dtype). Defaults to `torch.float32`. |
|
|
new_key_prefix: this prefix be prepended to the key to create a new key for the output and keep the value of |
|
|
key intact. By default not prefix is set and the corresponding array to the key will be replaced. |
|
|
allow_missing_keys: don't raise exception if key is missing. |
|
|
|
|
|
""" |
|
|
|
|
|
backend = SobelGradients.backend |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
keys: KeysCollection, |
|
|
kernel_size: int = 3, |
|
|
spatial_axes: Sequence[int] | int | None = None, |
|
|
normalize_kernels: bool = True, |
|
|
normalize_gradients: bool = False, |
|
|
padding_mode: str = "reflect", |
|
|
dtype: torch.dtype = torch.float32, |
|
|
new_key_prefix: str | None = None, |
|
|
allow_missing_keys: bool = False, |
|
|
) -> None: |
|
|
super().__init__(keys, allow_missing_keys) |
|
|
self.transform = SobelGradients( |
|
|
kernel_size=kernel_size, |
|
|
spatial_axes=spatial_axes, |
|
|
normalize_kernels=normalize_kernels, |
|
|
normalize_gradients=normalize_gradients, |
|
|
padding_mode=padding_mode, |
|
|
dtype=dtype, |
|
|
) |
|
|
self.new_key_prefix = new_key_prefix |
|
|
self.kernel_diff = self.transform.kernel_diff |
|
|
self.kernel_smooth = self.transform.kernel_smooth |
|
|
|
|
|
def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> dict[Hashable, NdarrayOrTensor]: |
|
|
d = dict(data) |
|
|
for key in self.key_iterator(d): |
|
|
new_key = key if self.new_key_prefix is None else self.new_key_prefix + key |
|
|
d[new_key] = self.transform(d[key]) |
|
|
|
|
|
return d |
|
|
|
|
|
|
|
|
class DistanceTransformEDTd(MapTransform): |
|
|
""" |
|
|
Applies the Euclidean distance transform on the input. |
|
|
Either GPU based with CuPy / cuCIM or CPU based with scipy. |
|
|
To use the GPU implementation, make sure cuCIM is available and that the data is a `torch.tensor` on a GPU device. |
|
|
|
|
|
Note that the results of the libraries can differ, so stick to one if possible. |
|
|
For details, check out the `SciPy`_ and `cuCIM`_ documentation and / or :func:`monai.transforms.utils.distance_transform_edt`. |
|
|
|
|
|
|
|
|
Note on the input shape: |
|
|
Has to be a channel first array, must have shape: (num_channels, H, W [,D]). |
|
|
Can be of any type but will be converted into binary: 1 wherever image equates to True, 0 elsewhere. |
|
|
Input gets passed channel-wise to the distance-transform, thus results from this function will differ |
|
|
from directly calling ``distance_transform_edt()`` in CuPy or SciPy. |
|
|
|
|
|
Args: |
|
|
keys: keys of the corresponding items to be transformed. |
|
|
allow_missing_keys: don't raise exception if key is missing. |
|
|
sampling: Spacing of elements along each dimension. If a sequence, must be of length equal to the input rank -1; |
|
|
if a single number, this is used for all axes. If not specified, a grid spacing of unity is implied. |
|
|
|
|
|
.. _SciPy: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.distance_transform_edt.html |
|
|
.. _cuCIM: https://docs.rapids.ai/api/cucim/nightly/api/#cucim.core.operations.morphology.distance_transform_edt |
|
|
|
|
|
|
|
|
""" |
|
|
|
|
|
backend = DistanceTransformEDT.backend |
|
|
|
|
|
def __init__( |
|
|
self, keys: KeysCollection, allow_missing_keys: bool = False, sampling: None | float | list[float] = None |
|
|
) -> None: |
|
|
super().__init__(keys, allow_missing_keys) |
|
|
self.sampling = sampling |
|
|
self.distance_transform = DistanceTransformEDT(sampling=self.sampling) |
|
|
|
|
|
def __call__(self, data: Mapping[Hashable, NdarrayOrTensor]) -> Mapping[Hashable, NdarrayOrTensor]: |
|
|
d = dict(data) |
|
|
for key in self.key_iterator(d): |
|
|
d[key] = self.distance_transform(img=d[key]) |
|
|
|
|
|
return d |
|
|
|
|
|
|
|
|
ActivationsD = ActivationsDict = Activationsd |
|
|
AsDiscreteD = AsDiscreteDict = AsDiscreted |
|
|
FillHolesD = FillHolesDict = FillHolesd |
|
|
InvertD = InvertDict = Invertd |
|
|
KeepLargestConnectedComponentD = KeepLargestConnectedComponentDict = KeepLargestConnectedComponentd |
|
|
RemoveSmallObjectsD = RemoveSmallObjectsDict = RemoveSmallObjectsd |
|
|
LabelFilterD = LabelFilterDict = LabelFilterd |
|
|
LabelToContourD = LabelToContourDict = LabelToContourd |
|
|
MeanEnsembleD = MeanEnsembleDict = MeanEnsembled |
|
|
ProbNMSD = ProbNMSDict = ProbNMSd |
|
|
SaveClassificationD = SaveClassificationDict = SaveClassificationd |
|
|
VoteEnsembleD = VoteEnsembleDict = VoteEnsembled |
|
|
EnsembleD = EnsembleDict = Ensembled |
|
|
SobelGradientsD = SobelGradientsDict = SobelGradientsd |
|
|
DistanceTransformEDTD = DistanceTransformEDTDict = DistanceTransformEDTd |
|
|
|