|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
""" |
|
|
A collection of "vanilla" transforms for utility functions. |
|
|
""" |
|
|
|
|
|
from __future__ import annotations |
|
|
|
|
|
import logging |
|
|
import sys |
|
|
import time |
|
|
import warnings |
|
|
from collections.abc import Mapping, Sequence |
|
|
from copy import deepcopy |
|
|
from functools import partial |
|
|
from typing import Any, Callable |
|
|
|
|
|
import numpy as np |
|
|
import torch |
|
|
import torch.nn as nn |
|
|
|
|
|
from monai.config import DtypeLike |
|
|
from monai.config.type_definitions import NdarrayOrTensor |
|
|
from monai.data.meta_obj import get_track_meta |
|
|
from monai.data.meta_tensor import MetaTensor |
|
|
from monai.data.utils import is_no_channel, no_collation |
|
|
from monai.networks.layers.simplelayers import ( |
|
|
ApplyFilter, |
|
|
EllipticalFilter, |
|
|
GaussianFilter, |
|
|
LaplaceFilter, |
|
|
MeanFilter, |
|
|
SavitzkyGolayFilter, |
|
|
SharpenFilter, |
|
|
median_filter, |
|
|
) |
|
|
from monai.transforms.inverse import InvertibleTransform |
|
|
from monai.transforms.traits import MultiSampleTrait |
|
|
from monai.transforms.transform import Randomizable, RandomizableTrait, RandomizableTransform, Transform |
|
|
from monai.transforms.utils import ( |
|
|
extreme_points_to_image, |
|
|
get_extreme_points, |
|
|
map_binary_to_indices, |
|
|
map_classes_to_indices, |
|
|
) |
|
|
from monai.transforms.utils_pytorch_numpy_unification import concatenate, in1d, moveaxis, unravel_indices |
|
|
from monai.utils import ( |
|
|
MetaKeys, |
|
|
TraceKeys, |
|
|
convert_data_type, |
|
|
convert_to_cupy, |
|
|
convert_to_numpy, |
|
|
convert_to_tensor, |
|
|
ensure_tuple, |
|
|
look_up_option, |
|
|
min_version, |
|
|
optional_import, |
|
|
) |
|
|
from monai.utils.enums import TransformBackends |
|
|
from monai.utils.misc import is_module_ver_at_least |
|
|
from monai.utils.type_conversion import convert_to_dst_type, get_equivalent_dtype |
|
|
|
|
|
PILImageImage, has_pil = optional_import("PIL.Image", name="Image") |
|
|
pil_image_fromarray, _ = optional_import("PIL.Image", name="fromarray") |
|
|
cp, has_cp = optional_import("cupy") |
|
|
|
|
|
__all__ = [ |
|
|
"Identity", |
|
|
"RandIdentity", |
|
|
"AsChannelLast", |
|
|
"AddCoordinateChannels", |
|
|
"EnsureChannelFirst", |
|
|
"EnsureType", |
|
|
"RepeatChannel", |
|
|
"RemoveRepeatedChannel", |
|
|
"SplitDim", |
|
|
"CastToType", |
|
|
"ToTensor", |
|
|
"ToNumpy", |
|
|
"ToPIL", |
|
|
"Transpose", |
|
|
"SqueezeDim", |
|
|
"DataStats", |
|
|
"SimulateDelay", |
|
|
"Lambda", |
|
|
"RandLambda", |
|
|
"LabelToMask", |
|
|
"FgBgToIndices", |
|
|
"ClassesToIndices", |
|
|
"ConvertToMultiChannelBasedOnBratsClasses", |
|
|
"AddExtremePointsChannel", |
|
|
"TorchVision", |
|
|
"MapLabelValue", |
|
|
"IntensityStats", |
|
|
"ToDevice", |
|
|
"CuCIM", |
|
|
"RandCuCIM", |
|
|
"ToCupy", |
|
|
"ImageFilter", |
|
|
"RandImageFilter", |
|
|
] |
|
|
|
|
|
|
|
|
class Identity(Transform): |
|
|
""" |
|
|
Do nothing to the data. |
|
|
As the output value is same as input, it can be used as a testing tool to verify the transform chain, |
|
|
Compose or transform adaptor, etc. |
|
|
""" |
|
|
|
|
|
backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
|
|
|
|
|
def __call__(self, img: NdarrayOrTensor) -> NdarrayOrTensor: |
|
|
""" |
|
|
Apply the transform to `img`. |
|
|
""" |
|
|
return img |
|
|
|
|
|
|
|
|
class RandIdentity(RandomizableTrait): |
|
|
""" |
|
|
Do nothing to the data. This transform is random, so can be used to stop the caching of any |
|
|
subsequent transforms. |
|
|
""" |
|
|
|
|
|
backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
|
|
|
|
|
def __call__(self, data: Any) -> Any: |
|
|
return data |
|
|
|
|
|
|
|
|
class AsChannelLast(Transform): |
|
|
""" |
|
|
Change the channel dimension of the image to the last dimension. |
|
|
|
|
|
Some of other 3rd party transforms assume the input image is in the channel-last format with shape |
|
|
(spatial_dim_1[, spatial_dim_2, ...], num_channels). |
|
|
|
|
|
This transform could be used to convert, for example, a channel-first image array in shape |
|
|
(num_channels, spatial_dim_1[, spatial_dim_2, ...]) into the channel-last format, |
|
|
so that MONAI transforms can construct a chain with other 3rd party transforms together. |
|
|
|
|
|
Args: |
|
|
channel_dim: which dimension of input image is the channel, default is the first dimension. |
|
|
""" |
|
|
|
|
|
backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
|
|
|
|
|
def __init__(self, channel_dim: int = 0) -> None: |
|
|
if not (isinstance(channel_dim, int) and channel_dim >= -1): |
|
|
raise ValueError(f"invalid channel dimension ({channel_dim}).") |
|
|
self.channel_dim = channel_dim |
|
|
|
|
|
def __call__(self, img: NdarrayOrTensor) -> NdarrayOrTensor: |
|
|
""" |
|
|
Apply the transform to `img`. |
|
|
""" |
|
|
out: NdarrayOrTensor = convert_to_tensor(moveaxis(img, self.channel_dim, -1), track_meta=get_track_meta()) |
|
|
return out |
|
|
|
|
|
|
|
|
class EnsureChannelFirst(Transform): |
|
|
""" |
|
|
Adjust or add the channel dimension of input data to ensure `channel_first` shape. |
|
|
|
|
|
This extracts the `original_channel_dim` info from provided meta_data dictionary or MetaTensor input. This value |
|
|
should state which dimension is the channel dimension so that it can be moved forward, or contain "no_channel" to |
|
|
state no dimension is the channel and so a 1-size first dimension is to be added. |
|
|
|
|
|
Args: |
|
|
strict_check: whether to raise an error when the meta information is insufficient. |
|
|
channel_dim: This argument can be used to specify the original channel dimension (integer) of the input array. |
|
|
It overrides the `original_channel_dim` from provided MetaTensor input. |
|
|
If the input array doesn't have a channel dim, this value should be ``'no_channel'``. |
|
|
If this is set to `None`, this class relies on `img` or `meta_dict` to provide the channel dimension. |
|
|
""" |
|
|
|
|
|
backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
|
|
|
|
|
def __init__(self, strict_check: bool = True, channel_dim: None | str | int = None): |
|
|
self.strict_check = strict_check |
|
|
self.input_channel_dim = channel_dim |
|
|
|
|
|
def __call__(self, img: torch.Tensor, meta_dict: Mapping | None = None) -> torch.Tensor: |
|
|
""" |
|
|
Apply the transform to `img`. |
|
|
""" |
|
|
if not isinstance(img, MetaTensor) and not isinstance(meta_dict, Mapping): |
|
|
if self.input_channel_dim is None: |
|
|
msg = "Metadata not available and channel_dim=None, EnsureChannelFirst is not in use." |
|
|
if self.strict_check: |
|
|
raise ValueError(msg) |
|
|
warnings.warn(msg) |
|
|
return img |
|
|
else: |
|
|
img = MetaTensor(img) |
|
|
|
|
|
if isinstance(img, MetaTensor): |
|
|
meta_dict = img.meta |
|
|
|
|
|
channel_dim = meta_dict.get(MetaKeys.ORIGINAL_CHANNEL_DIM, None) if isinstance(meta_dict, Mapping) else None |
|
|
if self.input_channel_dim is not None: |
|
|
channel_dim = float("nan") if self.input_channel_dim == "no_channel" else self.input_channel_dim |
|
|
|
|
|
if channel_dim is None: |
|
|
msg = "Unknown original_channel_dim in the MetaTensor meta dict or `meta_dict` or `channel_dim`." |
|
|
if self.strict_check: |
|
|
raise ValueError(msg) |
|
|
warnings.warn(msg) |
|
|
return img |
|
|
|
|
|
|
|
|
if isinstance(meta_dict, dict): |
|
|
meta_dict[MetaKeys.ORIGINAL_CHANNEL_DIM] = channel_dim |
|
|
|
|
|
if is_no_channel(channel_dim): |
|
|
result = img[None] |
|
|
else: |
|
|
result = moveaxis(img, int(channel_dim), 0) |
|
|
|
|
|
return convert_to_tensor(result, track_meta=get_track_meta()) |
|
|
|
|
|
|
|
|
class RepeatChannel(Transform): |
|
|
""" |
|
|
Repeat channel data to construct expected input shape for models. |
|
|
The `repeats` count includes the origin data, for example: |
|
|
``RepeatChannel(repeats=2)([[1, 2], [3, 4]])`` generates: ``[[1, 2], [1, 2], [3, 4], [3, 4]]`` |
|
|
|
|
|
Args: |
|
|
repeats: the number of repetitions for each element. |
|
|
""" |
|
|
|
|
|
backend = [TransformBackends.TORCH] |
|
|
|
|
|
def __init__(self, repeats: int) -> None: |
|
|
if repeats <= 0: |
|
|
raise ValueError(f"repeats count must be greater than 0, got {repeats}.") |
|
|
self.repeats = repeats |
|
|
|
|
|
def __call__(self, img: NdarrayOrTensor) -> NdarrayOrTensor: |
|
|
""" |
|
|
Apply the transform to `img`, assuming `img` is a "channel-first" array. |
|
|
""" |
|
|
repeat_fn = torch.repeat_interleave if isinstance(img, torch.Tensor) else np.repeat |
|
|
return convert_to_tensor(repeat_fn(img, self.repeats, 0), track_meta=get_track_meta()) |
|
|
|
|
|
|
|
|
class RemoveRepeatedChannel(Transform): |
|
|
""" |
|
|
RemoveRepeatedChannel data to undo RepeatChannel |
|
|
The `repeats` count specifies the deletion of the origin data, for example: |
|
|
``RemoveRepeatedChannel(repeats=2)([[1, 2], [1, 2], [3, 4], [3, 4]])`` generates: ``[[1, 2], [3, 4]]`` |
|
|
|
|
|
Args: |
|
|
repeats: the number of repetitions to be deleted for each element. |
|
|
""" |
|
|
|
|
|
backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
|
|
|
|
|
def __init__(self, repeats: int) -> None: |
|
|
if repeats <= 0: |
|
|
raise ValueError(f"repeats count must be greater than 0, got {repeats}.") |
|
|
|
|
|
self.repeats = repeats |
|
|
|
|
|
def __call__(self, img: NdarrayOrTensor) -> NdarrayOrTensor: |
|
|
""" |
|
|
Apply the transform to `img`, assuming `img` is a "channel-first" array. |
|
|
""" |
|
|
if img.shape[0] < 2: |
|
|
raise ValueError(f"Image must have more than one channel, got {img.shape[0]} channels.") |
|
|
|
|
|
out: NdarrayOrTensor = convert_to_tensor(img[:: self.repeats, :], track_meta=get_track_meta()) |
|
|
return out |
|
|
|
|
|
|
|
|
class SplitDim(Transform, MultiSampleTrait): |
|
|
""" |
|
|
Given an image of size X along a certain dimension, return a list of length X containing |
|
|
images. Useful for converting 3D images into a stack of 2D images, splitting multichannel inputs into |
|
|
single channels, for example. |
|
|
|
|
|
Note: `torch.split`/`np.split` is used, so the outputs are views of the input (shallow copy). |
|
|
|
|
|
Args: |
|
|
dim: dimension on which to split |
|
|
keepdim: if `True`, output will have singleton in the split dimension. If `False`, this |
|
|
dimension will be squeezed. |
|
|
update_meta: whether to update the MetaObj in each split result. |
|
|
""" |
|
|
|
|
|
backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
|
|
|
|
|
def __init__(self, dim: int = -1, keepdim: bool = True, update_meta=True) -> None: |
|
|
self.dim = dim |
|
|
self.keepdim = keepdim |
|
|
self.update_meta = update_meta |
|
|
|
|
|
def __call__(self, img: torch.Tensor) -> list[torch.Tensor]: |
|
|
""" |
|
|
Apply the transform to `img`. |
|
|
""" |
|
|
n_out = img.shape[self.dim] |
|
|
if isinstance(img, torch.Tensor): |
|
|
outputs = list(torch.split(img, 1, self.dim)) |
|
|
else: |
|
|
outputs = np.split(img, n_out, self.dim) |
|
|
for idx, item in enumerate(outputs): |
|
|
if not self.keepdim: |
|
|
outputs[idx] = item.squeeze(self.dim) |
|
|
if self.update_meta and isinstance(img, MetaTensor): |
|
|
if not isinstance(item, MetaTensor): |
|
|
item = MetaTensor(item, meta=img.meta) |
|
|
if self.dim == 0: |
|
|
continue |
|
|
ndim = len(item.affine) |
|
|
shift = torch.eye(ndim, device=item.affine.device, dtype=item.affine.dtype) |
|
|
shift[self.dim - 1, -1] = idx |
|
|
item.affine = item.affine @ shift |
|
|
return outputs |
|
|
|
|
|
|
|
|
class CastToType(Transform): |
|
|
""" |
|
|
Cast the Numpy data to specified numpy data type, or cast the PyTorch Tensor to |
|
|
specified PyTorch data type. |
|
|
|
|
|
Example: |
|
|
>>> import numpy as np |
|
|
>>> import torch |
|
|
>>> transform = CastToType(dtype=np.float32) |
|
|
|
|
|
>>> # Example with a numpy array |
|
|
>>> img_np = np.array([0, 127, 255], dtype=np.uint8) |
|
|
>>> img_np_casted = transform(img_np) |
|
|
>>> img_np_casted |
|
|
array([ 0. , 127. , 255. ], dtype=float32) |
|
|
|
|
|
>>> # Example with a PyTorch tensor |
|
|
>>> img_tensor = torch.tensor([0, 127, 255], dtype=torch.uint8) |
|
|
>>> img_tensor_casted = transform(img_tensor) |
|
|
>>> img_tensor_casted |
|
|
tensor([ 0., 127., 255.]) # dtype is float32 |
|
|
""" |
|
|
|
|
|
backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
|
|
|
|
|
def __init__(self, dtype=np.float32) -> None: |
|
|
""" |
|
|
Args: |
|
|
dtype: convert image to this data type, default is `np.float32`. |
|
|
""" |
|
|
self.dtype = dtype |
|
|
|
|
|
def __call__(self, img: NdarrayOrTensor, dtype: DtypeLike | torch.dtype = None) -> NdarrayOrTensor: |
|
|
""" |
|
|
Apply the transform to `img`, assuming `img` is a numpy array or PyTorch Tensor. |
|
|
|
|
|
Args: |
|
|
dtype: convert image to this data type, default is `self.dtype`. |
|
|
|
|
|
Raises: |
|
|
TypeError: When ``img`` type is not in ``Union[numpy.ndarray, torch.Tensor]``. |
|
|
|
|
|
""" |
|
|
return convert_data_type(img, output_type=type(img), dtype=dtype or self.dtype)[0] |
|
|
|
|
|
|
|
|
class ToTensor(Transform): |
|
|
""" |
|
|
Converts the input image to a tensor without applying any other transformations. |
|
|
Input data can be PyTorch Tensor, numpy array, list, dictionary, int, float, bool, str, etc. |
|
|
Will convert Tensor, Numpy array, float, int, bool to Tensor, strings and objects keep the original. |
|
|
For dictionary, list or tuple, convert every item to a Tensor if applicable and `wrap_sequence=False`. |
|
|
|
|
|
Args: |
|
|
dtype: target data type to when converting to Tensor. |
|
|
device: target device to put the converted Tensor data. |
|
|
wrap_sequence: if `False`, then lists will recursively call this function, default to `True`. |
|
|
E.g., if `False`, `[1, 2]` -> `[tensor(1), tensor(2)]`, if `True`, then `[1, 2]` -> `tensor([1, 2])`. |
|
|
track_meta: whether to convert to `MetaTensor` or regular tensor, default to `None`, |
|
|
use the return value of ``get_track_meta``. |
|
|
|
|
|
""" |
|
|
|
|
|
backend = [TransformBackends.TORCH] |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
dtype: torch.dtype | None = None, |
|
|
device: torch.device | str | None = None, |
|
|
wrap_sequence: bool = True, |
|
|
track_meta: bool | None = None, |
|
|
) -> None: |
|
|
super().__init__() |
|
|
self.dtype = dtype |
|
|
self.device = device |
|
|
self.wrap_sequence = wrap_sequence |
|
|
self.track_meta = get_track_meta() if track_meta is None else bool(track_meta) |
|
|
|
|
|
def __call__(self, img: NdarrayOrTensor): |
|
|
""" |
|
|
Apply the transform to `img` and make it contiguous. |
|
|
""" |
|
|
if isinstance(img, MetaTensor): |
|
|
img.applied_operations = [] |
|
|
return convert_to_tensor( |
|
|
img, dtype=self.dtype, device=self.device, wrap_sequence=self.wrap_sequence, track_meta=self.track_meta |
|
|
) |
|
|
|
|
|
|
|
|
class EnsureType(Transform): |
|
|
""" |
|
|
Ensure the input data to be a PyTorch Tensor or numpy array, support: `numpy array`, `PyTorch Tensor`, |
|
|
`float`, `int`, `bool`, `string` and `object` keep the original. |
|
|
If passing a dictionary, list or tuple, still return dictionary, list or tuple will recursively convert |
|
|
every item to the expected data type if `wrap_sequence=False`. |
|
|
|
|
|
Args: |
|
|
data_type: target data type to convert, should be "tensor" or "numpy". |
|
|
dtype: target data content type to convert, for example: np.float32, torch.float, etc. |
|
|
device: for Tensor data type, specify the target device. |
|
|
wrap_sequence: if `False`, then lists will recursively call this function, default to `True`. |
|
|
track_meta: if `True` convert to ``MetaTensor``, otherwise to Pytorch ``Tensor``, |
|
|
if ``None`` behave according to return value of py:func:`monai.data.meta_obj.get_track_meta`. |
|
|
|
|
|
Example with wrap_sequence=True: |
|
|
>>> import numpy as np |
|
|
>>> import torch |
|
|
>>> transform = EnsureType(data_type="tensor", wrap_sequence=True) |
|
|
>>> # Converting a list to a tensor |
|
|
>>> data_list = [1, 2., 3] |
|
|
>>> tensor_data = transform(data_list) |
|
|
>>> tensor_data |
|
|
tensor([1., 2., 3.]) # All elements have dtype float32 |
|
|
|
|
|
Example with wrap_sequence=False: |
|
|
>>> transform = EnsureType(data_type="tensor", wrap_sequence=False) |
|
|
>>> # Converting each element in a list to individual tensors |
|
|
>>> data_list = [1, 2, 3] |
|
|
>>> tensors_list = transform(data_list) |
|
|
>>> tensors_list |
|
|
[tensor(1), tensor(2.), tensor(3)] # Only second element is float32 rest are int64 |
|
|
""" |
|
|
|
|
|
backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
data_type: str = "tensor", |
|
|
dtype: DtypeLike | torch.dtype = None, |
|
|
device: torch.device | None = None, |
|
|
wrap_sequence: bool = True, |
|
|
track_meta: bool | None = None, |
|
|
) -> None: |
|
|
self.data_type = look_up_option(data_type.lower(), {"tensor", "numpy"}) |
|
|
self.dtype = dtype |
|
|
self.device = device |
|
|
self.wrap_sequence = wrap_sequence |
|
|
self.track_meta = get_track_meta() if track_meta is None else bool(track_meta) |
|
|
|
|
|
def __call__(self, data: NdarrayOrTensor, dtype: DtypeLike | torch.dtype = None): |
|
|
""" |
|
|
Args: |
|
|
data: input data can be PyTorch Tensor, numpy array, list, dictionary, int, float, bool, str, etc. |
|
|
will ensure Tensor, Numpy array, float, int, bool as Tensors or numpy arrays, strings and |
|
|
objects keep the original. for dictionary, list or tuple, ensure every item as expected type |
|
|
if applicable and `wrap_sequence=False`. |
|
|
dtype: target data content type to convert, for example: np.float32, torch.float, etc. |
|
|
|
|
|
""" |
|
|
if self.data_type == "tensor": |
|
|
output_type = MetaTensor if self.track_meta else torch.Tensor |
|
|
else: |
|
|
output_type = np.ndarray |
|
|
out: NdarrayOrTensor |
|
|
out, *_ = convert_data_type( |
|
|
data=data, |
|
|
output_type=output_type, |
|
|
dtype=self.dtype if dtype is None else dtype, |
|
|
device=self.device, |
|
|
wrap_sequence=self.wrap_sequence, |
|
|
) |
|
|
return out |
|
|
|
|
|
|
|
|
class ToNumpy(Transform): |
|
|
""" |
|
|
Converts the input data to numpy array, can support list or tuple of numbers and PyTorch Tensor. |
|
|
|
|
|
Args: |
|
|
dtype: target data type when converting to numpy array. |
|
|
wrap_sequence: if `False`, then lists will recursively call this function, default to `True`. |
|
|
E.g., if `False`, `[1, 2]` -> `[array(1), array(2)]`, if `True`, then `[1, 2]` -> `array([1, 2])`. |
|
|
|
|
|
""" |
|
|
|
|
|
backend = [TransformBackends.NUMPY] |
|
|
|
|
|
def __init__(self, dtype: DtypeLike = None, wrap_sequence: bool = True) -> None: |
|
|
super().__init__() |
|
|
self.dtype = dtype |
|
|
self.wrap_sequence = wrap_sequence |
|
|
|
|
|
def __call__(self, img: NdarrayOrTensor): |
|
|
""" |
|
|
Apply the transform to `img` and make it contiguous. |
|
|
""" |
|
|
return convert_to_numpy(img, dtype=self.dtype, wrap_sequence=self.wrap_sequence) |
|
|
|
|
|
|
|
|
class ToCupy(Transform): |
|
|
""" |
|
|
Converts the input data to CuPy array, can support list or tuple of numbers, NumPy and PyTorch Tensor. |
|
|
|
|
|
Args: |
|
|
dtype: data type specifier. It is inferred from the input by default. |
|
|
if not None, must be an argument of `numpy.dtype`, for more details: |
|
|
https://docs.cupy.dev/en/stable/reference/generated/cupy.array.html. |
|
|
wrap_sequence: if `False`, then lists will recursively call this function, default to `True`. |
|
|
E.g., if `False`, `[1, 2]` -> `[array(1), array(2)]`, if `True`, then `[1, 2]` -> `array([1, 2])`. |
|
|
|
|
|
""" |
|
|
|
|
|
backend = [TransformBackends.CUPY] |
|
|
|
|
|
def __init__(self, dtype: np.dtype | None = None, wrap_sequence: bool = True) -> None: |
|
|
super().__init__() |
|
|
self.dtype = dtype |
|
|
self.wrap_sequence = wrap_sequence |
|
|
|
|
|
def __call__(self, data: NdarrayOrTensor): |
|
|
""" |
|
|
Create a CuPy array from `data` and make it contiguous |
|
|
""" |
|
|
return convert_to_cupy(data, dtype=self.dtype, wrap_sequence=self.wrap_sequence) |
|
|
|
|
|
|
|
|
class ToPIL(Transform): |
|
|
""" |
|
|
Converts the input image (in the form of NumPy array or PyTorch Tensor) to PIL image |
|
|
""" |
|
|
|
|
|
backend = [TransformBackends.NUMPY] |
|
|
|
|
|
def __call__(self, img): |
|
|
""" |
|
|
Apply the transform to `img`. |
|
|
""" |
|
|
if isinstance(img, PILImageImage): |
|
|
return img |
|
|
if isinstance(img, torch.Tensor): |
|
|
img = img.detach().cpu().numpy() |
|
|
return pil_image_fromarray(img) |
|
|
|
|
|
|
|
|
class Transpose(Transform): |
|
|
""" |
|
|
Transposes the input image based on the given `indices` dimension ordering. |
|
|
""" |
|
|
|
|
|
backend = [TransformBackends.TORCH] |
|
|
|
|
|
def __init__(self, indices: Sequence[int] | None) -> None: |
|
|
self.indices = None if indices is None else tuple(indices) |
|
|
|
|
|
def __call__(self, img: NdarrayOrTensor) -> NdarrayOrTensor: |
|
|
""" |
|
|
Apply the transform to `img`. |
|
|
""" |
|
|
img = convert_to_tensor(img, track_meta=get_track_meta()) |
|
|
return img.permute(self.indices or tuple(range(img.ndim)[::-1])) |
|
|
|
|
|
|
|
|
class SqueezeDim(Transform): |
|
|
""" |
|
|
Squeeze a unitary dimension. |
|
|
""" |
|
|
|
|
|
backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
|
|
|
|
|
def __init__(self, dim: int | None = 0, update_meta=True) -> None: |
|
|
""" |
|
|
Args: |
|
|
dim: dimension to be squeezed. Default = 0 |
|
|
"None" works when the input is numpy array. |
|
|
update_meta: whether to update the meta info if the input is a metatensor. Default is ``True``. |
|
|
|
|
|
Raises: |
|
|
TypeError: When ``dim`` is not an ``Optional[int]``. |
|
|
|
|
|
""" |
|
|
if dim is not None and not isinstance(dim, int): |
|
|
raise TypeError(f"dim must be None or a int but is {type(dim).__name__}.") |
|
|
self.dim = dim |
|
|
self.update_meta = update_meta |
|
|
|
|
|
def __call__(self, img: NdarrayOrTensor) -> NdarrayOrTensor: |
|
|
""" |
|
|
Args: |
|
|
img: numpy arrays with required dimension `dim` removed |
|
|
""" |
|
|
img = convert_to_tensor(img, track_meta=get_track_meta()) |
|
|
if self.dim is None: |
|
|
if self.update_meta: |
|
|
warnings.warn("update_meta=True is ignored when dim=None.") |
|
|
return img.squeeze() |
|
|
dim = (self.dim + len(img.shape)) if self.dim < 0 else self.dim |
|
|
|
|
|
if img.shape[dim] != 1: |
|
|
raise ValueError(f"Can only squeeze singleton dimension, got shape {img.shape[dim]} of {img.shape}.") |
|
|
img = img.squeeze(dim) |
|
|
if self.update_meta and isinstance(img, MetaTensor) and dim > 0 and len(img.affine.shape) == 2: |
|
|
h, w = img.affine.shape |
|
|
affine, device = img.affine, img.affine.device if isinstance(img.affine, torch.Tensor) else None |
|
|
if h > dim: |
|
|
affine = affine[torch.arange(0, h, device=device) != dim - 1] |
|
|
if w > dim: |
|
|
affine = affine[:, torch.arange(0, w, device=device) != dim - 1] |
|
|
if (affine.shape[0] == affine.shape[1]) and not np.linalg.det(convert_to_numpy(affine, wrap_sequence=True)): |
|
|
warnings.warn(f"After SqueezeDim, img.affine is ill-posed: \n{img.affine}.") |
|
|
img.affine = affine |
|
|
return img |
|
|
|
|
|
|
|
|
class DataStats(Transform): |
|
|
""" |
|
|
Utility transform to show the statistics of data for debug or analysis. |
|
|
It can be inserted into any place of a transform chain and check results of previous transforms. |
|
|
It support both `numpy.ndarray` and `torch.tensor` as input data, |
|
|
so it can be used in pre-processing and post-processing. |
|
|
|
|
|
It gets logger from `logging.getLogger(name)`, we can setup a logger outside first with the same `name`. |
|
|
If the log level of `logging.RootLogger` is higher than `INFO`, will add a separate `StreamHandler` |
|
|
log handler with `INFO` level and record to `stdout`. |
|
|
|
|
|
""" |
|
|
|
|
|
backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
prefix: str = "Data", |
|
|
data_type: bool = True, |
|
|
data_shape: bool = True, |
|
|
value_range: bool = True, |
|
|
data_value: bool = False, |
|
|
additional_info: Callable | None = None, |
|
|
name: str = "DataStats", |
|
|
) -> None: |
|
|
""" |
|
|
Args: |
|
|
prefix: will be printed in format: "{prefix} statistics". |
|
|
data_type: whether to show the type of input data. |
|
|
data_shape: whether to show the shape of input data. |
|
|
value_range: whether to show the value range of input data. |
|
|
data_value: whether to show the raw value of input data. |
|
|
a typical example is to print some properties of Nifti image: affine, pixdim, etc. |
|
|
additional_info: user can define callable function to extract additional info from input data. |
|
|
name: identifier of `logging.logger` to use, defaulting to "DataStats". |
|
|
|
|
|
Raises: |
|
|
TypeError: When ``additional_info`` is not an ``Optional[Callable]``. |
|
|
|
|
|
""" |
|
|
if not isinstance(prefix, str): |
|
|
raise ValueError(f"prefix must be a string, got {type(prefix)}.") |
|
|
self.prefix = prefix |
|
|
self.data_type = data_type |
|
|
self.data_shape = data_shape |
|
|
self.value_range = value_range |
|
|
self.data_value = data_value |
|
|
if additional_info is not None and not callable(additional_info): |
|
|
raise TypeError(f"additional_info must be None or callable but is {type(additional_info).__name__}.") |
|
|
self.additional_info = additional_info |
|
|
self._logger_name = name |
|
|
_logger = logging.getLogger(self._logger_name) |
|
|
_logger.setLevel(logging.INFO) |
|
|
if logging.root.getEffectiveLevel() > logging.INFO: |
|
|
|
|
|
has_console_handler = any( |
|
|
hasattr(h, "is_data_stats_handler") and h.is_data_stats_handler for h in _logger.handlers |
|
|
) |
|
|
if not has_console_handler: |
|
|
|
|
|
console = logging.StreamHandler(sys.stdout) |
|
|
console.setLevel(logging.INFO) |
|
|
console.is_data_stats_handler = True |
|
|
_logger.addHandler(console) |
|
|
|
|
|
def __call__( |
|
|
self, |
|
|
img: NdarrayOrTensor, |
|
|
prefix: str | None = None, |
|
|
data_type: bool | None = None, |
|
|
data_shape: bool | None = None, |
|
|
value_range: bool | None = None, |
|
|
data_value: bool | None = None, |
|
|
additional_info: Callable | None = None, |
|
|
) -> NdarrayOrTensor: |
|
|
""" |
|
|
Apply the transform to `img`, optionally take arguments similar to the class constructor. |
|
|
""" |
|
|
lines = [f"{prefix or self.prefix} statistics:"] |
|
|
|
|
|
if self.data_type if data_type is None else data_type: |
|
|
lines.append(f"Type: {type(img)} {img.dtype if hasattr(img, 'dtype') else None}") |
|
|
if self.data_shape if data_shape is None else data_shape: |
|
|
lines.append(f"Shape: {img.shape if hasattr(img, 'shape') else None}") |
|
|
if self.value_range if value_range is None else value_range: |
|
|
if isinstance(img, np.ndarray): |
|
|
lines.append(f"Value range: ({np.min(img)}, {np.max(img)})") |
|
|
elif isinstance(img, torch.Tensor): |
|
|
lines.append(f"Value range: ({torch.min(img)}, {torch.max(img)})") |
|
|
else: |
|
|
lines.append(f"Value range: (not a PyTorch or Numpy array, type: {type(img)})") |
|
|
if self.data_value if data_value is None else data_value: |
|
|
lines.append(f"Value: {img}") |
|
|
additional_info = self.additional_info if additional_info is None else additional_info |
|
|
if additional_info is not None: |
|
|
lines.append(f"Additional info: {additional_info(img)}") |
|
|
separator = "\n" |
|
|
output = f"{separator.join(lines)}" |
|
|
logging.getLogger(self._logger_name).info(output) |
|
|
return img |
|
|
|
|
|
|
|
|
class SimulateDelay(Transform): |
|
|
""" |
|
|
This is a pass through transform to be used for testing purposes. It allows |
|
|
adding fake behaviors that are useful for testing purposes to simulate |
|
|
how large datasets behave without needing to test on large data sets. |
|
|
|
|
|
For example, simulating slow NFS data transfers, or slow network transfers |
|
|
in testing by adding explicit timing delays. Testing of small test data |
|
|
can lead to incomplete understanding of real world issues, and may lead |
|
|
to sub-optimal design choices. |
|
|
""" |
|
|
|
|
|
backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
|
|
|
|
|
def __init__(self, delay_time: float = 0.0) -> None: |
|
|
""" |
|
|
Args: |
|
|
delay_time: The minimum amount of time, in fractions of seconds, |
|
|
to accomplish this delay task. |
|
|
""" |
|
|
super().__init__() |
|
|
self.delay_time: float = delay_time |
|
|
|
|
|
def __call__(self, img: NdarrayOrTensor, delay_time: float | None = None) -> NdarrayOrTensor: |
|
|
""" |
|
|
Args: |
|
|
img: data remain unchanged throughout this transform. |
|
|
delay_time: The minimum amount of time, in fractions of seconds, |
|
|
to accomplish this delay task. |
|
|
""" |
|
|
time.sleep(self.delay_time if delay_time is None else delay_time) |
|
|
return img |
|
|
|
|
|
|
|
|
class Lambda(InvertibleTransform): |
|
|
""" |
|
|
Apply a user-defined lambda as a transform. |
|
|
|
|
|
For example: |
|
|
|
|
|
.. code-block:: python |
|
|
:emphasize-lines: 2 |
|
|
|
|
|
image = np.ones((10, 2, 2)) |
|
|
lambd = Lambda(func=lambda x: x[:4, :, :]) |
|
|
print(lambd(image).shape) |
|
|
(4, 2, 2) |
|
|
|
|
|
Args: |
|
|
func: Lambda/function to be applied. |
|
|
inv_func: Lambda/function of inverse operation, default to `lambda x: x`. |
|
|
track_meta: If `False`, then standard data objects will be returned (e.g., torch.Tensor` and `np.ndarray`) |
|
|
as opposed to MONAI's enhanced objects. By default, this is `True`. |
|
|
|
|
|
Raises: |
|
|
TypeError: When ``func`` is not an ``Optional[Callable]``. |
|
|
|
|
|
""" |
|
|
|
|
|
backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
|
|
|
|
|
def __init__( |
|
|
self, func: Callable | None = None, inv_func: Callable = no_collation, track_meta: bool = True |
|
|
) -> None: |
|
|
if func is not None and not callable(func): |
|
|
raise TypeError(f"func must be None or callable but is {type(func).__name__}.") |
|
|
self.func = func |
|
|
self.inv_func = inv_func |
|
|
self.track_meta = track_meta |
|
|
|
|
|
def __call__(self, img: NdarrayOrTensor, func: Callable | None = None): |
|
|
""" |
|
|
Apply `self.func` to `img`. |
|
|
|
|
|
Args: |
|
|
func: Lambda/function to be applied. Defaults to `self.func`. |
|
|
|
|
|
Raises: |
|
|
TypeError: When ``func`` is not an ``Optional[Callable]``. |
|
|
|
|
|
""" |
|
|
fn = func if func is not None else self.func |
|
|
if not callable(fn): |
|
|
raise TypeError(f"func must be None or callable but is {type(fn).__name__}.") |
|
|
out = fn(img) |
|
|
|
|
|
if isinstance(out, (np.ndarray, torch.Tensor)) and not isinstance(out, MetaTensor) and self.track_meta: |
|
|
out = MetaTensor(out) |
|
|
if isinstance(out, MetaTensor): |
|
|
self.push_transform(out) |
|
|
return out |
|
|
|
|
|
def inverse(self, data: torch.Tensor): |
|
|
if isinstance(data, MetaTensor): |
|
|
self.pop_transform(data) |
|
|
return self.inv_func(data) |
|
|
|
|
|
|
|
|
class RandLambda(Lambda, RandomizableTransform): |
|
|
""" |
|
|
Randomizable version :py:class:`monai.transforms.Lambda`, the input `func` may contain random logic, |
|
|
or randomly execute the function based on `prob`. |
|
|
|
|
|
Args: |
|
|
func: Lambda/function to be applied. |
|
|
prob: probability of executing the random function, default to 1.0, with 100% probability to execute. |
|
|
inv_func: Lambda/function of inverse operation, default to `lambda x: x`. |
|
|
track_meta: If `False`, then standard data objects will be returned (e.g., torch.Tensor` and `np.ndarray`) |
|
|
as opposed to MONAI's enhanced objects. By default, this is `True`. |
|
|
|
|
|
For more details, please check :py:class:`monai.transforms.Lambda`. |
|
|
""" |
|
|
|
|
|
backend = Lambda.backend |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
func: Callable | None = None, |
|
|
prob: float = 1.0, |
|
|
inv_func: Callable = no_collation, |
|
|
track_meta: bool = True, |
|
|
) -> None: |
|
|
Lambda.__init__(self=self, func=func, inv_func=inv_func, track_meta=track_meta) |
|
|
RandomizableTransform.__init__(self=self, prob=prob) |
|
|
|
|
|
def __call__(self, img: NdarrayOrTensor, func: Callable | None = None): |
|
|
self.randomize(img) |
|
|
out = deepcopy(super().__call__(img, func) if self._do_transform else img) |
|
|
|
|
|
if not isinstance(out, MetaTensor) and self.track_meta: |
|
|
out = MetaTensor(out) |
|
|
if isinstance(out, MetaTensor): |
|
|
lambda_info = self.pop_transform(out) if self._do_transform else {} |
|
|
self.push_transform(out, extra_info=lambda_info) |
|
|
return out |
|
|
|
|
|
def inverse(self, data: torch.Tensor): |
|
|
do_transform = self.get_most_recent_transform(data).pop(TraceKeys.DO_TRANSFORM) |
|
|
if do_transform: |
|
|
data = super().inverse(data) |
|
|
else: |
|
|
self.pop_transform(data) |
|
|
return data |
|
|
|
|
|
|
|
|
class LabelToMask(Transform): |
|
|
""" |
|
|
Convert labels to mask for other tasks. A typical usage is to convert segmentation labels |
|
|
to mask data to pre-process images and then feed the images into classification network. |
|
|
It can support single channel labels or One-Hot labels with specified `select_labels`. |
|
|
For example, users can select `label value = [2, 3]` to construct mask data, or select the |
|
|
second and the third channels of labels to construct mask data. |
|
|
The output mask data can be a multiple channels binary data or a single channel binary |
|
|
data that merges all the channels. |
|
|
|
|
|
Args: |
|
|
select_labels: labels to generate mask from. for 1 channel label, the `select_labels` |
|
|
is the expected label values, like: [1, 2, 3]. for One-Hot format label, the |
|
|
`select_labels` is the expected channel indices. |
|
|
merge_channels: whether to use `np.any()` to merge the result on channel dim. if yes, |
|
|
will return a single channel mask with binary data. |
|
|
|
|
|
""" |
|
|
|
|
|
backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
|
|
|
|
|
def __init__( |
|
|
self, select_labels: Sequence[int] | int, merge_channels: bool = False |
|
|
) -> None: |
|
|
self.select_labels = ensure_tuple(select_labels) |
|
|
self.merge_channels = merge_channels |
|
|
|
|
|
def __call__( |
|
|
self, img: NdarrayOrTensor, select_labels: Sequence[int] | int | None = None, merge_channels: bool = False |
|
|
) -> NdarrayOrTensor: |
|
|
""" |
|
|
Args: |
|
|
select_labels: labels to generate mask from. for 1 channel label, the `select_labels` |
|
|
is the expected label values, like: [1, 2, 3]. for One-Hot format label, the |
|
|
`select_labels` is the expected channel indices. |
|
|
merge_channels: whether to use `np.any()` to merge the result on channel dim. if yes, |
|
|
will return a single channel mask with binary data. |
|
|
""" |
|
|
img = convert_to_tensor(img, track_meta=get_track_meta()) |
|
|
if select_labels is None: |
|
|
select_labels = self.select_labels |
|
|
else: |
|
|
select_labels = ensure_tuple(select_labels) |
|
|
|
|
|
if img.shape[0] > 1: |
|
|
data = img[[*select_labels]] |
|
|
else: |
|
|
where: Callable = np.where if isinstance(img, np.ndarray) else torch.where |
|
|
if isinstance(img, np.ndarray) or is_module_ver_at_least(torch, (1, 8, 0)): |
|
|
data = where(in1d(img, select_labels), True, False).reshape(img.shape) |
|
|
|
|
|
else: |
|
|
data = where( |
|
|
in1d(img, select_labels), torch.tensor(1, device=img.device), torch.tensor(0, device=img.device) |
|
|
).reshape(img.shape) |
|
|
|
|
|
if merge_channels or self.merge_channels: |
|
|
if isinstance(img, np.ndarray) or is_module_ver_at_least(torch, (1, 8, 0)): |
|
|
return data.any(0)[None] |
|
|
|
|
|
return data.to(torch.uint8).any(0)[None].to(bool) |
|
|
|
|
|
return data |
|
|
|
|
|
|
|
|
class FgBgToIndices(Transform, MultiSampleTrait): |
|
|
""" |
|
|
Compute foreground and background of the input label data, return the indices. |
|
|
If no output_shape specified, output data will be 1 dim indices after flattening. |
|
|
This transform can help pre-compute foreground and background regions for other transforms. |
|
|
A typical usage is to randomly select foreground and background to crop. |
|
|
The main logic is based on :py:class:`monai.transforms.utils.map_binary_to_indices`. |
|
|
|
|
|
Args: |
|
|
image_threshold: if enabled `image` at runtime, use ``image > image_threshold`` to |
|
|
determine the valid image content area and select background only in this area. |
|
|
output_shape: expected shape of output indices. if not None, unravel indices to specified shape. |
|
|
|
|
|
""" |
|
|
|
|
|
backend = [TransformBackends.NUMPY, TransformBackends.TORCH] |
|
|
|
|
|
def __init__(self, image_threshold: float = 0.0, output_shape: Sequence[int] | None = None) -> None: |
|
|
self.image_threshold = image_threshold |
|
|
self.output_shape = output_shape |
|
|
|
|
|
def __call__( |
|
|
self, label: NdarrayOrTensor, image: NdarrayOrTensor | None = None, output_shape: Sequence[int] | None = None |
|
|
) -> tuple[NdarrayOrTensor, NdarrayOrTensor]: |
|
|
""" |
|
|
Args: |
|
|
label: input data to compute foreground and background indices. |
|
|
image: if image is not None, use ``label = 0 & image > image_threshold`` |
|
|
to define background. so the output items will not map to all the voxels in the label. |
|
|
output_shape: expected shape of output indices. if None, use `self.output_shape` instead. |
|
|
|
|
|
""" |
|
|
if output_shape is None: |
|
|
output_shape = self.output_shape |
|
|
fg_indices, bg_indices = map_binary_to_indices(label, image, self.image_threshold) |
|
|
if output_shape is not None: |
|
|
fg_indices = unravel_indices(fg_indices, output_shape) |
|
|
bg_indices = unravel_indices(bg_indices, output_shape) |
|
|
return fg_indices, bg_indices |
|
|
|
|
|
|
|
|
class ClassesToIndices(Transform, MultiSampleTrait): |
|
|
backend = [TransformBackends.NUMPY, TransformBackends.TORCH] |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
num_classes: int | None = None, |
|
|
image_threshold: float = 0.0, |
|
|
output_shape: Sequence[int] | None = None, |
|
|
max_samples_per_class: int | None = None, |
|
|
) -> None: |
|
|
""" |
|
|
Compute indices of every class of the input label data, return a list of indices. |
|
|
If no output_shape specified, output data will be 1 dim indices after flattening. |
|
|
This transform can help pre-compute indices of the class regions for other transforms. |
|
|
A typical usage is to randomly select indices of classes to crop. |
|
|
The main logic is based on :py:class:`monai.transforms.utils.map_classes_to_indices`. |
|
|
|
|
|
Args: |
|
|
num_classes: number of classes for argmax label, not necessary for One-Hot label. |
|
|
image_threshold: if enabled `image` at runtime, use ``image > image_threshold`` to |
|
|
determine the valid image content area and select only the indices of classes in this area. |
|
|
output_shape: expected shape of output indices. if not None, unravel indices to specified shape. |
|
|
max_samples_per_class: maximum length of indices to sample in each class to reduce memory consumption. |
|
|
Default is None, no subsampling. |
|
|
|
|
|
""" |
|
|
self.num_classes = num_classes |
|
|
self.image_threshold = image_threshold |
|
|
self.output_shape = output_shape |
|
|
self.max_samples_per_class = max_samples_per_class |
|
|
|
|
|
def __call__( |
|
|
self, label: NdarrayOrTensor, image: NdarrayOrTensor | None = None, output_shape: Sequence[int] | None = None |
|
|
) -> list[NdarrayOrTensor]: |
|
|
""" |
|
|
Args: |
|
|
label: input data to compute the indices of every class. |
|
|
image: if image is not None, use ``image > image_threshold`` to define valid region, and only select |
|
|
the indices within the valid region. |
|
|
output_shape: expected shape of output indices. if None, use `self.output_shape` instead. |
|
|
|
|
|
""" |
|
|
|
|
|
if output_shape is None: |
|
|
output_shape = self.output_shape |
|
|
indices: list[NdarrayOrTensor] |
|
|
indices = map_classes_to_indices( |
|
|
label, self.num_classes, image, self.image_threshold, self.max_samples_per_class |
|
|
) |
|
|
if output_shape is not None: |
|
|
indices = [unravel_indices(cls_indices, output_shape) for cls_indices in indices] |
|
|
|
|
|
return indices |
|
|
|
|
|
|
|
|
class ConvertToMultiChannelBasedOnBratsClasses(Transform): |
|
|
""" |
|
|
Convert labels to multi channels based on brats18 classes: |
|
|
label 1 is the necrotic and non-enhancing tumor core |
|
|
label 2 is the peritumoral edema |
|
|
label 4 is the GD-enhancing tumor |
|
|
The possible classes are TC (Tumor core), WT (Whole tumor) |
|
|
and ET (Enhancing tumor). |
|
|
""" |
|
|
|
|
|
backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
|
|
|
|
|
def __call__(self, img: NdarrayOrTensor) -> NdarrayOrTensor: |
|
|
|
|
|
if img.ndim == 4 and img.shape[0] == 1: |
|
|
img = img.squeeze(0) |
|
|
|
|
|
result = [(img == 1) | (img == 4), (img == 1) | (img == 4) | (img == 2), img == 4] |
|
|
|
|
|
|
|
|
return torch.stack(result, dim=0) if isinstance(img, torch.Tensor) else np.stack(result, axis=0) |
|
|
|
|
|
|
|
|
class AddExtremePointsChannel(Randomizable, Transform): |
|
|
""" |
|
|
Add extreme points of label to the image as a new channel. This transform generates extreme |
|
|
point from label and applies a gaussian filter. The pixel values in points image are rescaled |
|
|
to range [rescale_min, rescale_max] and added as a new channel to input image. The algorithm is |
|
|
described in Roth et al., Going to Extremes: Weakly Supervised Medical Image Segmentation |
|
|
https://arxiv.org/abs/2009.11988. |
|
|
|
|
|
This transform only supports single channel labels (1, spatial_dim1, [spatial_dim2, ...]). The |
|
|
background ``index`` is ignored when calculating extreme points. |
|
|
|
|
|
Args: |
|
|
background: Class index of background label, defaults to 0. |
|
|
pert: Random perturbation amount to add to the points, defaults to 0.0. |
|
|
|
|
|
Raises: |
|
|
ValueError: When no label image provided. |
|
|
ValueError: When label image is not single channel. |
|
|
""" |
|
|
|
|
|
backend = [TransformBackends.TORCH] |
|
|
|
|
|
def __init__(self, background: int = 0, pert: float = 0.0) -> None: |
|
|
self._background = background |
|
|
self._pert = pert |
|
|
self._points: list[tuple[int, ...]] = [] |
|
|
|
|
|
def randomize(self, label: NdarrayOrTensor) -> None: |
|
|
self._points = get_extreme_points(label, rand_state=self.R, background=self._background, pert=self._pert) |
|
|
|
|
|
def __call__( |
|
|
self, |
|
|
img: NdarrayOrTensor, |
|
|
label: NdarrayOrTensor | None = None, |
|
|
sigma: Sequence[float] | float | Sequence[torch.Tensor] | torch.Tensor = 3.0, |
|
|
rescale_min: float = -1.0, |
|
|
rescale_max: float = 1.0, |
|
|
) -> NdarrayOrTensor: |
|
|
""" |
|
|
Args: |
|
|
img: the image that we want to add new channel to. |
|
|
label: label image to get extreme points from. Shape must be |
|
|
(1, spatial_dim1, [, spatial_dim2, ...]). Doesn't support one-hot labels. |
|
|
sigma: if a list of values, must match the count of spatial dimensions of input data, |
|
|
and apply every value in the list to 1 spatial dimension. if only 1 value provided, |
|
|
use it for all spatial dimensions. |
|
|
rescale_min: minimum value of output data. |
|
|
rescale_max: maximum value of output data. |
|
|
""" |
|
|
if label is None: |
|
|
raise ValueError("This transform requires a label array!") |
|
|
if label.shape[0] != 1: |
|
|
raise ValueError("Only supports single channel labels!") |
|
|
|
|
|
|
|
|
self.randomize(label[0, :]) |
|
|
|
|
|
points_image = extreme_points_to_image( |
|
|
points=self._points, label=label, sigma=sigma, rescale_min=rescale_min, rescale_max=rescale_max |
|
|
) |
|
|
points_image, *_ = convert_to_dst_type(points_image, img) |
|
|
return concatenate((img, points_image), axis=0) |
|
|
|
|
|
|
|
|
class TorchVision: |
|
|
""" |
|
|
This is a wrapper transform for PyTorch TorchVision transform based on the specified transform name and args. |
|
|
As most of the TorchVision transforms only work for PIL image and PyTorch Tensor, this transform expects input |
|
|
data to be PyTorch Tensor, users can easily call `ToTensor` transform to convert a Numpy array to Tensor. |
|
|
|
|
|
""" |
|
|
|
|
|
backend = [TransformBackends.TORCH] |
|
|
|
|
|
def __init__(self, name: str, *args, **kwargs) -> None: |
|
|
""" |
|
|
Args: |
|
|
name: The transform name in TorchVision package. |
|
|
args: parameters for the TorchVision transform. |
|
|
kwargs: parameters for the TorchVision transform. |
|
|
|
|
|
""" |
|
|
super().__init__() |
|
|
self.name = name |
|
|
transform, _ = optional_import("torchvision.transforms", "0.8.0", min_version, name=name) |
|
|
self.trans = transform(*args, **kwargs) |
|
|
|
|
|
def __call__(self, img: NdarrayOrTensor): |
|
|
""" |
|
|
Args: |
|
|
img: PyTorch Tensor data for the TorchVision transform. |
|
|
|
|
|
""" |
|
|
img_t, *_ = convert_data_type(img, torch.Tensor) |
|
|
|
|
|
out = self.trans(img_t) |
|
|
out, *_ = convert_to_dst_type(src=out, dst=img) |
|
|
return out |
|
|
|
|
|
|
|
|
class MapLabelValue: |
|
|
""" |
|
|
Utility to map label values to another set of values. |
|
|
For example, map [3, 2, 1] to [0, 1, 2], [1, 2, 3] -> [0.5, 1.5, 2.5], ["label3", "label2", "label1"] -> [0, 1, 2], |
|
|
[3.5, 2.5, 1.5] -> ["label0", "label1", "label2"], etc. |
|
|
The label data must be numpy array or array-like data and the output data will be numpy array. |
|
|
|
|
|
""" |
|
|
|
|
|
backend = [TransformBackends.NUMPY, TransformBackends.TORCH] |
|
|
|
|
|
def __init__(self, orig_labels: Sequence, target_labels: Sequence, dtype: DtypeLike = np.float32) -> None: |
|
|
""" |
|
|
Args: |
|
|
orig_labels: original labels that map to others. |
|
|
target_labels: expected label values, 1: 1 map to the `orig_labels`. |
|
|
dtype: convert the output data to dtype, default to float32. |
|
|
if dtype is from PyTorch, the transform will use the pytorch backend, else with numpy backend. |
|
|
|
|
|
""" |
|
|
if len(orig_labels) != len(target_labels): |
|
|
raise ValueError("orig_labels and target_labels must have the same length.") |
|
|
|
|
|
self.orig_labels = orig_labels |
|
|
self.target_labels = target_labels |
|
|
self.pair = tuple((o, t) for o, t in zip(self.orig_labels, self.target_labels) if o != t) |
|
|
type_dtype = type(dtype) |
|
|
if getattr(type_dtype, "__module__", "") == "torch": |
|
|
self.use_numpy = False |
|
|
self.dtype = get_equivalent_dtype(dtype, data_type=torch.Tensor) |
|
|
else: |
|
|
self.use_numpy = True |
|
|
self.dtype = get_equivalent_dtype(dtype, data_type=np.ndarray) |
|
|
|
|
|
def __call__(self, img: NdarrayOrTensor): |
|
|
if self.use_numpy: |
|
|
img_np, *_ = convert_data_type(img, np.ndarray) |
|
|
_out_shape = img_np.shape |
|
|
img_flat = img_np.flatten() |
|
|
try: |
|
|
out_flat = img_flat.astype(self.dtype) |
|
|
except ValueError: |
|
|
|
|
|
out_flat = np.zeros(shape=img_flat.shape, dtype=self.dtype) |
|
|
for o, t in self.pair: |
|
|
out_flat[img_flat == o] = t |
|
|
out_t = out_flat.reshape(_out_shape) |
|
|
else: |
|
|
img_t, *_ = convert_data_type(img, torch.Tensor) |
|
|
out_t = img_t.detach().clone().to(self.dtype) |
|
|
for o, t in self.pair: |
|
|
out_t[img_t == o] = t |
|
|
out, *_ = convert_to_dst_type(src=out_t, dst=img, dtype=self.dtype) |
|
|
return out |
|
|
|
|
|
|
|
|
class IntensityStats(Transform): |
|
|
""" |
|
|
Compute statistics for the intensity values of input image and store into the metadata dictionary. |
|
|
For example: if `ops=[lambda x: np.mean(x), "max"]` and `key_prefix="orig"`, may generate below stats: |
|
|
`{"orig_custom_0": 1.5, "orig_max": 3.0}`. |
|
|
|
|
|
Args: |
|
|
ops: expected operations to compute statistics for the intensity. |
|
|
if a string, will map to the predefined operations, supported: ["mean", "median", "max", "min", "std"] |
|
|
mapping to `np.nanmean`, `np.nanmedian`, `np.nanmax`, `np.nanmin`, `np.nanstd`. |
|
|
if a callable function, will execute the function on input image. |
|
|
key_prefix: the prefix to combine with `ops` name to generate the key to store the results in the |
|
|
metadata dictionary. if some `ops` are callable functions, will use "{key_prefix}_custom_{index}" |
|
|
as the key, where index counts from 0. |
|
|
channel_wise: whether to compute statistics for every channel of input image separately. |
|
|
if True, return a list of values for every operation, default to False. |
|
|
|
|
|
""" |
|
|
|
|
|
backend = [TransformBackends.NUMPY] |
|
|
|
|
|
def __init__(self, ops: Sequence[str | Callable], key_prefix: str, channel_wise: bool = False) -> None: |
|
|
self.ops = ensure_tuple(ops) |
|
|
self.key_prefix = key_prefix |
|
|
self.channel_wise = channel_wise |
|
|
|
|
|
def __call__( |
|
|
self, img: NdarrayOrTensor, meta_data: dict | None = None, mask: np.ndarray | None = None |
|
|
) -> tuple[NdarrayOrTensor, dict]: |
|
|
""" |
|
|
Compute statistics for the intensity of input image. |
|
|
|
|
|
Args: |
|
|
img: input image to compute intensity stats. |
|
|
meta_data: metadata dictionary to store the statistics data, if None, will create an empty dictionary. |
|
|
mask: if not None, mask the image to extract only the interested area to compute statistics. |
|
|
mask must have the same shape as input `img`. |
|
|
|
|
|
""" |
|
|
img_np, *_ = convert_data_type(img, np.ndarray) |
|
|
if meta_data is None: |
|
|
meta_data = {} |
|
|
|
|
|
if mask is not None: |
|
|
if mask.shape != img_np.shape: |
|
|
raise ValueError(f"mask must have the same shape as input `img`, got {mask.shape} and {img_np.shape}.") |
|
|
if mask.dtype != bool: |
|
|
raise TypeError(f"mask must be bool array, got type {mask.dtype}.") |
|
|
img_np = img_np[mask] |
|
|
|
|
|
supported_ops = { |
|
|
"mean": np.nanmean, |
|
|
"median": np.nanmedian, |
|
|
"max": np.nanmax, |
|
|
"min": np.nanmin, |
|
|
"std": np.nanstd, |
|
|
} |
|
|
|
|
|
def _compute(op: Callable, data: np.ndarray): |
|
|
if self.channel_wise: |
|
|
return [op(c) for c in data] |
|
|
return op(data) |
|
|
|
|
|
custom_index = 0 |
|
|
for o in self.ops: |
|
|
if isinstance(o, str): |
|
|
o = look_up_option(o, supported_ops.keys()) |
|
|
meta_data[self.key_prefix + "_" + o] = _compute(supported_ops[o], img_np) |
|
|
elif callable(o): |
|
|
meta_data[self.key_prefix + "_custom_" + str(custom_index)] = _compute(o, img_np) |
|
|
custom_index += 1 |
|
|
else: |
|
|
raise ValueError("ops must be key string for predefined operations or callable function.") |
|
|
|
|
|
return img, meta_data |
|
|
|
|
|
|
|
|
class ToDevice(Transform): |
|
|
""" |
|
|
Move PyTorch Tensor to the specified device. |
|
|
It can help cache data into GPU and execute following logic on GPU directly. |
|
|
|
|
|
Note: |
|
|
If moving data to GPU device in the multi-processing workers of DataLoader, may got below CUDA error: |
|
|
"RuntimeError: Cannot re-initialize CUDA in forked subprocess. To use CUDA with multiprocessing, |
|
|
you must use the 'spawn' start method." |
|
|
So usually suggest to set `num_workers=0` in the `DataLoader` or `ThreadDataLoader`. |
|
|
|
|
|
""" |
|
|
|
|
|
backend = [TransformBackends.TORCH] |
|
|
|
|
|
def __init__(self, device: torch.device | str, **kwargs) -> None: |
|
|
""" |
|
|
Args: |
|
|
device: target device to move the Tensor, for example: "cuda:1". |
|
|
kwargs: other args for the PyTorch `Tensor.to()` API, for more details: |
|
|
https://pytorch.org/docs/stable/generated/torch.Tensor.to.html. |
|
|
|
|
|
""" |
|
|
self.device = device |
|
|
self.kwargs = kwargs |
|
|
|
|
|
def __call__(self, img: torch.Tensor): |
|
|
if not isinstance(img, torch.Tensor): |
|
|
raise ValueError("img must be PyTorch Tensor, consider converting img by `EnsureType` transform first.") |
|
|
|
|
|
return img.to(self.device, **self.kwargs) |
|
|
|
|
|
|
|
|
class CuCIM(Transform): |
|
|
""" |
|
|
Wrap a non-randomized cuCIM transform, defined based on the transform name and args. |
|
|
For randomized transforms use :py:class:`monai.transforms.RandCuCIM`. |
|
|
|
|
|
Args: |
|
|
name: the transform name in CuCIM package |
|
|
args: parameters for the CuCIM transform |
|
|
kwargs: parameters for the CuCIM transform |
|
|
|
|
|
Note: |
|
|
CuCIM transform only work with CuPy arrays, so this transform expects input data to be `cupy.ndarray`. |
|
|
Users can call `ToCuPy` transform to convert a numpy array or torch tensor to cupy array. |
|
|
""" |
|
|
|
|
|
def __init__(self, name: str, *args, **kwargs) -> None: |
|
|
super().__init__() |
|
|
self.name = name |
|
|
self.transform, _ = optional_import("cucim.core.operations.expose.transform", name=name) |
|
|
self.args = args |
|
|
self.kwargs = kwargs |
|
|
|
|
|
def __call__(self, data): |
|
|
""" |
|
|
Args: |
|
|
data: a CuPy array (`cupy.ndarray`) for the cuCIM transform |
|
|
|
|
|
Returns: |
|
|
`cupy.ndarray` |
|
|
|
|
|
""" |
|
|
return self.transform(data, *self.args, **self.kwargs) |
|
|
|
|
|
|
|
|
class RandCuCIM(CuCIM, RandomizableTrait): |
|
|
""" |
|
|
Wrap a randomized cuCIM transform, defined based on the transform name and args |
|
|
For deterministic non-randomized transforms use :py:class:`monai.transforms.CuCIM`. |
|
|
|
|
|
Args: |
|
|
name: the transform name in CuCIM package. |
|
|
args: parameters for the CuCIM transform. |
|
|
kwargs: parameters for the CuCIM transform. |
|
|
|
|
|
Note: |
|
|
- CuCIM transform only work with CuPy arrays, so this transform expects input data to be `cupy.ndarray`. |
|
|
Users can call `ToCuPy` transform to convert a numpy array or torch tensor to cupy array. |
|
|
- If the random factor of the underlying cuCIM transform is not derived from `self.R`, |
|
|
the results may not be deterministic. See Also: :py:class:`monai.transforms.Randomizable`. |
|
|
""" |
|
|
|
|
|
def __init__(self, name: str, *args, **kwargs) -> None: |
|
|
CuCIM.__init__(self, name, *args, **kwargs) |
|
|
|
|
|
|
|
|
class AddCoordinateChannels(Transform): |
|
|
""" |
|
|
Appends additional channels encoding coordinates of the input. Useful when e.g. training using patch-based sampling, |
|
|
to allow feeding of the patch's location into the network. |
|
|
|
|
|
This can be seen as a input-only version of CoordConv: |
|
|
|
|
|
Liu, R. et al. An Intriguing Failing of Convolutional Neural Networks and the CoordConv Solution, NeurIPS 2018. |
|
|
|
|
|
Args: |
|
|
spatial_dims: the spatial dimensions that are to have their coordinates encoded in a channel and |
|
|
appended to the input image. E.g., `(0, 1, 2)` represents `H, W, D` dims and append three channels |
|
|
to the input image, encoding the coordinates of the input's three spatial dimensions. |
|
|
|
|
|
""" |
|
|
|
|
|
backend = [TransformBackends.NUMPY] |
|
|
|
|
|
def __init__(self, spatial_dims: Sequence[int]) -> None: |
|
|
self.spatial_dims = spatial_dims |
|
|
|
|
|
def __call__(self, img: NdarrayOrTensor) -> NdarrayOrTensor: |
|
|
""" |
|
|
Args: |
|
|
img: data to be transformed, assuming `img` is channel first. |
|
|
""" |
|
|
if max(self.spatial_dims) > img.ndim - 2 or min(self.spatial_dims) < 0: |
|
|
raise ValueError(f"`spatial_dims` values must be within [0, {img.ndim - 2}]") |
|
|
|
|
|
spatial_size = img.shape[1:] |
|
|
coord_channels = np.array(np.meshgrid(*tuple(np.linspace(-0.5, 0.5, s) for s in spatial_size), indexing="ij")) |
|
|
coord_channels, *_ = convert_to_dst_type(coord_channels, img) |
|
|
coord_channels = coord_channels[list(self.spatial_dims)] |
|
|
return concatenate((img, coord_channels), axis=0) |
|
|
|
|
|
|
|
|
class ImageFilter(Transform): |
|
|
""" |
|
|
Applies a convolution filter to the input image. |
|
|
|
|
|
Args: |
|
|
filter: |
|
|
A string specifying the filter, a custom filter as ``torch.Tenor`` or ``np.ndarray`` or a ``nn.Module``. |
|
|
Available options for string are: ``mean``, ``laplace``, ``elliptical``, ``sobel``, ``sharpen``, ``median``, ``gauss`` |
|
|
See below for short explanations on every filter. |
|
|
filter_size: |
|
|
A single integer value specifying the size of the quadratic or cubic filter. |
|
|
Computational complexity scales to the power of 2 (2D filter) or 3 (3D filter), which |
|
|
should be considered when choosing filter size. |
|
|
kwargs: |
|
|
Additional arguments passed to filter function, required by ``sobel`` and ``gauss``. |
|
|
See below for details. |
|
|
|
|
|
Raises: |
|
|
ValueError: When ``filter_size`` is not an uneven integer |
|
|
ValueError: When ``filter`` is an array and ``ndim`` is not in [1,2,3] |
|
|
ValueError: When ``filter`` is an array and any dimension has an even shape |
|
|
NotImplementedError: When ``filter`` is a string and not in ``self.supported_filters`` |
|
|
KeyError: When necessary ``kwargs`` are not passed to a filter that requires additional arguments. |
|
|
|
|
|
|
|
|
**Mean Filtering:** ``filter='mean'`` |
|
|
|
|
|
Mean filtering can smooth edges and remove aliasing artifacts in an segmentation image. |
|
|
See also py:func:`monai.networks.layers.simplelayers.MeanFilter` |
|
|
Example 2D filter (5 x 5):: |
|
|
|
|
|
[[1, 1, 1, 1, 1], |
|
|
[1, 1, 1, 1, 1], |
|
|
[1, 1, 1, 1, 1], |
|
|
[1, 1, 1, 1, 1], |
|
|
[1, 1, 1, 1, 1]] |
|
|
|
|
|
If smoothing labels with this filter, ensure they are in one-hot format. |
|
|
|
|
|
**Outline Detection:** ``filter='laplace'`` |
|
|
|
|
|
Laplacian filtering for outline detection in images. Can be used to transform labels to contours. |
|
|
See also py:func:`monai.networks.layers.simplelayers.LaplaceFilter` |
|
|
|
|
|
Example 2D filter (5x5):: |
|
|
|
|
|
[[-1., -1., -1., -1., -1.], |
|
|
[-1., -1., -1., -1., -1.], |
|
|
[-1., -1., 24., -1., -1.], |
|
|
[-1., -1., -1., -1., -1.], |
|
|
[-1., -1., -1., -1., -1.]] |
|
|
|
|
|
|
|
|
**Dilation:** ``filter='elliptical'`` |
|
|
|
|
|
An elliptical filter can be used to dilate labels or label-contours. |
|
|
Example 2D filter (5x5):: |
|
|
|
|
|
[[0., 0., 1., 0., 0.], |
|
|
[1., 1., 1., 1., 1.], |
|
|
[1., 1., 1., 1., 1.], |
|
|
[1., 1., 1., 1., 1.], |
|
|
[0., 0., 1., 0., 0.]] |
|
|
|
|
|
|
|
|
**Edge Detection:** ``filter='sobel'`` |
|
|
|
|
|
This filter allows for additional arguments passed as ``kwargs`` during initialization. |
|
|
See also py:func:`monai.transforms.post.SobelGradients` |
|
|
|
|
|
*kwargs* |
|
|
|
|
|
* ``spatial_axes``: the axes that define the direction of the gradient to be calculated. |
|
|
It calculates the gradient along each of the provide axis. |
|
|
By default it calculate the gradient for all spatial axes. |
|
|
* ``normalize_kernels``: if normalize the Sobel kernel to provide proper gradients. Defaults to True. |
|
|
* ``normalize_gradients``: if normalize the output gradient to 0 and 1. Defaults to False. |
|
|
* ``padding_mode``: the padding mode of the image when convolving with Sobel kernels. Defaults to ``"reflect"``. |
|
|
Acceptable values are ``'zeros'``, ``'reflect'``, ``'replicate'`` or ``'circular'``. |
|
|
See ``torch.nn.Conv1d()`` for more information. |
|
|
* ``dtype``: kernel data type (torch.dtype). Defaults to ``torch.float32``. |
|
|
|
|
|
|
|
|
**Sharpening:** ``filter='sharpen'`` |
|
|
|
|
|
Sharpen an image with a 2D or 3D filter. |
|
|
Example 2D filter (5x5):: |
|
|
|
|
|
[[ 0., 0., -1., 0., 0.], |
|
|
[-1., -1., -1., -1., -1.], |
|
|
[-1., -1., 17., -1., -1.], |
|
|
[-1., -1., -1., -1., -1.], |
|
|
[ 0., 0., -1., 0., 0.]] |
|
|
|
|
|
|
|
|
**Gaussian Smooth:** ``filter='gauss'`` |
|
|
|
|
|
Blur/smooth an image with 2D or 3D gaussian filter. |
|
|
This filter requires additional arguments passed as ``kwargs`` during initialization. |
|
|
See also py:func:`monai.networks.layers.simplelayers.GaussianFilter` |
|
|
|
|
|
*kwargs* |
|
|
|
|
|
* ``sigma``: std. could be a single value, or spatial_dims number of values. |
|
|
* ``truncated``: spreads how many stds. |
|
|
* ``approx``: discrete Gaussian kernel type, available options are "erf", "sampled", and "scalespace". |
|
|
|
|
|
|
|
|
**Median Filter:** ``filter='median'`` |
|
|
|
|
|
Blur an image with 2D or 3D median filter to remove noise. |
|
|
Useful in image preprocessing to improve results of later processing. |
|
|
See also py:func:`monai.networks.layers.simplelayers.MedianFilter` |
|
|
|
|
|
|
|
|
**Savitzky Golay Filter:** ``filter = 'savitzky_golay'`` |
|
|
|
|
|
Convolve a Tensor along a particular axis with a Savitzky-Golay kernel. |
|
|
This filter requires additional arguments passed as ``kwargs`` during initialization. |
|
|
See also py:func:`monai.networks.layers.simplelayers.SavitzkyGolayFilter` |
|
|
|
|
|
*kwargs* |
|
|
|
|
|
* ``order``: Order of the polynomial to fit to each window, must be less than ``window_length``. |
|
|
* ``axis``: (optional): Axis along which to apply the filter kernel. Default 2 (first spatial dimension). |
|
|
* ``mode``: (string, optional): padding mode passed to convolution class. ``'zeros'``, ``'reflect'``, ``'replicate'`` or |
|
|
``'circular'``. Default: ``'zeros'``. See torch.nn.Conv1d() for more information. |
|
|
|
|
|
""" |
|
|
|
|
|
backend = [TransformBackends.TORCH, TransformBackends.NUMPY] |
|
|
supported_filters = sorted( |
|
|
["mean", "laplace", "elliptical", "sobel", "sharpen", "median", "gauss", "savitzky_golay"] |
|
|
) |
|
|
|
|
|
def __init__(self, filter: str | NdarrayOrTensor | nn.Module, filter_size: int | None = None, **kwargs) -> None: |
|
|
self._check_filter_format(filter, filter_size) |
|
|
self._check_kwargs_are_present(filter, **kwargs) |
|
|
self.filter = filter |
|
|
self.filter_size = filter_size |
|
|
self.additional_args_for_filter = kwargs |
|
|
|
|
|
def __call__( |
|
|
self, img: NdarrayOrTensor, meta_dict: dict | None = None, applied_operations: list | None = None |
|
|
) -> NdarrayOrTensor: |
|
|
""" |
|
|
Args: |
|
|
img: torch tensor data to apply filter to with shape: [channels, height, width[, depth]] |
|
|
meta_dict: An optional dictionary with metadata |
|
|
applied_operations: An optional list of operations that have been applied to the data |
|
|
|
|
|
Returns: |
|
|
A MetaTensor with the same shape as `img` and identical metadata |
|
|
""" |
|
|
if isinstance(img, MetaTensor): |
|
|
meta_dict = img.meta |
|
|
applied_operations = img.applied_operations |
|
|
|
|
|
img_, prev_type, device = convert_data_type(img, torch.Tensor) |
|
|
ndim = img_.ndim - 1 |
|
|
|
|
|
if isinstance(self.filter, str): |
|
|
self.filter = self._get_filter_from_string(self.filter, self.filter_size, ndim) |
|
|
elif isinstance(self.filter, (torch.Tensor, np.ndarray)): |
|
|
self.filter = ApplyFilter(self.filter) |
|
|
|
|
|
img_ = self._apply_filter(img_) |
|
|
if meta_dict is not None or applied_operations is not None: |
|
|
img_ = MetaTensor(img_, meta=meta_dict, applied_operations=applied_operations) |
|
|
else: |
|
|
img_, *_ = convert_data_type(img_, prev_type, device) |
|
|
return img_ |
|
|
|
|
|
def _check_all_values_uneven(self, x: tuple) -> None: |
|
|
for value in x: |
|
|
if value % 2 == 0: |
|
|
raise ValueError(f"Only uneven filters are supported, but filter size is {x}") |
|
|
|
|
|
def _check_filter_format(self, filter: str | NdarrayOrTensor | nn.Module, filter_size: int | None = None) -> None: |
|
|
if isinstance(filter, str): |
|
|
if not filter_size: |
|
|
raise ValueError("`filter_size` must be specified when specifying filters by string.") |
|
|
if filter_size % 2 == 0: |
|
|
raise ValueError("`filter_size` should be a single uneven integer.") |
|
|
if filter not in self.supported_filters: |
|
|
raise NotImplementedError(f"{filter}. Supported filters are {self.supported_filters}.") |
|
|
elif isinstance(filter, (torch.Tensor, np.ndarray)): |
|
|
if filter.ndim not in [1, 2, 3]: |
|
|
raise ValueError("Only 1D, 2D, and 3D filters are supported.") |
|
|
self._check_all_values_uneven(filter.shape) |
|
|
elif not isinstance(filter, (nn.Module, Transform)): |
|
|
raise TypeError( |
|
|
f"{type(filter)} is not supported." |
|
|
"Supported types are `class 'str'`, `class 'torch.Tensor'`, `class 'np.ndarray'`, " |
|
|
"`class 'torch.nn.modules.module.Module'`, `class 'monai.transforms.Transform'`" |
|
|
) |
|
|
|
|
|
def _check_kwargs_are_present(self, filter: str | NdarrayOrTensor | nn.Module, **kwargs: Any) -> None: |
|
|
""" |
|
|
Perform sanity checks on the kwargs if the filter contains the required keys. |
|
|
If the filter is ``gauss``, kwargs should contain ``sigma``. |
|
|
If the filter is ``savitzky_golay``, kwargs should contain ``order``. |
|
|
|
|
|
Args: |
|
|
filter: A string specifying the filter, a custom filter as ``torch.Tenor`` or ``np.ndarray`` or a ``nn.Module``. |
|
|
kwargs: additional arguments defining the filter. |
|
|
|
|
|
Raises: |
|
|
KeyError if the filter doesn't contain the requirement key. |
|
|
""" |
|
|
|
|
|
if not isinstance(filter, str): |
|
|
return |
|
|
if filter == "gauss" and "sigma" not in kwargs.keys(): |
|
|
raise KeyError("`filter='gauss', requires the additional keyword argument `sigma`") |
|
|
if filter == "savitzky_golay" and "order" not in kwargs.keys(): |
|
|
raise KeyError("`filter='savitzky_golay', requires the additional keyword argument `order`") |
|
|
|
|
|
def _get_filter_from_string(self, filter: str, size: int, ndim: int) -> nn.Module | Callable: |
|
|
if filter == "mean": |
|
|
return MeanFilter(ndim, size) |
|
|
elif filter == "laplace": |
|
|
return LaplaceFilter(ndim, size) |
|
|
elif filter == "elliptical": |
|
|
return EllipticalFilter(ndim, size) |
|
|
elif filter == "sobel": |
|
|
from monai.transforms.post.array import SobelGradients |
|
|
|
|
|
allowed_keys = SobelGradients.__init__.__annotations__.keys() |
|
|
kwargs = {k: v for k, v in self.additional_args_for_filter.items() if k in allowed_keys} |
|
|
return SobelGradients(size, **kwargs) |
|
|
elif filter == "sharpen": |
|
|
return SharpenFilter(ndim, size) |
|
|
elif filter == "gauss": |
|
|
allowed_keys = GaussianFilter.__init__.__annotations__.keys() |
|
|
kwargs = {k: v for k, v in self.additional_args_for_filter.items() if k in allowed_keys} |
|
|
return GaussianFilter(ndim, **kwargs) |
|
|
elif filter == "median": |
|
|
return partial(median_filter, kernel_size=size, spatial_dims=ndim) |
|
|
elif filter == "savitzky_golay": |
|
|
allowed_keys = SavitzkyGolayFilter.__init__.__annotations__.keys() |
|
|
kwargs = {k: v for k, v in self.additional_args_for_filter.items() if k in allowed_keys} |
|
|
return SavitzkyGolayFilter(size, **kwargs) |
|
|
else: |
|
|
raise NotImplementedError(f"Filter {filter} not implemented") |
|
|
|
|
|
def _apply_filter(self, img: torch.Tensor) -> torch.Tensor: |
|
|
if isinstance(self.filter, Transform): |
|
|
img = self.filter(img) |
|
|
else: |
|
|
img = self.filter(img.unsqueeze(0)) |
|
|
img = img[0] |
|
|
return img |
|
|
|
|
|
|
|
|
class RandImageFilter(RandomizableTransform): |
|
|
""" |
|
|
Randomly apply a convolutional filter to the input data. |
|
|
|
|
|
Args: |
|
|
filter: |
|
|
A string specifying the filter or a custom filter as `torch.Tenor` or `np.ndarray`. |
|
|
Available options are: `mean`, `laplace`, `elliptical`, `gaussian`` |
|
|
See below for short explanations on every filter. |
|
|
filter_size: |
|
|
A single integer value specifying the size of the quadratic or cubic filter. |
|
|
Computational complexity scales to the power of 2 (2D filter) or 3 (3D filter), which |
|
|
should be considered when choosing filter size. |
|
|
prob: |
|
|
Probability the transform is applied to the data |
|
|
""" |
|
|
|
|
|
backend = ImageFilter.backend |
|
|
|
|
|
def __init__( |
|
|
self, filter: str | NdarrayOrTensor, filter_size: int | None = None, prob: float = 0.1, **kwargs |
|
|
) -> None: |
|
|
super().__init__(prob) |
|
|
self.filter = ImageFilter(filter, filter_size, **kwargs) |
|
|
|
|
|
def __call__(self, img: NdarrayOrTensor, meta_dict: Mapping | None = None) -> NdarrayOrTensor: |
|
|
""" |
|
|
Args: |
|
|
img: torch tensor data to apply filter to with shape: [channels, height, width[, depth]] |
|
|
meta_dict: An optional dictionary with metadata |
|
|
kwargs: optional arguments required by specific filters. E.g. `sigma`if filter is `gauss`. |
|
|
see py:func:`monai.transforms.utility.array.ImageFilter` for more details |
|
|
|
|
|
Returns: |
|
|
A MetaTensor with the same shape as `img` and identical metadata |
|
|
""" |
|
|
self.randomize(None) |
|
|
if self._do_transform: |
|
|
img = self.filter(img) |
|
|
return img |
|
|
|