| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | """ |
| | A collection of dictionary-based wrappers around the "vanilla" transforms for intensity adjustment |
| | defined in :py:class:`monai.transforms.intensity.array`. |
| | |
| | Class names are ended with 'd' to denote dictionary-based transforms. |
| | """ |
| |
|
| | from collections.abc import Iterable |
| | from typing import Any, Dict, Hashable, Mapping, Optional, Sequence, Tuple, Union |
| |
|
| | import numpy as np |
| |
|
| | from monai.config import KeysCollection |
| | from monai.transforms.compose import MapTransform, Randomizable |
| | from monai.transforms.intensity.array import ( |
| | AdjustContrast, |
| | GaussianSharpen, |
| | GaussianSmooth, |
| | MaskIntensity, |
| | NormalizeIntensity, |
| | ScaleIntensity, |
| | ScaleIntensityRange, |
| | ScaleIntensityRangePercentiles, |
| | ShiftIntensity, |
| | ThresholdIntensity, |
| | ) |
| | from monai.utils import ensure_tuple_size |
| |
|
| |
|
| | class RandGaussianNoised(Randomizable, MapTransform): |
| | """ |
| | Dictionary-based version :py:class:`monai.transforms.RandGaussianNoise`. |
| | Add Gaussian noise to image. This transform assumes all the expected fields have same shape. |
| | |
| | Args: |
| | keys: keys of the corresponding items to be transformed. |
| | See also: :py:class:`monai.transforms.compose.MapTransform` |
| | prob: Probability to add Gaussian noise. |
| | mean: Mean or “centre” of the distribution. |
| | std: Standard deviation (spread) of distribution. |
| | """ |
| |
|
| | def __init__( |
| | self, keys: KeysCollection, prob: float = 0.1, mean: Union[Sequence[float], float] = 0.0, std: float = 0.1 |
| | ) -> None: |
| | super().__init__(keys) |
| | self.prob = prob |
| | self.mean = ensure_tuple_size(mean, len(self.keys)) |
| | self.std = std |
| | self._do_transform = False |
| | self._noise: Optional[np.ndarray] = None |
| |
|
| | def randomize(self, im_shape: Sequence[int]) -> None: |
| | self._do_transform = self.R.random() < self.prob |
| | self._noise = self.R.normal(self.mean, self.R.uniform(0, self.std), size=im_shape) |
| |
|
| | def __call__(self, data: Mapping[Hashable, np.ndarray]) -> Dict[Hashable, np.ndarray]: |
| | d = dict(data) |
| |
|
| | image_shape = d[self.keys[0]].shape |
| | self.randomize(image_shape) |
| | assert self._noise is not None |
| | if not self._do_transform: |
| | return d |
| | for key in self.keys: |
| | d[key] = d[key] + self._noise.astype(d[key].dtype) |
| | return d |
| |
|
| |
|
| | class ShiftIntensityd(MapTransform): |
| | """ |
| | Dictionary-based wrapper of :py:class:`monai.transforms.ShiftIntensity`. |
| | """ |
| |
|
| | def __init__(self, keys: KeysCollection, offset: float) -> None: |
| | """ |
| | Args: |
| | keys: keys of the corresponding items to be transformed. |
| | See also: :py:class:`monai.transforms.compose.MapTransform` |
| | offset: offset value to shift the intensity of image. |
| | """ |
| | super().__init__(keys) |
| | self.shifter = ShiftIntensity(offset) |
| |
|
| | def __call__(self, data: Mapping[Hashable, np.ndarray]) -> Dict[Hashable, np.ndarray]: |
| | d = dict(data) |
| | for key in self.keys: |
| | d[key] = self.shifter(d[key]) |
| | return d |
| |
|
| |
|
| | class RandShiftIntensityd(Randomizable, MapTransform): |
| | """ |
| | Dictionary-based version :py:class:`monai.transforms.RandShiftIntensity`. |
| | """ |
| |
|
| | def __init__(self, keys: KeysCollection, offsets: Union[Tuple[float, float], float], prob: float = 0.1) -> None: |
| | """ |
| | Args: |
| | keys: keys of the corresponding items to be transformed. |
| | See also: :py:class:`monai.transforms.compose.MapTransform` |
| | offsets: offset range to randomly shift. |
| | if single number, offset value is picked from (-offsets, offsets). |
| | prob: probability of rotating. |
| | (Default 0.1, with 10% probability it returns a rotated array.) |
| | """ |
| | super().__init__(keys) |
| |
|
| | if isinstance(offsets, (int, float)): |
| | self.offsets = (min(-offsets, offsets), max(-offsets, offsets)) |
| | else: |
| | assert len(offsets) == 2, "offsets should be a number or pair of numbers." |
| | self.offsets = (min(offsets), max(offsets)) |
| |
|
| | self.prob = prob |
| | self._do_transform = False |
| |
|
| | def randomize(self, data: Optional[Any] = None) -> None: |
| | self._offset = self.R.uniform(low=self.offsets[0], high=self.offsets[1]) |
| | self._do_transform = self.R.random() < self.prob |
| |
|
| | def __call__(self, data: Mapping[Hashable, np.ndarray]) -> Dict[Hashable, np.ndarray]: |
| | d = dict(data) |
| | self.randomize() |
| | if not self._do_transform: |
| | return d |
| | shifter = ShiftIntensity(self._offset) |
| | for key in self.keys: |
| | d[key] = shifter(d[key]) |
| | return d |
| |
|
| |
|
| | class ScaleIntensityd(MapTransform): |
| | """ |
| | Dictionary-based wrapper of :py:class:`monai.transforms.ScaleIntensity`. |
| | Scale the intensity of input image to the given value range (minv, maxv). |
| | If `minv` and `maxv` not provided, use `factor` to scale image by ``v = v * (1 + factor)``. |
| | """ |
| |
|
| | def __init__( |
| | self, keys: KeysCollection, minv: float = 0.0, maxv: float = 1.0, factor: Optional[float] = None |
| | ) -> None: |
| | """ |
| | Args: |
| | keys: keys of the corresponding items to be transformed. |
| | See also: :py:class:`monai.transforms.compose.MapTransform` |
| | minv: minimum value of output data. |
| | maxv: maximum value of output data. |
| | factor: factor scale by ``v = v * (1 + factor)``. |
| | |
| | """ |
| | super().__init__(keys) |
| | self.scaler = ScaleIntensity(minv, maxv, factor) |
| |
|
| | def __call__(self, data: Mapping[Hashable, np.ndarray]) -> Dict[Hashable, np.ndarray]: |
| | d = dict(data) |
| | for key in self.keys: |
| | d[key] = self.scaler(d[key]) |
| | return d |
| |
|
| |
|
| | class RandScaleIntensityd(Randomizable, MapTransform): |
| | """ |
| | Dictionary-based version :py:class:`monai.transforms.RandScaleIntensity`. |
| | """ |
| |
|
| | def __init__(self, keys: KeysCollection, factors: Union[Tuple[float, float], float], prob: float = 0.1) -> None: |
| | """ |
| | Args: |
| | keys: keys of the corresponding items to be transformed. |
| | See also: :py:class:`monai.transforms.compose.MapTransform` |
| | factors: factor range to randomly scale by ``v = v * (1 + factor)``. |
| | if single number, factor value is picked from (-factors, factors). |
| | prob: probability of rotating. |
| | (Default 0.1, with 10% probability it returns a rotated array.) |
| | |
| | """ |
| | super().__init__(keys) |
| |
|
| | if isinstance(factors, (int, float)): |
| | self.factors = (min(-factors, factors), max(-factors, factors)) |
| | else: |
| | assert len(factors) == 2, "factors should be a number or pair of numbers." |
| | self.factors = (min(factors), max(factors)) |
| |
|
| | self.prob = prob |
| | self._do_transform = False |
| |
|
| | def randomize(self, data: Optional[Any] = None) -> None: |
| | self.factor = self.R.uniform(low=self.factors[0], high=self.factors[1]) |
| | self._do_transform = self.R.random() < self.prob |
| |
|
| | def __call__(self, data: Mapping[Hashable, np.ndarray]) -> Dict[Hashable, np.ndarray]: |
| | d = dict(data) |
| | self.randomize() |
| | if not self._do_transform: |
| | return d |
| | scaler = ScaleIntensity(minv=None, maxv=None, factor=self.factor) |
| | for key in self.keys: |
| | d[key] = scaler(d[key]) |
| | return d |
| |
|
| |
|
| | class NormalizeIntensityd(MapTransform): |
| | """ |
| | Dictionary-based wrapper of :py:class:`monai.transforms.NormalizeIntensity`. |
| | This transform can normalize only non-zero values or entire image, and can also calculate |
| | mean and std on each channel separately. |
| | |
| | Args: |
| | keys: keys of the corresponding items to be transformed. |
| | See also: monai.transforms.MapTransform |
| | subtrahend: the amount to subtract by (usually the mean) |
| | divisor: the amount to divide by (usually the standard deviation) |
| | nonzero: whether only normalize non-zero values. |
| | channel_wise: if using calculated mean and std, calculate on each channel separately |
| | or calculate on the entire image directly. |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | keys: KeysCollection, |
| | subtrahend: Optional[np.ndarray] = None, |
| | divisor: Optional[np.ndarray] = None, |
| | nonzero: bool = False, |
| | channel_wise: bool = False, |
| | ) -> None: |
| | super().__init__(keys) |
| | self.normalizer = NormalizeIntensity(subtrahend, divisor, nonzero, channel_wise) |
| |
|
| | def __call__(self, data: Mapping[Hashable, np.ndarray]) -> Dict[Hashable, np.ndarray]: |
| | d = dict(data) |
| | for key in self.keys: |
| | d[key] = self.normalizer(d[key]) |
| | return d |
| |
|
| |
|
| | class ThresholdIntensityd(MapTransform): |
| | """ |
| | Dictionary-based wrapper of :py:class:`monai.transforms.ThresholdIntensity`. |
| | |
| | Args: |
| | keys: keys of the corresponding items to be transformed. |
| | See also: monai.transforms.MapTransform |
| | threshold: the threshold to filter intensity values. |
| | above: filter values above the threshold or below the threshold, default is True. |
| | cval: value to fill the remaining parts of the image, default is 0. |
| | """ |
| |
|
| | def __init__(self, keys: KeysCollection, threshold: float, above: bool = True, cval: float = 0.0) -> None: |
| | super().__init__(keys) |
| | self.filter = ThresholdIntensity(threshold, above, cval) |
| |
|
| | def __call__(self, data: Mapping[Hashable, np.ndarray]) -> Dict[Hashable, np.ndarray]: |
| | d = dict(data) |
| | for key in self.keys: |
| | d[key] = self.filter(d[key]) |
| | return d |
| |
|
| |
|
| | class ScaleIntensityRanged(MapTransform): |
| | """ |
| | Dictionary-based wrapper of :py:class:`monai.transforms.ScaleIntensityRange`. |
| | |
| | Args: |
| | keys: keys of the corresponding items to be transformed. |
| | See also: monai.transforms.MapTransform |
| | a_min: intensity original range min. |
| | a_max: intensity original range max. |
| | b_min: intensity target range min. |
| | b_max: intensity target range max. |
| | clip: whether to perform clip after scaling. |
| | """ |
| |
|
| | def __init__( |
| | self, keys: KeysCollection, a_min: float, a_max: float, b_min: float, b_max: float, clip: bool = False |
| | ) -> None: |
| | super().__init__(keys) |
| | self.scaler = ScaleIntensityRange(a_min, a_max, b_min, b_max, clip) |
| |
|
| | def __call__(self, data: Mapping[Hashable, np.ndarray]) -> Dict[Hashable, np.ndarray]: |
| | d = dict(data) |
| | for key in self.keys: |
| | d[key] = self.scaler(d[key]) |
| | return d |
| |
|
| |
|
| | class AdjustContrastd(MapTransform): |
| | """ |
| | Dictionary-based wrapper of :py:class:`monai.transforms.AdjustContrast`. |
| | Changes image intensity by gamma. Each pixel/voxel intensity is updated as: |
| | |
| | `x = ((x - min) / intensity_range) ^ gamma * intensity_range + min` |
| | |
| | Args: |
| | keys: keys of the corresponding items to be transformed. |
| | See also: monai.transforms.MapTransform |
| | gamma: gamma value to adjust the contrast as function. |
| | """ |
| |
|
| | def __init__(self, keys: KeysCollection, gamma: float) -> None: |
| | super().__init__(keys) |
| | self.adjuster = AdjustContrast(gamma) |
| |
|
| | def __call__(self, data: Mapping[Hashable, np.ndarray]) -> Dict[Hashable, np.ndarray]: |
| | d = dict(data) |
| | for key in self.keys: |
| | d[key] = self.adjuster(d[key]) |
| | return d |
| |
|
| |
|
| | class RandAdjustContrastd(Randomizable, MapTransform): |
| | """ |
| | Dictionary-based version :py:class:`monai.transforms.RandAdjustContrast`. |
| | Randomly changes image intensity by gamma. Each pixel/voxel intensity is updated as: |
| | |
| | `x = ((x - min) / intensity_range) ^ gamma * intensity_range + min` |
| | |
| | Args: |
| | keys: keys of the corresponding items to be transformed. |
| | See also: monai.transforms.MapTransform |
| | prob: Probability of adjustment. |
| | gamma: Range of gamma values. |
| | If single number, value is picked from (0.5, gamma), default is (0.5, 4.5). |
| | """ |
| |
|
| | def __init__( |
| | self, keys: KeysCollection, prob: float = 0.1, gamma: Union[Tuple[float, float], float] = (0.5, 4.5) |
| | ) -> None: |
| | super().__init__(keys) |
| | self.prob: float = prob |
| |
|
| | if isinstance(gamma, (int, float)): |
| | assert gamma > 0.5, "if gamma is single number, must greater than 0.5 and value is picked from (0.5, gamma)" |
| | self.gamma = (0.5, gamma) |
| | else: |
| | assert len(gamma) == 2, "gamma should be a number or pair of numbers." |
| | self.gamma = (min(gamma), max(gamma)) |
| |
|
| | self._do_transform = False |
| | self.gamma_value: Optional[float] = None |
| |
|
| | def randomize(self, data: Optional[Any] = None) -> None: |
| | self._do_transform = self.R.random_sample() < self.prob |
| | self.gamma_value = self.R.uniform(low=self.gamma[0], high=self.gamma[1]) |
| |
|
| | def __call__(self, data: Mapping[Hashable, np.ndarray]) -> Dict[Hashable, np.ndarray]: |
| | d = dict(data) |
| | self.randomize() |
| | assert self.gamma_value is not None |
| | if not self._do_transform: |
| | return d |
| | adjuster = AdjustContrast(self.gamma_value) |
| | for key in self.keys: |
| | d[key] = adjuster(d[key]) |
| | return d |
| |
|
| |
|
| | class ScaleIntensityRangePercentilesd(MapTransform): |
| | """ |
| | Dictionary-based wrapper of :py:class:`monai.transforms.ScaleIntensityRangePercentiles`. |
| | |
| | Args: |
| | keys: keys of the corresponding items to be transformed. |
| | See also: monai.transforms.MapTransform |
| | lower: lower percentile. |
| | upper: upper percentile. |
| | b_min: intensity target range min. |
| | b_max: intensity target range max. |
| | clip: whether to perform clip after scaling. |
| | relative: whether to scale to the corresponding percentiles of [b_min, b_max] |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | keys: KeysCollection, |
| | lower: float, |
| | upper: float, |
| | b_min: float, |
| | b_max: float, |
| | clip: bool = False, |
| | relative: bool = False, |
| | ) -> None: |
| | super().__init__(keys) |
| | self.scaler = ScaleIntensityRangePercentiles(lower, upper, b_min, b_max, clip, relative) |
| |
|
| | def __call__(self, data: Mapping[Hashable, np.ndarray]) -> Dict[Hashable, np.ndarray]: |
| | d = dict(data) |
| | for key in self.keys: |
| | d[key] = self.scaler(d[key]) |
| | return d |
| |
|
| |
|
| | class MaskIntensityd(MapTransform): |
| | """ |
| | Dictionary-based wrapper of :py:class:`monai.transforms.MaskIntensity`. |
| | |
| | Args: |
| | keys: keys of the corresponding items to be transformed. |
| | See also: :py:class:`monai.transforms.compose.MapTransform` |
| | mask_data: if mask data is single channel, apply to evey channel |
| | of input image. if multiple channels, the channel number must |
| | match input data. mask_data will be converted to `bool` values |
| | by `mask_data > 0` before applying transform to input image. |
| | |
| | """ |
| |
|
| | def __init__(self, keys: KeysCollection, mask_data: np.ndarray) -> None: |
| | super().__init__(keys) |
| | self.converter = MaskIntensity(mask_data) |
| |
|
| | def __call__(self, data: Mapping[Hashable, np.ndarray]) -> Dict[Hashable, np.ndarray]: |
| | d = dict(data) |
| | for key in self.keys: |
| | d[key] = self.converter(d[key]) |
| | return d |
| |
|
| |
|
| | class GaussianSmoothd(MapTransform): |
| | """ |
| | Dictionary-based wrapper of :py:class:`monai.transforms.GaussianSmooth`. |
| | |
| | Args: |
| | keys: keys of the corresponding items to be transformed. |
| | See also: :py:class:`monai.transforms.compose.MapTransform` |
| | sigma: if a list of values, must match the count of spatial dimensions of input data, |
| | and apply every value in the list to 1 spatial dimension. if only 1 value provided, |
| | use it for all spatial dimensions. |
| | |
| | """ |
| |
|
| | def __init__(self, keys: KeysCollection, sigma: Union[Sequence[float], float]) -> None: |
| | super().__init__(keys) |
| | self.converter = GaussianSmooth(sigma) |
| |
|
| | def __call__(self, data: Mapping[Hashable, np.ndarray]) -> Dict[Hashable, np.ndarray]: |
| | d = dict(data) |
| | for key in self.keys: |
| | d[key] = self.converter(d[key]) |
| | return d |
| |
|
| |
|
| | class RandGaussianSmoothd(Randomizable, MapTransform): |
| | """ |
| | Dictionary-based wrapper of :py:class:`monai.transforms.GaussianSmooth`. |
| | |
| | Args: |
| | keys: keys of the corresponding items to be transformed. |
| | See also: :py:class:`monai.transforms.compose.MapTransform` |
| | sigma_x: randomly select sigma value for the first spatial dimension. |
| | sigma_y: randomly select sigma value for the second spatial dimension if have. |
| | sigma_z: randomly select sigma value for the third spatial dimension if have. |
| | prob: probability of Gaussian smooth. |
| | |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | keys: KeysCollection, |
| | sigma_x: Tuple[float, float] = (0.25, 1.5), |
| | sigma_y: Tuple[float, float] = (0.25, 1.5), |
| | sigma_z: Tuple[float, float] = (0.25, 1.5), |
| | prob: float = 0.1, |
| | ) -> None: |
| | super().__init__(keys) |
| | self.sigma_x = sigma_x |
| | self.sigma_y = sigma_y |
| | self.sigma_z = sigma_z |
| | self.prob = prob |
| | self._do_transform = False |
| |
|
| | def randomize(self, data: Optional[Any] = None) -> None: |
| | self._do_transform = self.R.random_sample() < self.prob |
| | self.x = self.R.uniform(low=self.sigma_x[0], high=self.sigma_x[1]) |
| | self.y = self.R.uniform(low=self.sigma_y[0], high=self.sigma_y[1]) |
| | self.z = self.R.uniform(low=self.sigma_z[0], high=self.sigma_z[1]) |
| |
|
| | def __call__(self, data: Mapping[Hashable, np.ndarray]) -> Dict[Hashable, np.ndarray]: |
| | d = dict(data) |
| | self.randomize() |
| | if not self._do_transform: |
| | return d |
| | for key in self.keys: |
| | sigma = ensure_tuple_size(tup=(self.x, self.y, self.z), dim=d[key].ndim - 1) |
| | d[key] = GaussianSmooth(sigma=sigma)(d[key]) |
| | return d |
| |
|
| |
|
| | class GaussianSharpend(MapTransform): |
| | """ |
| | Dictionary-based wrapper of :py:class:`monai.transforms.GaussianSharpen`. |
| | |
| | Args: |
| | keys: keys of the corresponding items to be transformed. |
| | See also: :py:class:`monai.transforms.compose.MapTransform` |
| | sigma1: sigma parameter for the first gaussian kernel. if a list of values, must match the count |
| | of spatial dimensions of input data, and apply every value in the list to 1 spatial dimension. |
| | if only 1 value provided, use it for all spatial dimensions. |
| | sigma2: sigma parameter for the second gaussian kernel. if a list of values, must match the count |
| | of spatial dimensions of input data, and apply every value in the list to 1 spatial dimension. |
| | if only 1 value provided, use it for all spatial dimensions. |
| | alpha: weight parameter to compute the final result. |
| | |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | keys: KeysCollection, |
| | sigma1: Union[Sequence[float], float] = 3.0, |
| | sigma2: Union[Sequence[float], float] = 1.0, |
| | alpha: float = 30.0, |
| | ) -> None: |
| | super().__init__(keys) |
| | self.converter = GaussianSharpen(sigma1, sigma2, alpha) |
| |
|
| | def __call__(self, data: Mapping[Hashable, np.ndarray]) -> Dict[Hashable, np.ndarray]: |
| | d = dict(data) |
| | for key in self.keys: |
| | d[key] = self.converter(d[key]) |
| | return d |
| |
|
| |
|
| | class RandGaussianSharpend(Randomizable, MapTransform): |
| | """ |
| | Dictionary-based wrapper of :py:class:`monai.transforms.GaussianSharpen`. |
| | |
| | Args: |
| | keys: keys of the corresponding items to be transformed. |
| | See also: :py:class:`monai.transforms.compose.MapTransform` |
| | sigma1_x: randomly select sigma value for the first spatial dimension of first gaussian kernel. |
| | sigma1_y: randomly select sigma value for the second spatial dimension(if have) of first gaussian kernel. |
| | sigma1_z: randomly select sigma value for the third spatial dimension(if have) of first gaussian kernel. |
| | sigma2_x: randomly select sigma value for the first spatial dimension of second gaussian kernel. |
| | if only 1 value `X` provided, it must be smaller than `sigma1_x` and randomly select from [X, sigma1_x]. |
| | sigma2_y: randomly select sigma value for the second spatial dimension(if have) of second gaussian kernel. |
| | if only 1 value `Y` provided, it must be smaller than `sigma1_y` and randomly select from [Y, sigma1_y]. |
| | sigma2_z: randomly select sigma value for the third spatial dimension(if have) of second gaussian kernel. |
| | if only 1 value `Z` provided, it must be smaller than `sigma1_z` and randomly select from [Z, sigma1_z]. |
| | alpha: randomly select weight parameter to compute the final result. |
| | prob: probability of Gaussian sharpen. |
| | |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | keys: KeysCollection, |
| | sigma1_x: Tuple[float, float] = (0.5, 1.0), |
| | sigma1_y: Tuple[float, float] = (0.5, 1.0), |
| | sigma1_z: Tuple[float, float] = (0.5, 1.0), |
| | sigma2_x: Union[Tuple[float, float], float] = 0.5, |
| | sigma2_y: Union[Tuple[float, float], float] = 0.5, |
| | sigma2_z: Union[Tuple[float, float], float] = 0.5, |
| | alpha: Tuple[float, float] = (10.0, 30.0), |
| | prob: float = 0.1, |
| | ): |
| | super().__init__(keys) |
| | self.sigma1_x = sigma1_x |
| | self.sigma1_y = sigma1_y |
| | self.sigma1_z = sigma1_z |
| | self.sigma2_x = sigma2_x |
| | self.sigma2_y = sigma2_y |
| | self.sigma2_z = sigma2_z |
| | self.alpha = alpha |
| | self.prob = prob |
| | self._do_transform = False |
| |
|
| | def randomize(self, data: Optional[Any] = None) -> None: |
| | self._do_transform = self.R.random_sample() < self.prob |
| | self.x1 = self.R.uniform(low=self.sigma1_x[0], high=self.sigma1_x[1]) |
| | self.y1 = self.R.uniform(low=self.sigma1_y[0], high=self.sigma1_y[1]) |
| | self.z1 = self.R.uniform(low=self.sigma1_z[0], high=self.sigma1_z[1]) |
| | sigma2_x = (self.sigma2_x, self.x1) if not isinstance(self.sigma2_x, Iterable) else self.sigma2_x |
| | sigma2_y = (self.sigma2_y, self.y1) if not isinstance(self.sigma2_y, Iterable) else self.sigma2_y |
| | sigma2_z = (self.sigma2_z, self.z1) if not isinstance(self.sigma2_z, Iterable) else self.sigma2_z |
| | self.x2 = self.R.uniform(low=sigma2_x[0], high=sigma2_x[1]) |
| | self.y2 = self.R.uniform(low=sigma2_y[0], high=sigma2_y[1]) |
| | self.z2 = self.R.uniform(low=sigma2_z[0], high=sigma2_z[1]) |
| | self.a = self.R.uniform(low=self.alpha[0], high=self.alpha[1]) |
| |
|
| | def __call__(self, data): |
| | d = dict(data) |
| | self.randomize() |
| | if not self._do_transform: |
| | return d |
| | for key in self.keys: |
| | sigma1 = ensure_tuple_size(tup=(self.x1, self.y1, self.z1), dim=d[key].ndim - 1) |
| | sigma2 = ensure_tuple_size(tup=(self.x2, self.y2, self.z2), dim=d[key].ndim - 1) |
| | d[key] = GaussianSharpen(sigma1=sigma1, sigma2=sigma2, alpha=self.a)(d[key]) |
| | return d |
| |
|
| |
|
| | RandGaussianNoiseD = RandGaussianNoiseDict = RandGaussianNoised |
| | ShiftIntensityD = ShiftIntensityDict = ShiftIntensityd |
| | RandShiftIntensityD = RandShiftIntensityDict = RandShiftIntensityd |
| | ScaleIntensityD = ScaleIntensityDict = ScaleIntensityd |
| | RandScaleIntensityD = RandScaleIntensityDict = RandScaleIntensityd |
| | NormalizeIntensityD = NormalizeIntensityDict = NormalizeIntensityd |
| | ThresholdIntensityD = ThresholdIntensityDict = ThresholdIntensityd |
| | ScaleIntensityRangeD = ScaleIntensityRangeDict = ScaleIntensityRanged |
| | AdjustContrastD = AdjustContrastDict = AdjustContrastd |
| | RandAdjustContrastD = RandAdjustContrastDict = RandAdjustContrastd |
| | ScaleIntensityRangePercentilesD = ScaleIntensityRangePercentilesDict = ScaleIntensityRangePercentilesd |
| | MaskIntensityD = MaskIntensityDict = MaskIntensityd |
| | GaussianSmoothD = GaussianSmoothDict = GaussianSmoothd |
| | RandGaussianSmoothD = RandGaussianSmoothDict = RandGaussianSmoothd |
| | GaussianSharpenD = GaussianSharpenDict = GaussianSharpend |
| | RandGaussianSharpenD = RandGaussianSharpenDict = RandGaussianSharpend |
| |
|