# Copyright (c) MONAI Consortium # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # http://www.apache.org/licenses/LICENSE-2.0 # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from __future__ import annotations from typing import Any, Mapping, Sequence import torch from monai.apps.utils import get_logger from monai.config import NdarrayOrTensor from monai.data.meta_tensor import MetaTensor from monai.data.utils import to_affine_nd from monai.transforms.lazy.utils import ( affine_from_pending, combine_transforms, is_compatible_apply_kwargs, kwargs_from_pending, resample, ) from monai.transforms.traits import LazyTrait from monai.transforms.transform import MapTransform from monai.utils import LazyAttr, look_up_option __all__ = ["apply_pending_transforms", "apply_pending_transforms_in_order", "apply_pending"] __override_keywords = {"mode", "padding_mode", "dtype", "align_corners", "resample_mode", "device"} def _log_pending_info( transform: Any, data: Any, activity: str, *, lazy: bool | None = None, key: str | None = None, logger_name: bool | str = False, ): if logger_name is False: return logger_name = logger_name if isinstance(logger_name, str) else "apply_pending_transforms" logger = get_logger(logger_name) tcname = type(transform).__name__ if isinstance(transform, LazyTrait): tlazy = f", transform.lazy: {transform.lazy}" if lazy is not None and lazy != transform.lazy: tlazy += " (overridden)" else: tlazy = ", transform is not lazy" msg = f"{activity} - lazy: {lazy}, {{key_msg}}pending: {{pcount}}, upcoming '{tcname}'{tlazy}" if isinstance(transform, MapTransform): transform_keys = transform.keys if key is None else (key,) for k in transform_keys: if k in data: pcount = len(data[k].pending_operations) if isinstance(data[k], MetaTensor) else 0 logger.info(msg.format(pcount=pcount, key_msg=f"key: '{k}', ")) else: pcount = len(data.pending_operations) if isinstance(data, MetaTensor) else 0 logger.info(msg.format(pcount=pcount, key_msg="" if key is None else f"key: '{key}', ")) def _log_applied_info(data: Any, key=None, logger_name: bool | str = False): if logger_name is False: return logger_name = logger_name if isinstance(logger_name, str) else "apply_pending_transforms" logger = get_logger(logger_name) key_str = "" if key is None else f"key: '{key}', " logger.info(f"Pending transforms applied: {key_str}applied_operations: {len(data.applied_operations)}") def apply_pending_transforms( data: NdarrayOrTensor | Sequence[Any | NdarrayOrTensor] | Mapping[Any, NdarrayOrTensor], keys: tuple | None, overrides: dict | None = None, logger_name: bool | str = False, ): """ apply_pending_transforms is called with either a tensor or a dictionary, some entries of which contain tensors. When operating on a dictionary of tensors, the 'keys' parameter determines what tensors should be checked. If 'keys' is not set, all keys of 'data' are considered. This method optionally takes a set of overrides that can be used to change specific parameters on the transform pipeline. See ``Compose`` for more details. This method takes a logger_name that can be used to override the default logger, to provide telemetry during the execution of pending transforms. This method is intended primarily for use by ``execute_compose`` and other methods that handle the underlying execution of transform pipelines. You should not need to use it in the general case, unless you are developing functionality to perform such operations. Args: data: a ``torch.Tensor`` or ``MetaTensor``, or dictionary of tensors. keys: an optional tuple of keys that filters the keys on 'data' if it is a dict overrides: An optional dictionary that specifies parameters that can be used to override transform arguments when they are called. When 'data' is a dict, this dictionary should contain a dictionary of overrides for each key that needs them logger_name: An optional name for a logger to be used when applying pending transforms. If None, logging is suppressed. Returns: an object of the same type as data if pending transforms were applied, or 'data' if they were not """ if isinstance(data, list): return [apply_pending_transforms(d, keys, overrides, logger_name) for d in data] if isinstance(data, tuple): return tuple(apply_pending_transforms(d, keys, overrides, logger_name) for d in data) if isinstance(data, dict): # get the keys from 'data' for metatensors with pending operations. If 'keys' is set, select # only data keys that are in 'keys' active_keys = [k for k in data.keys() if keys is None or k in keys] keys_to_update = [k for k in active_keys if isinstance(data[k], MetaTensor) and data[k].has_pending_operations] if len(keys_to_update) > 0: rdata = dict(data) for k in keys_to_update: overrides_ = None if overrides is None else overrides.get(k, None) rdata[k], _ = apply_pending(data[k], overrides=overrides_) _log_applied_info(rdata[k], key=k, logger_name=logger_name) return rdata else: if isinstance(data, MetaTensor) and data.has_pending_operations: rdata, _ = apply_pending(data, overrides=overrides) _log_applied_info(rdata, logger_name=logger_name) return rdata return data def apply_pending_transforms_in_order( transform, data, lazy: bool | None = None, overrides: dict | None = None, logger_name: bool | str = False ): """ This method causes "in order" processing of pending transforms to occur. "in order" processing of pending transforms ensures that all pending transforms have been applied to the tensor before a non-lazy transform (or lazy transform that is executing non-lazily) is carried out. It ensures that no operations will be added to a metatensor's apply_operations while there are outstanding pending_operations. Note that there is only one mechanism for executing lazy resampling at present but this is expected to change in future releases. Evaluation of pending transforms is performed under the following circumstances: * If the transform is a lazy transform and: * The transform checks data as part of its execution, or * the transform is not executing lazily * If the transform is an ApplyPending[d] transform * If the transform is not a lazy transform This method is designed to be used only in the context of implementing lazy resampling functionality. In general you should not need to interact with or use this method directly, and its API may change without warning between releases. See the :ref:`Lazy Resampling topic for more information about lazy resampling. Args: transform: a transform that should be evaluated to determine whether pending transforms should be applied data: a tensor / MetaTensor, or dictionary containing tensors / MetaTensors whose pending transforms may need to be applied lazy: The lazy mode that is being applied (this can be False, True or None) overrides: An optional dictionary containing overrides to be applied to the pending transforms when they are lazily executed. If data is a dict, it should contain a dictionary of overrides for each key that needs them logger_name: An optional name for a logger to be used when applying pending transforms. If None, logging is suppressed. Returns: an object of the same type as data if pending transforms were applied, or 'data' if they were not """ from monai.transforms.lazy.dictionary import ApplyPendingd must_apply_pending = True keys = transform.keys if isinstance(transform, ApplyPendingd) else None if isinstance(transform, LazyTrait) and not transform.requires_current_data: must_apply_pending = not (transform.lazy if lazy is None else lazy) if must_apply_pending is True: _log_pending_info(transform, data, "Apply pending transforms", lazy=lazy, logger_name=logger_name) return apply_pending_transforms(data, keys, overrides, logger_name) _log_pending_info(transform, data, "Accumulate pending transforms", lazy=lazy, logger_name=logger_name) return data def apply_pending(data: torch.Tensor | MetaTensor, pending: list | None = None, overrides: dict | None = None): """ This method applies pending transforms to `data` tensors. Currently, only 2d and 3d inputs are supported. This method is designed to be called by ``apply_pending_transforms`` and other methods / classes that are part of the implementation of lazy resampling. In general, you should not need to call this method unless you are directly developing custom lazy execution strategies. It works by calculating the overall effect of the accumulated pending transforms. When it runs out of pending transforms or when it finds incompatibilities between the accumulated pending transform and the next pending transform, it then applies the accumulated transform in a call to ``resample``. Pending transforms are incompatible with each other if one or more of the arguments in the pending transforms differ. These are parameters such as 'mode', 'padding_mode', 'dtype' and so forth. If a pending transform doesn't have a given parameter, it is considered compatible with the accumulated transform. If a subsequent transform has a parameter that is incompatible with the accumulated transform (e.g. 'mode' of 'bilinear' vs. 'mode' of 'nearest'), an intermediate resample will be performed and the accumulated transform reset to its starting state. After resampling, the pending transforms are pushed to the ``applied_transforms`` field of the resulting MetaTensor. Note, if a torch.tensor is passed to this method along with a list of pending transforms, the resampled tensor will be wrapped in a MetaTensor before being returned. Args: data: A torch Tensor or a monai MetaTensor. pending: pending transforms. This must be set if data is a Tensor, but is optional if data is a MetaTensor. overrides: a dictionary of overrides for the transform arguments. The keys must be one of: - mode: {``"bilinear"``, ``"nearest"``} or spline interpolation order ``0-5`` (integers). Interpolation mode to calculate output values. Defaults to None. See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html When it's `an integer`, the numpy (cpu tensor)/cupy (cuda tensor) backends will be used and the value represents the order of the spline interpolation. See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html - padding_mode: {``"zeros"``, ``"border"``, ``"reflection"``} Padding mode for outside grid values. Defaults to None. See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html When `mode` is an integer, using numpy/cupy backends, this argument accepts {'reflect', 'grid-mirror', 'constant', 'grid-constant', 'nearest', 'mirror', 'grid-wrap', 'wrap'}. See also: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.map_coordinates.html - dtype: data type for resampling computation. Defaults to ``float64``. If ``None``, use the data type of input data, this option may not be compatible the resampling backend. - align_corners: Geometrically, we consider the pixels of the input as squares rather than points, when using the PyTorch resampling backend. Defaults to ``False``. See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html - device: device for resampling computation. Defaults to ``None``. - resample_mode: the mode of resampling, currently support ``"auto"``. Setting to other values will use the :py:class:`monai.transforms.SpatialResample` for resampling (instead of potentially crop/pad). """ overrides = (overrides or {}).copy() for k in overrides: look_up_option(k, __override_keywords) # check existence of the key if isinstance(data, MetaTensor) and pending is None: pending = data.pending_operations.copy() data.clear_pending_operations() pending = [] if pending is None else pending if not pending: return data, [] cumulative_xform = affine_from_pending(pending[0]) if cumulative_xform.shape[0] == 3: cumulative_xform = to_affine_nd(3, cumulative_xform) cur_kwargs = kwargs_from_pending(pending[0]) override_kwargs: dict[str, Any] = {} if "mode" in overrides: override_kwargs[LazyAttr.INTERP_MODE] = overrides["mode"] if "padding_mode" in overrides: override_kwargs[LazyAttr.PADDING_MODE] = overrides["padding_mode"] if "align_corners" in overrides: override_kwargs[LazyAttr.ALIGN_CORNERS] = overrides["align_corners"] if "resample_mode" in overrides: override_kwargs[LazyAttr.RESAMPLE_MODE] = overrides["resample_mode"] override_dtype = overrides.get("dtype", torch.float64) override_kwargs[LazyAttr.DTYPE] = data.dtype if override_dtype is None else override_dtype device = overrides.get("device") for p in pending[1:]: new_kwargs = kwargs_from_pending(p) if not is_compatible_apply_kwargs(cur_kwargs, new_kwargs): # carry out an intermediate resample here due to incompatibility between arguments _cur_kwargs = cur_kwargs.copy() _cur_kwargs.update(override_kwargs) data = resample(data.to(device), cumulative_xform, _cur_kwargs) next_matrix = affine_from_pending(p) if next_matrix.shape[0] == 3: next_matrix = to_affine_nd(3, next_matrix) cumulative_xform = combine_transforms(cumulative_xform, next_matrix) cur_kwargs.update(new_kwargs) cur_kwargs.update(override_kwargs) data = resample(data.to(device), cumulative_xform, cur_kwargs) if isinstance(data, MetaTensor): for p in pending: data.push_applied_operation(p) return data, pending