|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
from __future__ import annotations |
|
|
|
|
|
import json |
|
|
from collections.abc import Callable, Hashable, Iterable, Sequence |
|
|
from typing import Any |
|
|
|
|
|
import numpy as np |
|
|
import torch |
|
|
|
|
|
from monai.config import IndexSelection, KeysCollection, NdarrayOrTensor |
|
|
from monai.networks.layers import GaussianFilter |
|
|
from monai.transforms import Resize, SpatialCrop |
|
|
from monai.transforms.transform import MapTransform, Randomizable, Transform |
|
|
from monai.transforms.utils import generate_spatial_bounding_box, is_positive |
|
|
from monai.utils import InterpolateMode, ensure_tuple, ensure_tuple_rep, min_version, optional_import |
|
|
from monai.utils.enums import PostFix |
|
|
|
|
|
measure, _ = optional_import("skimage.measure", "0.14.2", min_version) |
|
|
distance_transform_cdt, _ = optional_import("scipy.ndimage.morphology", name="distance_transform_cdt") |
|
|
|
|
|
DEFAULT_POST_FIX = PostFix.meta() |
|
|
|
|
|
|
|
|
|
|
|
class FindAllValidSlicesd(Transform): |
|
|
""" |
|
|
Find/List all valid slices in the label. |
|
|
Label is assumed to be a 4D Volume with shape CDHW, where C=1. |
|
|
|
|
|
Args: |
|
|
label: key to the label source. |
|
|
sids: key to store slices indices having valid label map. |
|
|
""" |
|
|
|
|
|
def __init__(self, label: str = "label", sids: str = "sids"): |
|
|
self.label = label |
|
|
self.sids = sids |
|
|
|
|
|
def _apply(self, label): |
|
|
sids = [] |
|
|
for sid in range(label.shape[1]): |
|
|
if np.sum(label[0][sid]) != 0: |
|
|
sids.append(sid) |
|
|
return np.asarray(sids) |
|
|
|
|
|
def __call__(self, data: Any) -> dict: |
|
|
d: dict = dict(data) |
|
|
label = d[self.label].numpy() if isinstance(data[self.label], torch.Tensor) else data[self.label] |
|
|
if label.shape[0] != 1: |
|
|
raise ValueError(f"Only supports single channel labels, got label shape {label.shape}!") |
|
|
|
|
|
if len(label.shape) != 4: |
|
|
raise ValueError(f"Only supports label with shape CDHW, got label shape {label.shape}!") |
|
|
|
|
|
sids = self._apply(label) |
|
|
if sids is not None and len(sids): |
|
|
d[self.sids] = sids |
|
|
return d |
|
|
|
|
|
|
|
|
class AddInitialSeedPointd(Randomizable, Transform): |
|
|
""" |
|
|
Add random guidance as initial seed point for a given label. |
|
|
|
|
|
Note that the label is of size (C, D, H, W) or (C, H, W) |
|
|
|
|
|
The guidance is of size (2, N, # of dims) where N is number of guidance added. |
|
|
# of dims = 4 when C, D, H, W; # of dims = 3 when (C, H, W) |
|
|
|
|
|
Args: |
|
|
label: label source. |
|
|
guidance: key to store guidance. |
|
|
sids: key that represents list of valid slice indices for the given label. |
|
|
sid: key that represents the slice to add initial seed point. If not present, random sid will be chosen. |
|
|
connected_regions: maximum connected regions to use for adding initial points. |
|
|
""" |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
label: str = "label", |
|
|
guidance: str = "guidance", |
|
|
sids: str = "sids", |
|
|
sid: str = "sid", |
|
|
connected_regions: int = 5, |
|
|
): |
|
|
self.label = label |
|
|
self.sids_key = sids |
|
|
self.sid_key = sid |
|
|
self.sid = None |
|
|
self.guidance = guidance |
|
|
self.connected_regions = connected_regions |
|
|
|
|
|
def randomize(self, data): |
|
|
sid = data.get(self.sid_key, None) |
|
|
sids = data.get(self.sids_key, None) |
|
|
if sids is not None: |
|
|
if sid is None or sid not in sids: |
|
|
sid = self.R.choice(sids, replace=False) |
|
|
else: |
|
|
sid = None |
|
|
self.sid = sid |
|
|
|
|
|
def _apply(self, label, sid): |
|
|
dimensions = 3 if len(label.shape) > 3 else 2 |
|
|
default_guidance = [-1] * (dimensions + 1) |
|
|
|
|
|
dims = dimensions |
|
|
if sid is not None and dimensions == 3: |
|
|
dims = 2 |
|
|
label = label[0][sid][np.newaxis] |
|
|
|
|
|
label = (label > 0.5).astype(np.float32) |
|
|
blobs_labels = measure.label(label.astype(int), background=0) if dims == 2 else label |
|
|
if np.max(blobs_labels) <= 0: |
|
|
raise AssertionError("Not a valid Label") |
|
|
|
|
|
pos_guidance = [] |
|
|
for ridx in range(1, 2 if dims == 3 else self.connected_regions + 1): |
|
|
if dims == 2: |
|
|
label = (blobs_labels == ridx).astype(np.float32) |
|
|
if np.sum(label) == 0: |
|
|
pos_guidance.append(default_guidance) |
|
|
continue |
|
|
|
|
|
distance = distance_transform_cdt(label).flatten() |
|
|
probability = np.exp(distance) - 1.0 |
|
|
|
|
|
idx = np.where(label.flatten() > 0)[0] |
|
|
seed = self.R.choice(idx, size=1, p=probability[idx] / np.sum(probability[idx])) |
|
|
dst = distance[seed] |
|
|
|
|
|
g = np.asarray(np.unravel_index(seed, label.shape)).transpose().tolist()[0] |
|
|
g[0] = dst[0] |
|
|
if dimensions == 2 or dims == 3: |
|
|
pos_guidance.append(g) |
|
|
else: |
|
|
pos_guidance.append([g[0], sid, g[-2], g[-1]]) |
|
|
|
|
|
return np.asarray([pos_guidance, [default_guidance] * len(pos_guidance)]) |
|
|
|
|
|
def __call__(self, data): |
|
|
d = dict(data) |
|
|
self.randomize(data) |
|
|
d[self.guidance] = json.dumps(self._apply(d[self.label], self.sid).astype(int, copy=False).tolist()) |
|
|
return d |
|
|
|
|
|
|
|
|
class AddGuidanceSignald(Transform): |
|
|
""" |
|
|
Add Guidance signal for input image. |
|
|
|
|
|
Based on the "guidance" points, apply gaussian to them and add them as new channel for input image. |
|
|
|
|
|
Args: |
|
|
image: key to the image source. |
|
|
guidance: key to store guidance. |
|
|
sigma: standard deviation for Gaussian kernel. |
|
|
number_intensity_ch: channel index. |
|
|
|
|
|
""" |
|
|
|
|
|
def __init__(self, image: str = "image", guidance: str = "guidance", sigma: int = 2, number_intensity_ch: int = 1): |
|
|
self.image = image |
|
|
self.guidance = guidance |
|
|
self.sigma = sigma |
|
|
self.number_intensity_ch = number_intensity_ch |
|
|
|
|
|
def _get_signal(self, image, guidance): |
|
|
dimensions = 3 if len(image.shape) > 3 else 2 |
|
|
guidance = guidance.tolist() if isinstance(guidance, np.ndarray) else guidance |
|
|
guidance = json.loads(guidance) if isinstance(guidance, str) else guidance |
|
|
if dimensions == 3: |
|
|
signal = np.zeros((len(guidance), image.shape[-3], image.shape[-2], image.shape[-1]), dtype=np.float32) |
|
|
else: |
|
|
signal = np.zeros((len(guidance), image.shape[-2], image.shape[-1]), dtype=np.float32) |
|
|
|
|
|
sshape = signal.shape |
|
|
for i, g_i in enumerate(guidance): |
|
|
for point in g_i: |
|
|
if np.any(np.asarray(point) < 0): |
|
|
continue |
|
|
|
|
|
if dimensions == 3: |
|
|
p1 = max(0, min(int(point[-3]), sshape[-3] - 1)) |
|
|
p2 = max(0, min(int(point[-2]), sshape[-2] - 1)) |
|
|
p3 = max(0, min(int(point[-1]), sshape[-1] - 1)) |
|
|
signal[i, p1, p2, p3] = 1.0 |
|
|
else: |
|
|
p1 = max(0, min(int(point[-2]), sshape[-2] - 1)) |
|
|
p2 = max(0, min(int(point[-1]), sshape[-1] - 1)) |
|
|
signal[i, p1, p2] = 1.0 |
|
|
|
|
|
if np.max(signal[i]) > 0: |
|
|
signal_tensor = torch.tensor(signal[i]) |
|
|
pt_gaussian = GaussianFilter(len(signal_tensor.shape), sigma=self.sigma) |
|
|
signal_tensor = pt_gaussian(signal_tensor.unsqueeze(0).unsqueeze(0)) |
|
|
signal_tensor = signal_tensor.squeeze(0).squeeze(0) |
|
|
signal[i] = signal_tensor.detach().cpu().numpy() |
|
|
signal[i] = (signal[i] - np.min(signal[i])) / (np.max(signal[i]) - np.min(signal[i])) |
|
|
return signal |
|
|
|
|
|
def _apply(self, image, guidance): |
|
|
signal = self._get_signal(image, guidance) |
|
|
|
|
|
if isinstance(image, torch.Tensor): |
|
|
image = image.detach().cpu().numpy() |
|
|
|
|
|
image = image[0 : 0 + self.number_intensity_ch, ...] |
|
|
return np.concatenate([image, signal], axis=0) |
|
|
|
|
|
def __call__(self, data): |
|
|
d = dict(data) |
|
|
image = d[self.image] |
|
|
guidance = d[self.guidance] |
|
|
|
|
|
d[self.image] = self._apply(image, guidance) |
|
|
return d |
|
|
|
|
|
|
|
|
class FindDiscrepancyRegionsd(Transform): |
|
|
""" |
|
|
Find discrepancy between prediction and actual during click interactions during training. |
|
|
|
|
|
Args: |
|
|
label: key to label source. |
|
|
pred: key to prediction source. |
|
|
discrepancy: key to store discrepancies found between label and prediction. |
|
|
|
|
|
""" |
|
|
|
|
|
def __init__(self, label: str = "label", pred: str = "pred", discrepancy: str = "discrepancy"): |
|
|
self.label = label |
|
|
self.pred = pred |
|
|
self.discrepancy = discrepancy |
|
|
|
|
|
@staticmethod |
|
|
def disparity(label, pred): |
|
|
label = (label > 0.5).astype(np.float32) |
|
|
pred = (pred > 0.5).astype(np.float32) |
|
|
disparity = label - pred |
|
|
|
|
|
pos_disparity = (disparity > 0).astype(np.float32) |
|
|
neg_disparity = (disparity < 0).astype(np.float32) |
|
|
return [pos_disparity, neg_disparity] |
|
|
|
|
|
def _apply(self, label, pred): |
|
|
return self.disparity(label, pred) |
|
|
|
|
|
def __call__(self, data): |
|
|
d = dict(data) |
|
|
label = d[self.label] |
|
|
pred = d[self.pred] |
|
|
|
|
|
d[self.discrepancy] = self._apply(label, pred) |
|
|
return d |
|
|
|
|
|
|
|
|
class AddRandomGuidanced(Randomizable, Transform): |
|
|
""" |
|
|
Add random guidance based on discrepancies that were found between label and prediction. |
|
|
input shape is as below: |
|
|
Guidance is of shape (2, N, # of dim) |
|
|
Discrepancy is of shape (2, C, D, H, W) or (2, C, H, W) |
|
|
Probability is of shape (1) |
|
|
|
|
|
Args: |
|
|
guidance: key to guidance source. |
|
|
discrepancy: key that represents discrepancies found between label and prediction. |
|
|
probability: key that represents click/interaction probability. |
|
|
|
|
|
""" |
|
|
|
|
|
def __init__(self, guidance: str = "guidance", discrepancy: str = "discrepancy", probability: str = "probability"): |
|
|
self.guidance = guidance |
|
|
self.discrepancy = discrepancy |
|
|
self.probability = probability |
|
|
self._will_interact = None |
|
|
|
|
|
def randomize(self, data=None): |
|
|
probability = data[self.probability] |
|
|
self._will_interact = self.R.choice([True, False], p=[probability, 1.0 - probability]) |
|
|
|
|
|
def find_guidance(self, discrepancy): |
|
|
distance = distance_transform_cdt(discrepancy).flatten() |
|
|
probability = np.exp(distance) - 1.0 |
|
|
idx = np.where(discrepancy.flatten() > 0)[0] |
|
|
|
|
|
if np.sum(discrepancy > 0) > 0: |
|
|
seed = self.R.choice(idx, size=1, p=probability[idx] / np.sum(probability[idx])) |
|
|
dst = distance[seed] |
|
|
|
|
|
g = np.asarray(np.unravel_index(seed, discrepancy.shape)).transpose().tolist()[0] |
|
|
g[0] = dst[0] |
|
|
return g |
|
|
return None |
|
|
|
|
|
def add_guidance(self, discrepancy, will_interact): |
|
|
if not will_interact: |
|
|
return None, None |
|
|
|
|
|
pos_discr = discrepancy[0] |
|
|
neg_discr = discrepancy[1] |
|
|
|
|
|
can_be_positive = np.sum(pos_discr) > 0 |
|
|
can_be_negative = np.sum(neg_discr) > 0 |
|
|
correct_pos = np.sum(pos_discr) >= np.sum(neg_discr) |
|
|
|
|
|
if correct_pos and can_be_positive: |
|
|
return self.find_guidance(pos_discr), None |
|
|
|
|
|
if not correct_pos and can_be_negative: |
|
|
return None, self.find_guidance(neg_discr) |
|
|
return None, None |
|
|
|
|
|
def _apply(self, guidance, discrepancy): |
|
|
guidance = guidance.tolist() if isinstance(guidance, np.ndarray) else guidance |
|
|
guidance = json.loads(guidance) if isinstance(guidance, str) else guidance |
|
|
pos, neg = self.add_guidance(discrepancy, self._will_interact) |
|
|
if pos: |
|
|
guidance[0].append(pos) |
|
|
guidance[1].append([-1] * len(pos)) |
|
|
if neg: |
|
|
guidance[0].append([-1] * len(neg)) |
|
|
guidance[1].append(neg) |
|
|
|
|
|
return json.dumps(np.asarray(guidance, dtype=int).tolist()) |
|
|
|
|
|
def __call__(self, data): |
|
|
d = dict(data) |
|
|
guidance = d[self.guidance] |
|
|
discrepancy = d[self.discrepancy] |
|
|
|
|
|
self.randomize(data) |
|
|
d[self.guidance] = self._apply(guidance, discrepancy) |
|
|
return d |
|
|
|
|
|
|
|
|
class SpatialCropForegroundd(MapTransform): |
|
|
""" |
|
|
Crop only the foreground object of the expected images. |
|
|
|
|
|
Difference VS :py:class:`monai.transforms.CropForegroundd`: |
|
|
|
|
|
1. If the bounding box is smaller than spatial size in all dimensions then this transform will crop the |
|
|
object using box's center and spatial_size. |
|
|
|
|
|
2. This transform will set "start_coord_key", "end_coord_key", "original_shape_key" and "cropped_shape_key" |
|
|
in data[{key}_{meta_key_postfix}] |
|
|
|
|
|
The typical usage is to help training and evaluation if the valid part is small in the whole medical image. |
|
|
The valid part can be determined by any field in the data with `source_key`, for example: |
|
|
|
|
|
- Select values > 0 in image field as the foreground and crop on all fields specified by `keys`. |
|
|
- Select label = 3 in label field as the foreground to crop on all fields specified by `keys`. |
|
|
- Select label > 0 in the third channel of a One-Hot label field as the foreground to crop all `keys` fields. |
|
|
|
|
|
Users can define arbitrary function to select expected foreground from the whole source image or specified |
|
|
channels. And it can also add margin to every dim of the bounding box of foreground object. |
|
|
|
|
|
Args: |
|
|
keys: keys of the corresponding items to be transformed. |
|
|
See also: :py:class:`monai.transforms.MapTransform` |
|
|
source_key: data source to generate the bounding box of foreground, can be image or label, etc. |
|
|
spatial_size: minimal spatial size of the image patch e.g. [128, 128, 128] to fit in. |
|
|
select_fn: function to select expected foreground, default is to select values > 0. |
|
|
channel_indices: if defined, select foreground only on the specified channels |
|
|
of image. if None, select foreground on the whole image. |
|
|
margin: add margin value to spatial dims of the bounding box, if only 1 value provided, use it for all dims. |
|
|
allow_smaller: when computing box size with `margin`, whether allow the image size to be smaller |
|
|
than box size, default to `True`. if the margined size is bigger than image size, will pad with |
|
|
specified `mode`. |
|
|
meta_keys: explicitly indicate the key of the corresponding metadata dictionary. |
|
|
for example, for data with key `image`, the metadata by default is in `image_meta_dict`. |
|
|
the metadata is a dictionary object which contains: filename, original_shape, etc. |
|
|
it can be a sequence of string, map to the `keys`. |
|
|
if None, will try to construct meta_keys by `key_{meta_key_postfix}`. |
|
|
meta_key_postfix: if meta_keys is None, use `{key}_{meta_key_postfix}` to fetch/store the metadata according |
|
|
to the key data, default is `meta_dict`, the metadata is a dictionary object. |
|
|
For example, to handle key `image`, read/write affine matrices from the |
|
|
metadata `image_meta_dict` dictionary's `affine` field. |
|
|
start_coord_key: key to record the start coordinate of spatial bounding box for foreground. |
|
|
end_coord_key: key to record the end coordinate of spatial bounding box for foreground. |
|
|
original_shape_key: key to record original shape for foreground. |
|
|
cropped_shape_key: key to record cropped shape for foreground. |
|
|
allow_missing_keys: don't raise exception if key is missing. |
|
|
""" |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
keys: KeysCollection, |
|
|
source_key: str, |
|
|
spatial_size: Sequence[int] | np.ndarray, |
|
|
select_fn: Callable = is_positive, |
|
|
channel_indices: IndexSelection | None = None, |
|
|
margin: int = 0, |
|
|
allow_smaller: bool = True, |
|
|
meta_keys: KeysCollection | None = None, |
|
|
meta_key_postfix: str = DEFAULT_POST_FIX, |
|
|
start_coord_key: str = "foreground_start_coord", |
|
|
end_coord_key: str = "foreground_end_coord", |
|
|
original_shape_key: str = "foreground_original_shape", |
|
|
cropped_shape_key: str = "foreground_cropped_shape", |
|
|
allow_missing_keys: bool = False, |
|
|
) -> None: |
|
|
super().__init__(keys, allow_missing_keys) |
|
|
|
|
|
self.source_key = source_key |
|
|
self.spatial_size = list(spatial_size) |
|
|
self.select_fn = select_fn |
|
|
self.channel_indices = channel_indices |
|
|
self.margin = margin |
|
|
self.allow_smaller = allow_smaller |
|
|
self.meta_keys = ensure_tuple_rep(None, len(self.keys)) if meta_keys is None else ensure_tuple(meta_keys) |
|
|
if len(self.keys) != len(self.meta_keys): |
|
|
raise ValueError("meta_keys should have the same length as keys.") |
|
|
self.meta_key_postfix = ensure_tuple_rep(meta_key_postfix, len(self.keys)) |
|
|
self.start_coord_key = start_coord_key |
|
|
self.end_coord_key = end_coord_key |
|
|
self.original_shape_key = original_shape_key |
|
|
self.cropped_shape_key = cropped_shape_key |
|
|
|
|
|
def __call__(self, data): |
|
|
d = dict(data) |
|
|
box_start, box_end = generate_spatial_bounding_box( |
|
|
d[self.source_key], self.select_fn, self.channel_indices, self.margin, self.allow_smaller |
|
|
) |
|
|
|
|
|
center = list(np.mean([box_start, box_end], axis=0).astype(int, copy=False)) |
|
|
current_size = list(np.subtract(box_end, box_start).astype(int, copy=False)) |
|
|
|
|
|
if np.all(np.less(current_size, self.spatial_size)): |
|
|
cropper = SpatialCrop(roi_center=center, roi_size=self.spatial_size) |
|
|
box_start = np.array([s.start for s in cropper.slices]) |
|
|
box_end = np.array([s.stop for s in cropper.slices]) |
|
|
else: |
|
|
cropper = SpatialCrop(roi_start=box_start, roi_end=box_end) |
|
|
|
|
|
for key, meta_key, meta_key_postfix in self.key_iterator(d, self.meta_keys, self.meta_key_postfix): |
|
|
meta_key = meta_key or f"{key}_{meta_key_postfix}" |
|
|
d[meta_key][self.start_coord_key] = box_start |
|
|
d[meta_key][self.end_coord_key] = box_end |
|
|
d[meta_key][self.original_shape_key] = d[key].shape |
|
|
|
|
|
image = cropper(d[key]) |
|
|
d[meta_key][self.cropped_shape_key] = image.shape |
|
|
d[key] = image |
|
|
return d |
|
|
|
|
|
|
|
|
|
|
|
class AddGuidanceFromPointsd(Transform): |
|
|
""" |
|
|
Add guidance based on user clicks. |
|
|
|
|
|
We assume the input is loaded by LoadImaged and has the shape of (H, W, D) originally. |
|
|
Clicks always specify the coordinates in (H, W, D) |
|
|
|
|
|
If depth_first is True: |
|
|
|
|
|
Input is now of shape (D, H, W), will return guidance that specifies the coordinates in (D, H, W) |
|
|
|
|
|
else: |
|
|
|
|
|
Input is now of shape (H, W, D), will return guidance that specifies the coordinates in (H, W, D) |
|
|
|
|
|
Args: |
|
|
ref_image: key to reference image to fetch current and original image details. |
|
|
guidance: output key to store guidance. |
|
|
foreground: key that represents user foreground (+ve) clicks. |
|
|
background: key that represents user background (-ve) clicks. |
|
|
axis: axis that represents slices in 3D volume. (axis to Depth) |
|
|
depth_first: if depth (slices) is positioned at first dimension. |
|
|
spatial_dims: dimensions based on model used for deepgrow (2D vs 3D). |
|
|
slice_key: key that represents applicable slice to add guidance. |
|
|
meta_keys: explicitly indicate the key of the metadata dictionary of `ref_image`. |
|
|
for example, for data with key `image`, the metadata by default is in `image_meta_dict`. |
|
|
the metadata is a dictionary object which contains: filename, original_shape, etc. |
|
|
if None, will try to construct meta_keys by `{ref_image}_{meta_key_postfix}`. |
|
|
meta_key_postfix: if meta_key is None, use `{ref_image}_{meta_key_postfix}` to fetch the metadata according |
|
|
to the key data, default is `meta_dict`, the metadata is a dictionary object. |
|
|
For example, to handle key `image`, read/write affine matrices from the |
|
|
metadata `image_meta_dict` dictionary's `affine` field. |
|
|
|
|
|
""" |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
ref_image: str, |
|
|
guidance: str = "guidance", |
|
|
foreground: str = "foreground", |
|
|
background: str = "background", |
|
|
axis: int = 0, |
|
|
depth_first: bool = True, |
|
|
spatial_dims: int = 2, |
|
|
slice_key: str = "slice", |
|
|
meta_keys: str | None = None, |
|
|
meta_key_postfix: str = DEFAULT_POST_FIX, |
|
|
): |
|
|
self.ref_image = ref_image |
|
|
self.guidance = guidance |
|
|
self.foreground = foreground |
|
|
self.background = background |
|
|
self.axis = axis |
|
|
self.depth_first = depth_first |
|
|
self.dimensions = spatial_dims |
|
|
self.slice = slice_key |
|
|
self.meta_keys = meta_keys |
|
|
self.meta_key_postfix = meta_key_postfix |
|
|
|
|
|
def _apply(self, pos_clicks, neg_clicks, factor, slice_num): |
|
|
pos = neg = [] |
|
|
|
|
|
if self.dimensions == 2: |
|
|
points: list = list(pos_clicks) |
|
|
points.extend(neg_clicks) |
|
|
|
|
|
slices = list(np.unique(np.array(points)[:, self.axis])) |
|
|
slice_idx = slices[0] if slice_num is None else next(x for x in slices if x == slice_num) |
|
|
|
|
|
if len(pos_clicks): |
|
|
pos_clicks = np.array(pos_clicks) |
|
|
pos = (pos_clicks[np.where(pos_clicks[:, self.axis] == slice_idx)] * factor)[:, 1:].astype(int).tolist() |
|
|
if len(neg_clicks): |
|
|
neg_clicks = np.array(neg_clicks) |
|
|
neg = (neg_clicks[np.where(neg_clicks[:, self.axis] == slice_idx)] * factor)[:, 1:].astype(int).tolist() |
|
|
|
|
|
guidance = [pos, neg, slice_idx] |
|
|
else: |
|
|
if len(pos_clicks): |
|
|
pos = np.multiply(pos_clicks, factor).astype(int, copy=False).tolist() |
|
|
if len(neg_clicks): |
|
|
neg = np.multiply(neg_clicks, factor).astype(int, copy=False).tolist() |
|
|
guidance = [pos, neg] |
|
|
return guidance |
|
|
|
|
|
def __call__(self, data): |
|
|
d = dict(data) |
|
|
meta_dict_key = self.meta_keys or f"{self.ref_image}_{self.meta_key_postfix}" |
|
|
if meta_dict_key not in d: |
|
|
raise RuntimeError(f"Missing meta_dict {meta_dict_key} in data!") |
|
|
if "spatial_shape" not in d[meta_dict_key]: |
|
|
raise RuntimeError('Missing "spatial_shape" in meta_dict!') |
|
|
original_shape = d[meta_dict_key]["spatial_shape"] |
|
|
current_shape = list(d[self.ref_image].shape) |
|
|
|
|
|
if self.depth_first: |
|
|
if self.axis != 0: |
|
|
raise RuntimeError("Depth first means the depth axis should be 0.") |
|
|
|
|
|
original_shape = np.roll(original_shape, 1) |
|
|
|
|
|
factor = np.array(current_shape) / original_shape |
|
|
|
|
|
fg_bg_clicks = [] |
|
|
for key in [self.foreground, self.background]: |
|
|
clicks = d[key] |
|
|
clicks = list(np.array(clicks, dtype=int)) |
|
|
if self.depth_first: |
|
|
for i in range(len(clicks)): |
|
|
clicks[i] = list(np.roll(clicks[i], 1)) |
|
|
fg_bg_clicks.append(clicks) |
|
|
d[self.guidance] = self._apply(fg_bg_clicks[0], fg_bg_clicks[1], factor, d.get(self.slice)) |
|
|
return d |
|
|
|
|
|
|
|
|
class SpatialCropGuidanced(MapTransform): |
|
|
""" |
|
|
Crop image based on guidance with minimal spatial size. |
|
|
|
|
|
- If the bounding box is smaller than spatial size in all dimensions then this transform will crop the |
|
|
object using box's center and spatial_size. |
|
|
|
|
|
- This transform will set "start_coord_key", "end_coord_key", "original_shape_key" and "cropped_shape_key" |
|
|
in data[{key}_{meta_key_postfix}] |
|
|
|
|
|
Input data is of shape (C, spatial_1, [spatial_2, ...]) |
|
|
|
|
|
Args: |
|
|
keys: keys of the corresponding items to be transformed. |
|
|
guidance: key to the guidance. It is used to generate the bounding box of foreground |
|
|
spatial_size: minimal spatial size of the image patch e.g. [128, 128, 128] to fit in. |
|
|
margin: add margin value to spatial dims of the bounding box, if only 1 value provided, use it for all dims. |
|
|
meta_keys: explicitly indicate the key of the corresponding metadata dictionary. |
|
|
for example, for data with key `image`, the metadata by default is in `image_meta_dict`. |
|
|
the metadata is a dictionary object which contains: filename, original_shape, etc. |
|
|
it can be a sequence of string, map to the `keys`. |
|
|
if None, will try to construct meta_keys by `key_{meta_key_postfix}`. |
|
|
meta_key_postfix: if meta_keys is None, use `key_{postfix}` to fetch the metadata according |
|
|
to the key data, default is `meta_dict`, the metadata is a dictionary object. |
|
|
For example, to handle key `image`, read/write affine matrices from the |
|
|
metadata `image_meta_dict` dictionary's `affine` field. |
|
|
start_coord_key: key to record the start coordinate of spatial bounding box for foreground. |
|
|
end_coord_key: key to record the end coordinate of spatial bounding box for foreground. |
|
|
original_shape_key: key to record original shape for foreground. |
|
|
cropped_shape_key: key to record cropped shape for foreground. |
|
|
allow_missing_keys: don't raise exception if key is missing. |
|
|
""" |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
keys: KeysCollection, |
|
|
guidance: str, |
|
|
spatial_size: Iterable[int], |
|
|
margin: int = 20, |
|
|
meta_keys: KeysCollection | None = None, |
|
|
meta_key_postfix: str = DEFAULT_POST_FIX, |
|
|
start_coord_key: str = "foreground_start_coord", |
|
|
end_coord_key: str = "foreground_end_coord", |
|
|
original_shape_key: str = "foreground_original_shape", |
|
|
cropped_shape_key: str = "foreground_cropped_shape", |
|
|
allow_missing_keys: bool = False, |
|
|
) -> None: |
|
|
super().__init__(keys, allow_missing_keys) |
|
|
|
|
|
self.guidance = guidance |
|
|
self.spatial_size = list(spatial_size) |
|
|
self.margin = margin |
|
|
self.meta_keys = ensure_tuple_rep(None, len(self.keys)) if meta_keys is None else ensure_tuple(meta_keys) |
|
|
if len(self.keys) != len(self.meta_keys): |
|
|
raise ValueError("meta_keys should have the same length as keys.") |
|
|
self.meta_key_postfix = ensure_tuple_rep(meta_key_postfix, len(self.keys)) |
|
|
self.start_coord_key = start_coord_key |
|
|
self.end_coord_key = end_coord_key |
|
|
self.original_shape_key = original_shape_key |
|
|
self.cropped_shape_key = cropped_shape_key |
|
|
|
|
|
def bounding_box(self, points, img_shape): |
|
|
ndim = len(img_shape) |
|
|
margin = ensure_tuple_rep(self.margin, ndim) |
|
|
for m in margin: |
|
|
if m < 0: |
|
|
raise ValueError("margin value should not be negative number.") |
|
|
|
|
|
box_start = [0] * ndim |
|
|
box_end = [0] * ndim |
|
|
|
|
|
for di in range(ndim): |
|
|
dt = points[..., di] |
|
|
min_d = max(min(dt - margin[di]), 0) |
|
|
max_d = min(img_shape[di], max(dt + margin[di] + 1)) |
|
|
box_start[di], box_end[di] = min_d, max_d |
|
|
return box_start, box_end |
|
|
|
|
|
def __call__(self, data: Any) -> dict: |
|
|
d: dict = dict(data) |
|
|
first_key: Hashable = self.first_key(d) |
|
|
if first_key == (): |
|
|
return d |
|
|
|
|
|
guidance = d[self.guidance] |
|
|
original_spatial_shape = d[first_key].shape[1:] |
|
|
box_start, box_end = self.bounding_box(np.array(guidance[0] + guidance[1]), original_spatial_shape) |
|
|
center = list(np.mean([box_start, box_end], axis=0).astype(int, copy=False)) |
|
|
spatial_size = self.spatial_size |
|
|
|
|
|
box_size = list(np.subtract(box_end, box_start).astype(int, copy=False)) |
|
|
spatial_size = spatial_size[-len(box_size) :] |
|
|
|
|
|
if len(spatial_size) < len(box_size): |
|
|
|
|
|
|
|
|
diff = len(box_size) - len(spatial_size) |
|
|
spatial_size = list(original_spatial_shape[1 : (1 + diff)]) + spatial_size |
|
|
|
|
|
if np.all(np.less(box_size, spatial_size)): |
|
|
if len(center) == 3: |
|
|
|
|
|
center[0] = spatial_size[0] // 2 |
|
|
cropper = SpatialCrop(roi_center=center, roi_size=spatial_size) |
|
|
else: |
|
|
cropper = SpatialCrop(roi_start=box_start, roi_end=box_end) |
|
|
|
|
|
|
|
|
box_start = np.array([s.start for s in cropper.slices]) |
|
|
box_end = np.array([s.stop for s in cropper.slices]) |
|
|
for key, meta_key, meta_key_postfix in self.key_iterator(d, self.meta_keys, self.meta_key_postfix): |
|
|
if not np.array_equal(d[key].shape[1:], original_spatial_shape): |
|
|
raise RuntimeError("All the image specified in keys should have same spatial shape") |
|
|
meta_key = meta_key or f"{key}_{meta_key_postfix}" |
|
|
d[meta_key][self.start_coord_key] = box_start |
|
|
d[meta_key][self.end_coord_key] = box_end |
|
|
d[meta_key][self.original_shape_key] = d[key].shape |
|
|
|
|
|
image = cropper(d[key]) |
|
|
d[meta_key][self.cropped_shape_key] = image.shape |
|
|
d[key] = image |
|
|
|
|
|
pos_clicks, neg_clicks = guidance[0], guidance[1] |
|
|
pos = np.subtract(pos_clicks, box_start).tolist() if len(pos_clicks) else [] |
|
|
neg = np.subtract(neg_clicks, box_start).tolist() if len(neg_clicks) else [] |
|
|
|
|
|
d[self.guidance] = [pos, neg] |
|
|
return d |
|
|
|
|
|
|
|
|
class ResizeGuidanced(Transform): |
|
|
""" |
|
|
Resize the guidance based on cropped vs resized image. |
|
|
|
|
|
This transform assumes that the images have been cropped and resized. And the shape after cropped is store inside |
|
|
the meta dict of ref image. |
|
|
|
|
|
Args: |
|
|
guidance: key to guidance |
|
|
ref_image: key to reference image to fetch current and original image details |
|
|
meta_keys: explicitly indicate the key of the metadata dictionary of `ref_image`. |
|
|
for example, for data with key `image`, the metadata by default is in `image_meta_dict`. |
|
|
the metadata is a dictionary object which contains: filename, original_shape, etc. |
|
|
if None, will try to construct meta_keys by `{ref_image}_{meta_key_postfix}`. |
|
|
meta_key_postfix: if meta_key is None, use `{ref_image}_{meta_key_postfix}` to fetch the metadata according |
|
|
to the key data, default is `meta_dict`, the metadata is a dictionary object. |
|
|
For example, to handle key `image`, read/write affine matrices from the |
|
|
metadata `image_meta_dict` dictionary's `affine` field. |
|
|
cropped_shape_key: key that records cropped shape for foreground. |
|
|
""" |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
guidance: str, |
|
|
ref_image: str, |
|
|
meta_keys: str | None = None, |
|
|
meta_key_postfix: str = DEFAULT_POST_FIX, |
|
|
cropped_shape_key: str = "foreground_cropped_shape", |
|
|
) -> None: |
|
|
self.guidance = guidance |
|
|
self.ref_image = ref_image |
|
|
self.meta_keys = meta_keys |
|
|
self.meta_key_postfix = meta_key_postfix |
|
|
self.cropped_shape_key = cropped_shape_key |
|
|
|
|
|
def __call__(self, data: Any) -> dict: |
|
|
d = dict(data) |
|
|
guidance = d[self.guidance] |
|
|
meta_dict: dict = d[self.meta_keys or f"{self.ref_image}_{self.meta_key_postfix}"] |
|
|
current_shape = d[self.ref_image].shape[1:] |
|
|
cropped_shape = meta_dict[self.cropped_shape_key][1:] |
|
|
factor = np.divide(current_shape, cropped_shape) |
|
|
|
|
|
pos_clicks, neg_clicks = guidance[0], guidance[1] |
|
|
pos = np.multiply(pos_clicks, factor).astype(int, copy=False).tolist() if len(pos_clicks) else [] |
|
|
neg = np.multiply(neg_clicks, factor).astype(int, copy=False).tolist() if len(neg_clicks) else [] |
|
|
|
|
|
d[self.guidance] = [pos, neg] |
|
|
return d |
|
|
|
|
|
|
|
|
class RestoreLabeld(MapTransform): |
|
|
""" |
|
|
Restores label based on the ref image. |
|
|
|
|
|
The ref_image is assumed that it went through the following transforms: |
|
|
|
|
|
1. Fetch2DSliced (If 2D) |
|
|
2. Spacingd |
|
|
3. SpatialCropGuidanced |
|
|
4. Resized |
|
|
|
|
|
And its shape is assumed to be (C, D, H, W) |
|
|
|
|
|
This transform tries to undo these operation so that the result label can be overlapped with original volume. |
|
|
It does the following operation: |
|
|
|
|
|
1. Undo Resized |
|
|
2. Undo SpatialCropGuidanced |
|
|
3. Undo Spacingd |
|
|
4. Undo Fetch2DSliced |
|
|
|
|
|
The resulting label is of shape (D, H, W) |
|
|
|
|
|
Args: |
|
|
keys: keys of the corresponding items to be transformed. |
|
|
ref_image: reference image to fetch current and original image details |
|
|
slice_only: apply only to an applicable slice, in case of 2D model/prediction |
|
|
mode: {``"constant"``, ``"edge"``, ``"linear_ramp"``, ``"maximum"``, ``"mean"``, |
|
|
``"median"``, ``"minimum"``, ``"reflect"``, ``"symmetric"``, ``"wrap"``, ``"empty"``} |
|
|
One of the listed string values or a user supplied function for padding. Defaults to ``"constant"``. |
|
|
See also: https://numpy.org/doc/1.18/reference/generated/numpy.pad.html |
|
|
align_corners: Geometrically, we consider the pixels of the input as squares rather than points. |
|
|
See also: https://pytorch.org/docs/stable/generated/torch.nn.functional.grid_sample.html |
|
|
It also can be a sequence of bool, each element corresponds to a key in ``keys``. |
|
|
meta_keys: explicitly indicate the key of the corresponding metadata dictionary. |
|
|
for example, for data with key `image`, the metadata by default is in `image_meta_dict`. |
|
|
the metadata is a dictionary object which contains: filename, original_shape, etc. |
|
|
it can be a sequence of string, map to the `keys`. |
|
|
if None, will try to construct meta_keys by `key_{meta_key_postfix}`. |
|
|
meta_key_postfix: if meta_key is None, use `key_{meta_key_postfix} to fetch the metadata according |
|
|
to the key data, default is `meta_dict`, the metadata is a dictionary object. |
|
|
For example, to handle key `image`, read/write affine matrices from the |
|
|
metadata `image_meta_dict` dictionary's `affine` field. |
|
|
start_coord_key: key that records the start coordinate of spatial bounding box for foreground. |
|
|
end_coord_key: key that records the end coordinate of spatial bounding box for foreground. |
|
|
original_shape_key: key that records original shape for foreground. |
|
|
cropped_shape_key: key that records cropped shape for foreground. |
|
|
allow_missing_keys: don't raise exception if key is missing. |
|
|
""" |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
keys: KeysCollection, |
|
|
ref_image: str, |
|
|
slice_only: bool = False, |
|
|
mode: Sequence[InterpolateMode | str] | InterpolateMode | str = InterpolateMode.NEAREST, |
|
|
align_corners: Sequence[bool | None] | bool | None = None, |
|
|
meta_keys: str | None = None, |
|
|
meta_key_postfix: str = DEFAULT_POST_FIX, |
|
|
start_coord_key: str = "foreground_start_coord", |
|
|
end_coord_key: str = "foreground_end_coord", |
|
|
original_shape_key: str = "foreground_original_shape", |
|
|
cropped_shape_key: str = "foreground_cropped_shape", |
|
|
allow_missing_keys: bool = False, |
|
|
) -> None: |
|
|
super().__init__(keys, allow_missing_keys) |
|
|
self.ref_image = ref_image |
|
|
self.slice_only = slice_only |
|
|
self.mode = ensure_tuple_rep(mode, len(self.keys)) |
|
|
self.align_corners = ensure_tuple_rep(align_corners, len(self.keys)) |
|
|
self.meta_keys = ensure_tuple_rep(None, len(self.keys)) if meta_keys is None else ensure_tuple(meta_keys) |
|
|
if len(self.keys) != len(self.meta_keys): |
|
|
raise ValueError("meta_keys should have the same length as keys.") |
|
|
self.meta_key_postfix = meta_key_postfix |
|
|
self.start_coord_key = start_coord_key |
|
|
self.end_coord_key = end_coord_key |
|
|
self.original_shape_key = original_shape_key |
|
|
self.cropped_shape_key = cropped_shape_key |
|
|
|
|
|
def __call__(self, data: Any) -> dict: |
|
|
d = dict(data) |
|
|
meta_dict: dict = d[f"{self.ref_image}_{self.meta_key_postfix}"] |
|
|
|
|
|
for key, mode, align_corners, meta_key in self.key_iterator(d, self.mode, self.align_corners, self.meta_keys): |
|
|
image = d[key] |
|
|
|
|
|
|
|
|
current_shape = image.shape |
|
|
cropped_shape = meta_dict[self.cropped_shape_key] |
|
|
if np.any(np.not_equal(current_shape, cropped_shape)): |
|
|
resizer = Resize(spatial_size=cropped_shape[1:], mode=mode) |
|
|
image = resizer(image, mode=mode, align_corners=align_corners) |
|
|
|
|
|
|
|
|
original_shape = meta_dict[self.original_shape_key] |
|
|
result = np.zeros(original_shape, dtype=np.float32) |
|
|
box_start = meta_dict[self.start_coord_key] |
|
|
box_end = meta_dict[self.end_coord_key] |
|
|
|
|
|
spatial_dims = min(len(box_start), len(image.shape[1:])) |
|
|
slices = tuple( |
|
|
[slice(None)] + [slice(s, e) for s, e in zip(box_start[:spatial_dims], box_end[:spatial_dims])] |
|
|
) |
|
|
result[slices] = image |
|
|
|
|
|
|
|
|
current_size = result.shape[1:] |
|
|
|
|
|
spatial_shape = list(np.roll(meta_dict["spatial_shape"], 1)) |
|
|
spatial_size = spatial_shape[-len(current_size) :] |
|
|
|
|
|
if np.any(np.not_equal(current_size, spatial_size)): |
|
|
resizer = Resize(spatial_size=spatial_size, mode=mode) |
|
|
result = resizer(result, mode=mode, align_corners=align_corners) |
|
|
|
|
|
|
|
|
slice_idx = meta_dict.get("slice_idx") |
|
|
final_result: NdarrayOrTensor |
|
|
if slice_idx is None or self.slice_only: |
|
|
final_result = result if len(result.shape) <= 3 else result[0] |
|
|
else: |
|
|
slice_idx = meta_dict["slice_idx"][0] |
|
|
final_result = np.zeros(tuple(spatial_shape)) |
|
|
final_result[slice_idx] = result |
|
|
d[key] = final_result |
|
|
|
|
|
meta_key = meta_key or f"{key}_{self.meta_key_postfix}" |
|
|
meta = d.get(meta_key) |
|
|
if meta is None: |
|
|
meta = dict() |
|
|
d[meta_key] = meta |
|
|
meta["slice_idx"] = slice_idx |
|
|
meta["affine"] = meta_dict["original_affine"] |
|
|
return d |
|
|
|
|
|
|
|
|
class Fetch2DSliced(MapTransform): |
|
|
""" |
|
|
Fetch one slice in case of a 3D volume. |
|
|
|
|
|
The volume only contains spatial coordinates. |
|
|
|
|
|
Args: |
|
|
keys: keys of the corresponding items to be transformed. |
|
|
guidance: key that represents guidance. |
|
|
axis: axis that represents slice in 3D volume. |
|
|
meta_keys: explicitly indicate the key of the corresponding metadata dictionary. |
|
|
for example, for data with key `image`, the metadata by default is in `image_meta_dict`. |
|
|
the metadata is a dictionary object which contains: filename, original_shape, etc. |
|
|
it can be a sequence of string, map to the `keys`. |
|
|
if None, will try to construct meta_keys by `key_{meta_key_postfix}`. |
|
|
meta_key_postfix: use `key_{meta_key_postfix}` to fetch the metadata according to the key data, |
|
|
default is `meta_dict`, the metadata is a dictionary object. |
|
|
For example, to handle key `image`, read/write affine matrices from the |
|
|
metadata `image_meta_dict` dictionary's `affine` field. |
|
|
allow_missing_keys: don't raise exception if key is missing. |
|
|
""" |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
keys: KeysCollection, |
|
|
guidance: str = "guidance", |
|
|
axis: int = 0, |
|
|
meta_keys: KeysCollection | None = None, |
|
|
meta_key_postfix: str = DEFAULT_POST_FIX, |
|
|
allow_missing_keys: bool = False, |
|
|
): |
|
|
super().__init__(keys, allow_missing_keys) |
|
|
self.guidance = guidance |
|
|
self.axis = axis |
|
|
self.meta_keys = ensure_tuple_rep(None, len(self.keys)) if meta_keys is None else ensure_tuple(meta_keys) |
|
|
if len(self.keys) != len(self.meta_keys): |
|
|
raise ValueError("meta_keys should have the same length as keys.") |
|
|
self.meta_key_postfix = ensure_tuple_rep(meta_key_postfix, len(self.keys)) |
|
|
|
|
|
def _apply(self, image, guidance): |
|
|
slice_idx = guidance[2] |
|
|
idx = [] |
|
|
for i, size_i in enumerate(image.shape): |
|
|
idx.append(slice_idx) if i == self.axis else idx.append(slice(0, size_i)) |
|
|
|
|
|
return image[tuple(idx)], tuple(idx) |
|
|
|
|
|
def __call__(self, data): |
|
|
d = dict(data) |
|
|
guidance = d[self.guidance] |
|
|
if len(guidance) < 3: |
|
|
raise RuntimeError("Guidance does not container slice_idx!") |
|
|
for key, meta_key, meta_key_postfix in self.key_iterator(d, self.meta_keys, self.meta_key_postfix): |
|
|
img_slice, idx = self._apply(d[key], guidance) |
|
|
d[key] = img_slice |
|
|
d[meta_key or f"{key}_{meta_key_postfix}"]["slice_idx"] = idx |
|
|
return d |
|
|
|