|
|
from collections import OrderedDict |
|
|
from typing import Union, Tuple, List |
|
|
|
|
|
import numpy as np |
|
|
import pandas as pd |
|
|
import torch |
|
|
from batchgenerators.augmentations.utils import resize_segmentation |
|
|
from scipy.ndimage.interpolation import map_coordinates |
|
|
from skimage.transform import resize |
|
|
|
|
|
ANISO_THRESHOLD = 3 |
|
|
|
|
|
def get_do_separate_z(spacing: Union[Tuple[float, ...], List[float], np.ndarray], anisotropy_threshold=ANISO_THRESHOLD): |
|
|
do_separate_z = (np.max(spacing) / np.min(spacing)) > anisotropy_threshold |
|
|
return do_separate_z |
|
|
|
|
|
|
|
|
def get_lowres_axis(new_spacing: Union[Tuple[float, ...], List[float], np.ndarray]): |
|
|
axis = np.where(max(new_spacing) / np.array(new_spacing) == 1)[0] |
|
|
return axis |
|
|
|
|
|
|
|
|
def compute_new_shape(old_shape: Union[Tuple[int, ...], List[int], np.ndarray], |
|
|
old_spacing: Union[Tuple[float, ...], List[float], np.ndarray], |
|
|
new_spacing: Union[Tuple[float, ...], List[float], np.ndarray]) -> np.ndarray: |
|
|
assert len(old_spacing) == len(old_shape) |
|
|
assert len(old_shape) == len(new_spacing) |
|
|
|
|
|
new_shape = np.array([int(round(i / j * k)) for i, j, k in zip(old_spacing, new_spacing, old_shape)]) |
|
|
return new_shape |
|
|
|
|
|
|
|
|
def resample_data_or_seg_to_spacing(data: np.ndarray, |
|
|
current_spacing: Union[Tuple[float, ...], List[float], np.ndarray], |
|
|
new_spacing: Union[Tuple[float, ...], List[float], np.ndarray], |
|
|
is_seg: bool = False, |
|
|
order: int = 3, order_z: int = 0, |
|
|
force_separate_z: Union[bool, None] = False, |
|
|
separate_z_anisotropy_threshold: float = ANISO_THRESHOLD): |
|
|
if force_separate_z is not None: |
|
|
do_separate_z = force_separate_z |
|
|
if force_separate_z: |
|
|
axis = get_lowres_axis(current_spacing) |
|
|
else: |
|
|
axis = None |
|
|
else: |
|
|
if get_do_separate_z(current_spacing, separate_z_anisotropy_threshold): |
|
|
do_separate_z = True |
|
|
axis = get_lowres_axis(current_spacing) |
|
|
elif get_do_separate_z(new_spacing, separate_z_anisotropy_threshold): |
|
|
do_separate_z = True |
|
|
axis = get_lowres_axis(new_spacing) |
|
|
else: |
|
|
do_separate_z = False |
|
|
axis = None |
|
|
|
|
|
if axis is not None: |
|
|
if len(axis) == 3: |
|
|
|
|
|
do_separate_z = False |
|
|
elif len(axis) == 2: |
|
|
|
|
|
|
|
|
do_separate_z = False |
|
|
else: |
|
|
pass |
|
|
|
|
|
if data is not None: |
|
|
assert len(data.shape) == 4, "data must be c x y z" |
|
|
|
|
|
shape = np.array(data[0].shape) |
|
|
new_shape = compute_new_shape(shape[1:], current_spacing, new_spacing) |
|
|
|
|
|
data_reshaped = resample_data_or_seg(data, new_shape, is_seg, axis, order, do_separate_z, order_z=order_z) |
|
|
return data_reshaped |
|
|
|
|
|
|
|
|
def resample_data_or_seg_to_shape(data: Union[torch.Tensor, np.ndarray], |
|
|
new_shape: Union[Tuple[int, ...], List[int], np.ndarray], |
|
|
current_spacing: Union[Tuple[float, ...], List[float], np.ndarray], |
|
|
new_spacing: Union[Tuple[float, ...], List[float], np.ndarray], |
|
|
is_seg: bool = False, |
|
|
order: int = 3, order_z: int = 0, |
|
|
force_separate_z: Union[bool, None] = False, |
|
|
separate_z_anisotropy_threshold: float = ANISO_THRESHOLD): |
|
|
""" |
|
|
needed for segmentation export. Stupid, I know. Maybe we can fix that with Leos new resampling functions |
|
|
""" |
|
|
if isinstance(data, torch.Tensor): |
|
|
data = data.cpu().numpy() |
|
|
if force_separate_z is not None: |
|
|
do_separate_z = force_separate_z |
|
|
if force_separate_z: |
|
|
axis = get_lowres_axis(current_spacing) |
|
|
else: |
|
|
axis = None |
|
|
else: |
|
|
if get_do_separate_z(current_spacing, separate_z_anisotropy_threshold): |
|
|
do_separate_z = True |
|
|
axis = get_lowres_axis(current_spacing) |
|
|
elif get_do_separate_z(new_spacing, separate_z_anisotropy_threshold): |
|
|
do_separate_z = True |
|
|
axis = get_lowres_axis(new_spacing) |
|
|
else: |
|
|
do_separate_z = False |
|
|
axis = None |
|
|
|
|
|
if axis is not None: |
|
|
if len(axis) == 3: |
|
|
|
|
|
do_separate_z = False |
|
|
elif len(axis) == 2: |
|
|
|
|
|
|
|
|
do_separate_z = False |
|
|
else: |
|
|
pass |
|
|
|
|
|
if data is not None: |
|
|
assert len(data.shape) == 4, "data must be c x y z" |
|
|
|
|
|
data_reshaped = resample_data_or_seg(data, new_shape, is_seg, axis, order, do_separate_z, order_z=order_z) |
|
|
return data_reshaped |
|
|
|
|
|
|
|
|
def resample_data_or_seg(data: np.ndarray, new_shape: Union[Tuple[float, ...], List[float], np.ndarray], |
|
|
is_seg: bool = False, axis: Union[None, int] = None, order: int = 3, |
|
|
do_separate_z: bool = False, order_z: int = 0): |
|
|
""" |
|
|
separate_z=True will resample with order 0 along z |
|
|
:param data: |
|
|
:param new_shape: |
|
|
:param is_seg: |
|
|
:param axis: |
|
|
:param order: |
|
|
:param do_separate_z: |
|
|
:param order_z: only applies if do_separate_z is True |
|
|
:return: |
|
|
""" |
|
|
assert len(data.shape) == 4, "data must be (c, x, y, z)" |
|
|
assert len(new_shape) == len(data.shape) - 1 |
|
|
|
|
|
if is_seg: |
|
|
resize_fn = resize_segmentation |
|
|
kwargs = OrderedDict() |
|
|
else: |
|
|
resize_fn = resize |
|
|
kwargs = {'mode': 'edge', 'anti_aliasing': False} |
|
|
dtype_data = data.dtype |
|
|
shape = np.array(data[0].shape) |
|
|
new_shape = np.array(new_shape) |
|
|
if np.any(shape != new_shape): |
|
|
data = data.astype(float) |
|
|
if do_separate_z: |
|
|
|
|
|
assert len(axis) == 1, "only one anisotropic axis supported" |
|
|
axis = axis[0] |
|
|
if axis == 0: |
|
|
new_shape_2d = new_shape[1:] |
|
|
elif axis == 1: |
|
|
new_shape_2d = new_shape[[0, 2]] |
|
|
else: |
|
|
new_shape_2d = new_shape[:-1] |
|
|
|
|
|
reshaped_final_data = [] |
|
|
for c in range(data.shape[0]): |
|
|
reshaped_data = [] |
|
|
for slice_id in range(shape[axis]): |
|
|
if axis == 0: |
|
|
reshaped_data.append(resize_fn(data[c, slice_id], new_shape_2d, order, **kwargs)) |
|
|
elif axis == 1: |
|
|
reshaped_data.append(resize_fn(data[c, :, slice_id], new_shape_2d, order, **kwargs)) |
|
|
else: |
|
|
reshaped_data.append(resize_fn(data[c, :, :, slice_id], new_shape_2d, order, **kwargs)) |
|
|
reshaped_data = np.stack(reshaped_data, axis) |
|
|
if shape[axis] != new_shape[axis]: |
|
|
|
|
|
|
|
|
rows, cols, dim = new_shape[0], new_shape[1], new_shape[2] |
|
|
orig_rows, orig_cols, orig_dim = reshaped_data.shape |
|
|
|
|
|
row_scale = float(orig_rows) / rows |
|
|
col_scale = float(orig_cols) / cols |
|
|
dim_scale = float(orig_dim) / dim |
|
|
|
|
|
map_rows, map_cols, map_dims = np.mgrid[:rows, :cols, :dim] |
|
|
map_rows = row_scale * (map_rows + 0.5) - 0.5 |
|
|
map_cols = col_scale * (map_cols + 0.5) - 0.5 |
|
|
map_dims = dim_scale * (map_dims + 0.5) - 0.5 |
|
|
|
|
|
coord_map = np.array([map_rows, map_cols, map_dims]) |
|
|
if not is_seg or order_z == 0: |
|
|
reshaped_final_data.append(map_coordinates(reshaped_data, coord_map, order=order_z, |
|
|
mode='nearest')[None]) |
|
|
else: |
|
|
unique_labels = np.sort(pd.unique(reshaped_data.ravel())) |
|
|
reshaped = np.zeros(new_shape, dtype=dtype_data) |
|
|
|
|
|
for i, cl in enumerate(unique_labels): |
|
|
reshaped_multihot = np.round( |
|
|
map_coordinates((reshaped_data == cl).astype(float), coord_map, order=order_z, |
|
|
mode='nearest')) |
|
|
reshaped[reshaped_multihot > 0.5] = cl |
|
|
reshaped_final_data.append(reshaped[None]) |
|
|
else: |
|
|
reshaped_final_data.append(reshaped_data[None]) |
|
|
reshaped_final_data = np.vstack(reshaped_final_data) |
|
|
else: |
|
|
|
|
|
reshaped = [] |
|
|
for c in range(data.shape[0]): |
|
|
reshaped.append(resize_fn(data[c], new_shape, order, **kwargs)[None]) |
|
|
reshaped_final_data = np.vstack(reshaped) |
|
|
return reshaped_final_data.astype(dtype_data) |
|
|
else: |
|
|
|
|
|
return data |
|
|
|