# Copyright 2020 Division of Medical Image Computing, German Cancer Research Center (DKFZ), Heidelberg, Germany # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import multiprocessing import shutil from time import sleep from typing import Union, Tuple import glob import numpy as np from batchgenerators.utilities.file_and_folder_operations import * from light_training.preprocessing.cropping.cropping import crop_to_nonzero # from .default_resampling import resample_data_or_seg_to_spacing, resample_img from light_training.preprocessing.resampling.default_resampling import resample_data_or_seg_to_shape, compute_new_shape from tqdm import tqdm from light_training.preprocessing.normalization.default_normalization_schemes import CTNormalization, ZScoreNormalization import SimpleITK as sitk from tqdm import tqdm from copy import deepcopy import json def create_image(image_arr, spacing): image = sitk.GetImageFromArray(image_arr) image.SetSpacing(spacing) return image def get_shape_must_be_divisible_by(net_numpool_per_axis): return 2 ** np.array(net_numpool_per_axis) def pad_shape(shape, must_be_divisible_by): """ pads shape so that it is divisible by must_be_divisible_by :param shape: :param must_be_divisible_by: :return: """ if not isinstance(must_be_divisible_by, (tuple, list, np.ndarray)): must_be_divisible_by = [must_be_divisible_by] * len(shape) else: assert len(must_be_divisible_by) == len(shape) new_shp = [shape[i] + must_be_divisible_by[i] - shape[i] % must_be_divisible_by[i] for i in range(len(shape))] for i in range(len(shape)): if shape[i] % must_be_divisible_by[i] == 0: new_shp[i] -= must_be_divisible_by[i] new_shp = np.array(new_shp).astype(int) return new_shp def get_pool_and_conv_props(spacing, patch_size, min_feature_map_size, max_numpool): """ this is the same as get_pool_and_conv_props_v2 from old nnunet :param spacing: :param patch_size: :param min_feature_map_size: min edge length of feature maps in bottleneck :param max_numpool: :return: """ # todo review this code dim = len(spacing) current_spacing = deepcopy(list(spacing)) current_size = deepcopy(list(patch_size)) pool_op_kernel_sizes = [[1] * len(spacing)] conv_kernel_sizes = [] num_pool_per_axis = [0] * dim kernel_size = [1] * dim while True: # exclude axes that we cannot pool further because of min_feature_map_size constraint valid_axes_for_pool = [i for i in range(dim) if current_size[i] >= 2*min_feature_map_size] if len(valid_axes_for_pool) < 1: break spacings_of_axes = [current_spacing[i] for i in valid_axes_for_pool] # find axis that are within factor of 2 within smallest spacing min_spacing_of_valid = min(spacings_of_axes) valid_axes_for_pool = [i for i in valid_axes_for_pool if current_spacing[i] / min_spacing_of_valid < 2] # max_numpool constraint valid_axes_for_pool = [i for i in valid_axes_for_pool if num_pool_per_axis[i] < max_numpool] if len(valid_axes_for_pool) == 1: if current_size[valid_axes_for_pool[0]] >= 3 * min_feature_map_size: pass else: break if len(valid_axes_for_pool) < 1: break # now we need to find kernel sizes # kernel sizes are initialized to 1. They are successively set to 3 when their associated axis becomes within # factor 2 of min_spacing. Once they are 3 they remain 3 for d in range(dim): if kernel_size[d] == 3: continue else: if spacings_of_axes[d] / min(current_spacing) < 2: kernel_size[d] = 3 other_axes = [i for i in range(dim) if i not in valid_axes_for_pool] pool_kernel_sizes = [0] * dim for v in valid_axes_for_pool: pool_kernel_sizes[v] = 2 num_pool_per_axis[v] += 1 current_spacing[v] *= 2 current_size[v] = np.ceil(current_size[v] / 2) for nv in other_axes: pool_kernel_sizes[nv] = 1 pool_op_kernel_sizes.append(pool_kernel_sizes) conv_kernel_sizes.append(deepcopy(kernel_size)) #print(conv_kernel_sizes) must_be_divisible_by = get_shape_must_be_divisible_by(num_pool_per_axis) patch_size = pad_shape(patch_size, must_be_divisible_by) # we need to add one more conv_kernel_size for the bottleneck. We always use 3x3(x3) conv here conv_kernel_sizes.append([3]*dim) return num_pool_per_axis, pool_op_kernel_sizes, conv_kernel_sizes, patch_size, must_be_divisible_by class DefaultPreprocessor(object): def __init__(self, base_dir, image_dir, # output_dir, # out_spacing, label_dir=None, data_type="CT"): """ Everything we need is in the plans. Those are given when run() is called """ self.base_dir = base_dir self.image_dir = image_dir self.label_dir = label_dir self.data_type = data_type def run_case_npy(self, data: np.ndarray, seg, properties: dict): # let's not mess up the inputs! data = np.copy(data) old_shape = data.shape original_spacing = list(properties['spacing']) ## 由于old spacing读出来是反的,因此这里需要转置一下 original_spacing_trans = original_spacing[::-1] properties["original_spacing_trans"] = original_spacing_trans properties["target_spacing_trans"] = self.out_spacing shape_before_cropping = data.shape[1:] ## crop properties['shape_before_cropping'] = shape_before_cropping # this command will generate a segmentation. This is important because of the nonzero mask which we may need data, seg, bbox = crop_to_nonzero(data, seg) properties['bbox_used_for_cropping'] = bbox # crop, remember to store size before cropping! shape_before_resample = data.shape[1:] properties['shape_after_cropping_before_resample'] = shape_before_resample new_shape = compute_new_shape(data.shape[1:], original_spacing_trans, self.out_spacing) if seg is None : seg_norm = np.zeros_like(data) else : seg_norm = seg data = self._normalize(data, seg_norm, self.foreground_intensity_properties_per_channel) assert len(data.shape) == 4 data = resample_data_or_seg_to_shape(data, new_shape, original_spacing, self.out_spacing, order=3, order_z=0) properties['shape_after_resample'] = new_shape if seg is not None : assert len(seg.shape) == 4 seg = resample_data_or_seg_to_shape(seg, new_shape, original_spacing, self.out_spacing, is_seg=True, order=1, order_z=0) properties['class_locations'] = self._sample_foreground_locations(seg, self.all_labels, ) if np.max(seg) > 127: seg = seg.astype(np.int16) else: seg = seg.astype(np.int8) # assert len(seg.shape) == 4 # seg = create_image(seg[0], original_spacing) # seg = resample_img(seg, out_spacing=self.out_spacing, is_label=True) # seg = sitk.GetArrayFromImage(seg)[None] # print(f"all_labels is {np.unique(seg)}") # if np.max(seg) > 127: # seg = seg.astype(np.int16) # else: # seg = seg.astype(np.int8) # seg = resample_data_or_seg_to_spacing(seg, current_spacing=original_spacing, # new_spacing=self.out_spacing, is_seg=True) print(f'old shape: {old_shape}, new_shape after crop and resample: {new_shape}, old_spacing: {original_spacing}, ' f'new_spacing: {self.out_spacing}, boxes is {bbox}') return data, seg # need to modify def get_iterable_list(self): all_cases = os.listdir(os.path.join(self.base_dir, self.image_dir)) return all_cases def _normalize(self, data: np.ndarray, seg: np.ndarray, foreground_intensity_properties_per_channel: dict) -> np.ndarray: for c in range(data.shape[0]): normalizer_class = CTNormalization normalizer = normalizer_class(use_mask_for_norm=False, intensityproperties=foreground_intensity_properties_per_channel[str(c)]) data[c] = normalizer.run(data[c], seg[0]) return data # need to modify def read_data(self, case_name): ## only for CT dataset data = sitk.ReadImage(os.path.join(self.base_dir, self.image_dir, case_name)) seg_arr = None ## 一定要是float32!!!! data_arr = sitk.GetArrayFromImage(data).astype(np.float32) data_arr = data_arr[None] if self.label_dir is not None: seg = sitk.ReadImage(os.path.join(self.base_dir, self.label_dir, case_name)) ## 读出来以后一定转float32!!! seg_arr = sitk.GetArrayFromImage(seg).astype(np.float32) seg_arr = seg_arr[None] intensities_per_channel, intensity_statistics_per_channel = self.collect_foreground_intensities(seg_arr, data_arr) else : intensities_per_channel = [] intensity_statistics_per_channel = [] properties = {"spacing": data.GetSpacing(), "raw_size": data_arr.shape[1:], "name": case_name.split(".")[0], "intensities_per_channel": intensities_per_channel, "intensity_statistics_per_channel": intensity_statistics_per_channel} return data_arr, seg_arr, properties def run_case(self, case_name): """ seg file can be none (test cases) order of operations is: transpose -> crop -> resample so when we export we need to run the following order: resample -> crop -> transpose (we could also run transpose at a different place, but reverting the order of operations done during preprocessing seems cleaner) """ data, seg, properties = self.read_data(case_name) data, seg = self.run_case_npy(data, seg, properties) return data, seg, properties def run_case_save(self, case_name): print(case_name + "~~~~~~~~" * 10) data, seg, properties = self.run_case(case_name) # print('dtypes', data.dtype, seg.dtype) case_name = case_name.split(".")[0] np.savez_compressed(os.path.join(self.output_dir, case_name) + '.npz', data=data, seg=seg) write_pickle(properties, os.path.join(self.output_dir, case_name) + '.pkl') print(f"data is saved at: {os.path.join(self.output_dir, case_name) + '.npz'}") def experiment_plan(self, case_name): data, seg, properties = self.read_data(case_name) print(f"labels is {np.unique(seg)}") spacing = properties["spacing"] raw_size = properties["raw_size"] intensities_per_channel = properties["intensities_per_channel"] return spacing, raw_size, intensities_per_channel def determine_fullres_target_spacing(self, spacings, sizes) -> np.ndarray: # if self.overwrite_target_spacing is not None: # return np.array(self.overwrite_target_spacing) # spacings = self.dataset_fingerprint['spacings'] # sizes = self.dataset_fingerprint['shapes_after_crop'] target = np.percentile(np.vstack(spacings), 50, 0) target_size = np.percentile(np.vstack(sizes), 50, 0) # we need to identify datasets for which a different target spacing could be beneficial. These datasets have # the following properties: # - one axis which much lower resolution than the others # - the lowres axis has much less voxels than the others # - (the size in mm of the lowres axis is also reduced) worst_spacing_axis = np.argmax(target) other_axes = [i for i in range(len(target)) if i != worst_spacing_axis] other_spacings = [target[i] for i in other_axes] other_sizes = [target_size[i] for i in other_axes] has_aniso_spacing = target[worst_spacing_axis] > (3 * max(other_spacings)) has_aniso_voxels = target_size[worst_spacing_axis] * 3 < min(other_sizes) if has_aniso_spacing and has_aniso_voxels: spacings_of_that_axis = np.vstack(spacings)[:, worst_spacing_axis] target_spacing_of_that_axis = np.percentile(spacings_of_that_axis, 10) # don't let the spacing of that axis get higher than the other axes if target_spacing_of_that_axis < max(other_spacings): target_spacing_of_that_axis = max(max(other_spacings), target_spacing_of_that_axis) + 1e-5 target[worst_spacing_axis] = target_spacing_of_that_axis return target def compute_new_shape(self, old_shape: Union[Tuple[int, ...], List[int], np.ndarray], old_spacing: Union[Tuple[float, ...], List[float], np.ndarray], new_spacing: Union[Tuple[float, ...], List[float], np.ndarray]) -> np.ndarray: ## spacing need to be transposed old_spacing = list(old_spacing)[::-1] new_spacing = list(new_spacing)[::-1] assert len(old_spacing) == len(old_shape) assert len(old_shape) == len(new_spacing) new_shape = np.array([int(round(i / j * k)) for i, j, k in zip(old_spacing, new_spacing, old_shape)]) return new_shape def run_plan(self): all_iter = self.get_iterable_list() spacings = [] sizes = [] intensities_per_channels = [] print(f"analysing data......") for case in tqdm(all_iter, total=len(all_iter)): spacing, size, intensities_per_channel = self.experiment_plan(case) spacings.append(spacing) sizes.append(size) intensities_per_channels.append(intensities_per_channel) print(f"all spacing is {spacings}") print(f"all sizes is {sizes}") foreground_intensities_per_channel = [np.concatenate([r[i] for r in intensities_per_channels]) for i in range(len(intensities_per_channels[0]))] num_channels = len(intensities_per_channels[0]) intensity_statistics_per_channel = {} for i in range(num_channels): intensity_statistics_per_channel[i] = { 'mean': float(np.mean(foreground_intensities_per_channel[i])), 'median': float(np.median(foreground_intensities_per_channel[i])), 'std': float(np.std(foreground_intensities_per_channel[i])), 'min': float(np.min(foreground_intensities_per_channel[i])), 'max': float(np.max(foreground_intensities_per_channel[i])), 'percentile_99_5': float(np.percentile(foreground_intensities_per_channel[i], 99.5)), 'percentile_00_5': float(np.percentile(foreground_intensities_per_channel[i], 0.5)), } print(f"intensity_statistics_per_channel is {intensity_statistics_per_channel}") fullres_spacing = self.determine_fullres_target_spacing(spacings, sizes) print(f"fullres spacing is {fullres_spacing[::-1]}") # get transposed new median shape (what we would have after resampling) new_shapes = [self.compute_new_shape(j, i, fullres_spacing) for i, j in zip(spacings, sizes)] new_median_shape = np.median(new_shapes, 0) print(f"median_shape is {new_median_shape}") tmp = 1 / np.array(fullres_spacing) initial_patch_size = [round(i) for i in tmp * (256 ** 3 / np.prod(tmp)) ** (1 / 3)] print(f"initial_patch_size is {initial_patch_size[::-1]}") network_num_pool_per_axis, pool_op_kernel_sizes, conv_kernel_sizes, patch_size, \ shape_must_be_divisible_by = get_pool_and_conv_props(fullres_spacing, initial_patch_size, 4, 999999) print(f"target medium patch size is {patch_size[::-1]}") analysis_path = "./data_analysis_result.txt" with open(analysis_path, "w") as f: f.write(json.dumps({ "intensity_statistics_per_channel": intensity_statistics_per_channel, "fullres spacing": fullres_spacing.tolist(), "median_shape": new_median_shape.tolist(), "initial_patch_size": initial_patch_size, "target medium patch size": patch_size[::-1].tolist() })) print(f"Analysis done, save to {analysis_path}") def collect_foreground_intensities(self, segmentation: np.ndarray, images: np.ndarray, seed: int = 1234, num_samples: int = 10000): """ images=image with multiple channels = shape (c, x, y(, z)) """ assert len(images.shape) == 4 assert len(segmentation.shape) == 4 assert not np.any(np.isnan(segmentation)), "Segmentation contains NaN values. grrrr.... :-(" assert not np.any(np.isnan(images)), "Images contains NaN values. grrrr.... :-(" rs = np.random.RandomState(seed) intensities_per_channel = [] # we don't use the intensity_statistics_per_channel at all, it's just something that might be nice to have intensity_statistics_per_channel = [] # segmentation is 4d: 1,x,y,z. We need to remove the empty dimension for the following code to work foreground_mask = segmentation[0] > 0 for i in range(len(images)): foreground_pixels = images[i][foreground_mask] num_fg = len(foreground_pixels) # sample with replacement so that we don't get issues with cases that have less than num_samples # foreground_pixels. We could also just sample less in those cases but that would than cause these # training cases to be underrepresented intensities_per_channel.append( rs.choice(foreground_pixels, num_samples, replace=True) if num_fg > 0 else []) intensity_statistics_per_channel.append({ 'mean': np.mean(foreground_pixels) if num_fg > 0 else np.nan, 'median': np.median(foreground_pixels) if num_fg > 0 else np.nan, 'min': np.min(foreground_pixels) if num_fg > 0 else np.nan, 'max': np.max(foreground_pixels) if num_fg > 0 else np.nan, 'percentile_99_5': np.percentile(foreground_pixels, 99.5) if num_fg > 0 else np.nan, 'percentile_00_5': np.percentile(foreground_pixels, 0.5) if num_fg > 0 else np.nan, }) return intensities_per_channel, intensity_statistics_per_channel @staticmethod def _sample_foreground_locations(seg: np.ndarray, classes_or_regions: Union[List[int], List[Tuple[int, ...]]], seed: int = 1234, verbose: bool = False): num_samples = 10000 min_percent_coverage = 0.01 # at least 1% of the class voxels need to be selected, otherwise it may be too # sparse rndst = np.random.RandomState(seed) class_locs = {} for c in classes_or_regions: k = c if not isinstance(c, list) else tuple(c) if isinstance(c, (tuple, list)): ## region mask = seg == c[0] for cc in c[1:]: mask = mask | (seg == cc) all_locs = np.argwhere(mask) else: all_locs = np.argwhere(seg == c) if len(all_locs) == 0: class_locs[k] = [] continue target_num_samples = min(num_samples, len(all_locs)) target_num_samples = max(target_num_samples, int(np.ceil(len(all_locs) * min_percent_coverage))) selected = all_locs[rndst.choice(len(all_locs), target_num_samples, replace=False)] class_locs[k] = selected if verbose: print(c, target_num_samples) return class_locs def run(self, output_spacing, output_dir, all_labels, foreground_intensity_properties_per_channel=None, num_processes=8): self.out_spacing = output_spacing self.all_labels = all_labels self.output_dir = output_dir self.foreground_intensity_properties_per_channel = foreground_intensity_properties_per_channel all_iter = self.get_iterable_list() maybe_mkdir_p(self.output_dir) # test_run for case_name in all_iter: self.run_case_save(case_name) break # multiprocessing magic. r = [] with multiprocessing.get_context("spawn").Pool(num_processes) as p: for case_name in all_iter: r.append(p.starmap_async(self.run_case_save, ((case_name, ),))) remaining = list(range(len(all_iter))) # p is pretty nifti. If we kill workers they just respawn but don't do any work. # So we need to store the original pool of workers. workers = [j for j in p._pool] with tqdm(desc=None, total=len(all_iter)) as pbar: while len(remaining) > 0: all_alive = all([j.is_alive() for j in workers]) if not all_alive: raise RuntimeError('Some background worker is 6 feet under. Yuck. \n' 'OK jokes aside.\n' 'One of your background processes is missing. This could be because of ' 'an error (look for an error message) or because it was killed ' 'by your OS due to running out of RAM. If you don\'t see ' 'an error message, out of RAM is likely the problem. In that case ' 'reducing the number of workers might help') done = [i for i in remaining if r[i].ready()] for _ in done: pbar.update() remaining = [i for i in remaining if i not in done] sleep(0.1)