Datasets:

ArXiv:
emad2001's picture
Upload folder using huggingface_hub
b4d7ac8 verified
import torch
import numpy as np
import SimpleITK
import os
import sys
from monai.inferers import SlidingWindowInferer
class Customalgorithm(): # SegmentationAlgorithm is not inherited in this class anymore
def __init__(self):
"""
Do not modify the `self.input_dir` and `self.output_dir`.
(Check https://grand-challenge.org/algorithms/interfaces/)
"""
self.input_dir = "/input/"
self.output_dir = "/output/images/head-neck-segmentation/"
# self.out_spacing = [3.0, 0.54199219, 0.54199219]
self.out_spacing = [3.0, 1.0, 1.0]
# self.device = "cpu"
self.device = torch.device("cuda")
self.patch_size = [64, 128, 128]
def filte_state_dict(self, sd):
if "module" in sd :
sd = sd["module"]
new_sd = {}
for k, v in sd.items():
k = str(k)
new_k = k[7:] if k.startswith("module") else k
new_sd[new_k] = v
del sd
return new_sd
def convert_mha_to_nii(self, mha_input_path, nii_out_path): # nnUNet specific
img = SimpleITK.ReadImage(mha_input_path)
print(img.GetSize())
SimpleITK.WriteImage(img, nii_out_path, True)
def convert_nii_to_mha(self, nii_input_path, mha_out_path): # nnUNet specific
img = SimpleITK.ReadImage(nii_input_path)
SimpleITK.WriteImage(img, mha_out_path, True)
def read(self, mha_path):
img = SimpleITK.ReadImage(mha_path)
spacing = img.GetSpacing()
raw_size = SimpleITK.GetArrayFromImage(img).shape
img = SimpleITK.GetArrayFromImage(img)[None,].astype(np.float32)
properties = {
"spacing": spacing,
"raw_size": raw_size
}
return img, properties
def check_gpu(self):
"""
Check if GPU is available. Note that the Grand Challenge only has one available GPU.
"""
print('Checking GPU availability')
is_available = torch.cuda.is_available()
print('Available: ' + str(is_available))
print(f'Device count: {torch.cuda.device_count()}')
if is_available:
print(f'Current device: {torch.cuda.current_device()}')
print('Device name: ' + torch.cuda.get_device_name(0))
print('Device memory: ' +
str(torch.cuda.get_device_properties(0).total_memory))
def load_inputs(self): # use two modalities input data
"""
Read input data (two modalities) from `self.input_dir` (/input/).
Please do not modify the path for CT and contrast-CT images.
"""
ct_mha = os.listdir(os.path.join(self.input_dir, 'images/head-neck-ct/'))[0]
ctc_mha = os.listdir(os.path.join(self.input_dir, 'images/head-neck-contrast-enhanced-ct/'))[0]
uuid = os.path.splitext(ct_mha)[0]
img, properties = self.read(os.path.join(self.input_dir, 'images/head-neck-ct/', ct_mha))
img_c, _ = self.read(os.path.join(self.input_dir, 'images/head-neck-contrast-enhanced-ct/', ctc_mha))
data = np.concatenate([img, img_c], axis=0)
del img
del img_c
# data is (2, d, w, h)
return uuid, data, properties
def crop(self, data, properties):
from light_training.preprocessing.cropping.cropping import crop_to_nonzero
seg = np.zeros_like(data)
shape_before_cropping = data.shape[1:]
## crop
properties['shape_before_cropping'] = shape_before_cropping
# this command will generate a segmentation. This is important because of the nonzero mask which we may need
data, seg, bbox = crop_to_nonzero(data, seg)
del seg
properties['bbox_used_for_cropping'] = bbox
return data, properties
def resample(self, data, properties):
from light_training.preprocessing.resampling.default_resampling import compute_new_shape, resample_data_or_seg_to_shape
# crop, remember to store size before cropping!
shape_before_resample = data.shape[1:]
properties['shape_after_cropping_before_resample'] = shape_before_resample
new_shape = compute_new_shape(data.shape[1:], original_spacing_trans, self.out_spacing)
assert len(data.shape) == 4
data = resample_data_or_seg_to_shape(data, new_shape,
original_spacing,
self.out_spacing,
order=3,
order_z=0)
properties['shape_after_resample'] = new_shape
return data, properties
def preprocess(self, data, properties, crop_first=True):
from light_training.process_framework.norm import norm_func
original_spacing = list(properties['spacing'])
## 由于old spacing读出来是反的,因此这里需要转置一下
original_spacing_trans = original_spacing[::-1]
properties["original_spacing_trans"] = original_spacing_trans
properties["target_spacing_trans"] = self.out_spacing
if crop_first:
data, properties = self.crop(data, properties)
data = norm_func(data)
if not crop_first:
data, properties = self.crop(data, properties)
data, properties = self.resample(data, properties)
data = data[None,]
data = torch.from_numpy(data)
return data, properties
def predict(self, data, properties, uid):
torch.cuda.empty_cache()
from models.nnunet3d import NNUNetWrapper
model = NNUNetWrapper(norm="ins")
new_sd = self.filte_state_dict(torch.load("./weight/unet3d_0_addaug_bs2_ep1000_ds_gpu4/final_model_0.8552.pt", map_location="cpu"))
model.load_state_dict(new_sd)
del new_sd
torch.cuda.empty_cache()
# data = data.to(self.deivce)
# model.to(self.device)
model.eval()
window_infer = SlidingWindowInferer(roi_size=self.patch_size,
sw_batch_size=1,
overlap=0.5,
progress=True,
mode="gaussian")
predictor = Predictor(window_infer, mirror_axes=None)
try:
ensemble_output = predictor.maybe_mirror_and_predict(data, model, self.device)
except RuntimeError:
ensemble_output = predictor.maybe_mirror_and_predict(data, model, torch.device("cpu"))
torch.cuda.empty_cache()
del model
del data
print(f"prediction done")
ensemble_output = predictor.predict_raw_probability(ensemble_output, properties)
print(f"non linear....")
# ensemble_output = predictor.apply_nonlinear(ensemble_output, nonlinear_type="sigmoid")
ensemble_output = ensemble_output > 0
print(f"restore crop...")
ensemble_output = predictor.predict_noncrop_probability(ensemble_output, properties)
raw_spacing = properties["spacing"]
case_name = uid
print(f"uuid is {uid}")
os.makedirs(os.path.dirname(self.output_dir), exist_ok=True)
print(f"saving....")
predictor.save_to_nii_multi_organ(ensemble_output,
raw_spacing,
save_dir=self.output_dir,
case_name=case_name,
postprocess=False)
# """
# load the model and checkpoint, and generate the predictions. You can replace this part with your own model.
# """
# predict_from_folder_segrap2023(self.weight, self.nii_path, self.result_path, 0, 0, 1)
# print("nnUNet segmentation done!")
# if not os.path.exists(os.path.join(self.result_path, self.nii_seg_file)):
# print('waiting for nnUNet segmentation to be created')
# while not os.path.exists(os.path.join(self.result_path, self.nii_seg_file)):
# import time
# print('.', end='')
# time.sleep(5)
# # print(cproc) # since nnUNet_predict call is split into prediction and postprocess, a pre-mature exit code is received but segmentation file not yet written. This hack ensures that all spawned subprocesses are finished before being printed.
# print('Prediction finished !')
def post_process(self):
self.check_gpu()
print('Start processing')
uuid, data, properties = self.load_inputs()
data, properties = self.preprocess(data, properties)
print(properties)
print('Start prediction')
self.predict(data, properties, uuid)
# print('Start output writing')
# self.write_outputs(uuid)
def process(self):
"""
Read inputs from /input, process with your algorithm and write to /output
"""
self.post_process()
if __name__ == "__main__":
Customalgorithm().process()