|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
from __future__ import annotations |
|
|
|
|
|
import copy |
|
|
import os |
|
|
|
|
|
import numpy as np |
|
|
|
|
|
import monai |
|
|
from monai.bundle import ConfigParser |
|
|
from monai.utils import StrEnum, ensure_tuple, optional_import |
|
|
|
|
|
tqdm, has_tqdm = optional_import("tqdm", name="tqdm") |
|
|
nib, _ = optional_import("nibabel") |
|
|
|
|
|
logger = monai.apps.utils.get_logger(__name__) |
|
|
|
|
|
__all__ = ["analyze_data", "create_new_data_copy", "create_new_dataset_json", "NNUNETMode"] |
|
|
|
|
|
|
|
|
class NNUNETMode(StrEnum): |
|
|
N_2D = "2d" |
|
|
N_3D_FULLRES = "3d_fullres" |
|
|
N_3D_LOWRES = "3d_lowres" |
|
|
N_3D_CASCADE_FULLRES = "3d_cascade_fullres" |
|
|
|
|
|
|
|
|
def analyze_data(datalist_json: dict, data_dir: str) -> tuple[int, int]: |
|
|
""" |
|
|
Analyze (training) data |
|
|
|
|
|
Args: |
|
|
datalist_json: original data list .json (required by most monai tutorials). |
|
|
data_dir: raw data directory. |
|
|
""" |
|
|
img = monai.transforms.LoadImage(image_only=True, ensure_channel_first=True, simple_keys=True)( |
|
|
os.path.join(data_dir, datalist_json["training"][0]["image"]) |
|
|
) |
|
|
num_input_channels = img.size()[0] if img.dim() == 4 else 1 |
|
|
logger.info(f"num_input_channels: {num_input_channels}") |
|
|
|
|
|
num_foreground_classes = 0 |
|
|
for _i in range(len(datalist_json["training"])): |
|
|
seg = monai.transforms.LoadImage(image_only=True, ensure_channel_first=True, simple_keys=True)( |
|
|
os.path.join(data_dir, datalist_json["training"][_i]["label"]) |
|
|
) |
|
|
num_foreground_classes = max(num_foreground_classes, int(seg.max())) |
|
|
logger.info(f"num_foreground_classes: {num_foreground_classes}") |
|
|
|
|
|
return num_input_channels, num_foreground_classes |
|
|
|
|
|
|
|
|
def create_new_data_copy( |
|
|
test_key: str, datalist_json: dict, data_dir: str, num_input_channels: int, output_datafolder: str |
|
|
) -> None: |
|
|
""" |
|
|
Create and organize a new copy of data to meet the requirements of nnU-Net V2 |
|
|
|
|
|
Args: |
|
|
test_key: key for test data in the data list .json. |
|
|
datalist_json: original data list .json (required by most monai tutorials). |
|
|
data_dir: raw data directory. |
|
|
num_input_channels: number of input (image) channels. |
|
|
output_datafolder: output folder. |
|
|
""" |
|
|
_index = 0 |
|
|
new_datalist_json: dict = {"training": [], test_key: []} |
|
|
|
|
|
for _key, _folder, _label_folder in list( |
|
|
zip(["training", test_key], ["imagesTr", "imagesTs"], ["labelsTr", "labelsTs"]) |
|
|
): |
|
|
if _key is None: |
|
|
continue |
|
|
|
|
|
logger.info(f"converting data section: {_key}...") |
|
|
for _k in tqdm(range(len(datalist_json[_key]))) if has_tqdm else range(len(datalist_json[_key])): |
|
|
orig_img_name = ( |
|
|
datalist_json[_key][_k]["image"] |
|
|
if isinstance(datalist_json[_key][_k], dict) |
|
|
else datalist_json[_key][_k] |
|
|
) |
|
|
img_name = f"case_{_index}" |
|
|
_index += 1 |
|
|
|
|
|
|
|
|
nda = monai.transforms.LoadImage(image_only=True, ensure_channel_first=True, simple_keys=True)( |
|
|
os.path.join(data_dir, orig_img_name) |
|
|
) |
|
|
affine = nda.meta["original_affine"] |
|
|
nda = nda.numpy() |
|
|
for _l in range(num_input_channels): |
|
|
outimg = nib.Nifti1Image(nda[_l, ...], affine) |
|
|
index = "_" + str(_l + 10000)[-4:] |
|
|
nib.save(outimg, os.path.join(output_datafolder, _folder, img_name + index + ".nii.gz")) |
|
|
|
|
|
|
|
|
if isinstance(datalist_json[_key][_k], dict) and "label" in datalist_json[_key][_k]: |
|
|
nda = monai.transforms.LoadImage(image_only=True, ensure_channel_first=True, simple_keys=True)( |
|
|
os.path.join(data_dir, datalist_json[_key][_k]["label"]) |
|
|
) |
|
|
affine = nda.meta["original_affine"] |
|
|
nda = nda.numpy().astype(np.uint8) |
|
|
nda = nda[0, ...] if nda.ndim == 4 and nda.shape[0] == 1 else nda |
|
|
nib.save( |
|
|
nib.Nifti1Image(nda, affine), os.path.join(output_datafolder, _label_folder, img_name + ".nii.gz") |
|
|
) |
|
|
|
|
|
if isinstance(datalist_json[_key][_k], dict): |
|
|
_val = copy.deepcopy(datalist_json[_key][_k]) |
|
|
_val["new_name"] = img_name |
|
|
new_datalist_json[_key].append(_val) |
|
|
else: |
|
|
new_datalist_json[_key].append({"image": datalist_json[_key][_k], "new_name": img_name}) |
|
|
|
|
|
ConfigParser.export_config_file( |
|
|
config=new_datalist_json, |
|
|
filepath=os.path.join(output_datafolder, "datalist.json"), |
|
|
fmt="json", |
|
|
sort_keys=True, |
|
|
indent=4, |
|
|
ensure_ascii=False, |
|
|
) |
|
|
|
|
|
return |
|
|
|
|
|
|
|
|
def create_new_dataset_json( |
|
|
modality: str, num_foreground_classes: int, num_input_channels: int, num_training_data: int, output_filepath: str |
|
|
) -> None: |
|
|
""" |
|
|
Create a new copy of dataset .json to meet the requirements of nnU-Net V2 |
|
|
|
|
|
Args: |
|
|
modality: image modality, could a string or a list of strings. |
|
|
num_foreground_classes: number of foreground classes. |
|
|
num_input_channels: number of input (image) channels. |
|
|
num_training_data: number of training data. |
|
|
output_filepath: output file path/name. |
|
|
""" |
|
|
new_json_data: dict = {} |
|
|
|
|
|
|
|
|
modality = ensure_tuple(modality) |
|
|
|
|
|
new_json_data["channel_names"] = {} |
|
|
for _j in range(num_input_channels): |
|
|
new_json_data["channel_names"][str(_j)] = modality[_j] |
|
|
|
|
|
new_json_data["labels"] = {} |
|
|
new_json_data["labels"]["background"] = 0 |
|
|
for _j in range(num_foreground_classes): |
|
|
new_json_data["labels"][f"class{_j + 1}"] = _j + 1 |
|
|
|
|
|
|
|
|
new_json_data["numTraining"] = num_training_data |
|
|
new_json_data["file_ending"] = ".nii.gz" |
|
|
|
|
|
ConfigParser.export_config_file( |
|
|
config=new_json_data, |
|
|
|
|
|
filepath=output_filepath, |
|
|
fmt="json", |
|
|
sort_keys=True, |
|
|
indent=4, |
|
|
ensure_ascii=False, |
|
|
) |
|
|
|
|
|
return |
|
|
|