|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
from __future__ import annotations |
|
|
|
|
|
import logging |
|
|
import os |
|
|
from collections.abc import Sequence |
|
|
|
|
|
import numpy as np |
|
|
|
|
|
from monai.config import PathLike |
|
|
from monai.transforms import Compose, EnsureChannelFirstd, LoadImaged, Orientationd, Spacingd, SqueezeDimd, Transform |
|
|
from monai.utils import GridSampleMode |
|
|
|
|
|
|
|
|
def create_dataset( |
|
|
datalist: list[dict], |
|
|
output_dir: str, |
|
|
dimension: int, |
|
|
pixdim: Sequence[float] | float, |
|
|
image_key: str = "image", |
|
|
label_key: str = "label", |
|
|
base_dir: PathLike | None = None, |
|
|
limit: int = 0, |
|
|
relative_path: bool = False, |
|
|
transforms: Transform | None = None, |
|
|
) -> list[dict]: |
|
|
""" |
|
|
Utility to pre-process and create dataset list for Deepgrow training over on existing one. |
|
|
The input data list is normally a list of images and labels (3D volume) that needs pre-processing |
|
|
for Deepgrow training pipeline. |
|
|
|
|
|
Args: |
|
|
datalist: A list of data dictionary. Each entry should at least contain 'image_key': <image filename>. |
|
|
For example, typical input data can be a list of dictionaries:: |
|
|
|
|
|
[{'image': <image filename>, 'label': <label filename>}] |
|
|
|
|
|
output_dir: target directory to store the training data for Deepgrow Training |
|
|
pixdim: output voxel spacing. |
|
|
dimension: dimension for Deepgrow training. It can be 2 or 3. |
|
|
image_key: image key in input datalist. Defaults to 'image'. |
|
|
label_key: label key in input datalist. Defaults to 'label'. |
|
|
base_dir: base directory in case related path is used for the keys in datalist. Defaults to None. |
|
|
limit: limit number of inputs for pre-processing. Defaults to 0 (no limit). |
|
|
relative_path: output keys values should be based on relative path. Defaults to False. |
|
|
transforms: explicit transforms to execute operations on input data. |
|
|
|
|
|
Raises: |
|
|
ValueError: When ``dimension`` is not one of [2, 3] |
|
|
ValueError: When ``datalist`` is Empty |
|
|
|
|
|
Returns: |
|
|
A new datalist that contains path to the images/labels after pre-processing. |
|
|
|
|
|
Example:: |
|
|
|
|
|
datalist = create_dataset( |
|
|
datalist=[{'image': 'img1.nii', 'label': 'label1.nii'}], |
|
|
base_dir=None, |
|
|
output_dir=output_2d, |
|
|
dimension=2, |
|
|
image_key='image', |
|
|
label_key='label', |
|
|
pixdim=(1.0, 1.0), |
|
|
limit=0, |
|
|
relative_path=True |
|
|
) |
|
|
|
|
|
print(datalist[0]["image"], datalist[0]["label"]) |
|
|
""" |
|
|
|
|
|
if dimension not in [2, 3]: |
|
|
raise ValueError("Dimension can be only 2 or 3 as Deepgrow supports only 2D/3D Training") |
|
|
|
|
|
if not len(datalist): |
|
|
raise ValueError("Input datalist is empty") |
|
|
|
|
|
transforms = _default_transforms(image_key, label_key, pixdim) if transforms is None else transforms |
|
|
new_datalist = [] |
|
|
for idx, item in enumerate(datalist): |
|
|
if limit and idx >= limit: |
|
|
break |
|
|
|
|
|
image = item[image_key] |
|
|
label = item.get(label_key, None) |
|
|
if base_dir: |
|
|
image = os.path.join(base_dir, image) |
|
|
label = os.path.join(base_dir, label) if label else None |
|
|
|
|
|
image = os.path.abspath(image) |
|
|
label = os.path.abspath(label) if label else None |
|
|
|
|
|
logging.info(f"Image: {image}; Label: {label if label else None}") |
|
|
data = transforms({image_key: image, label_key: label}) |
|
|
|
|
|
vol_image = data[image_key] |
|
|
vol_label = data.get(label_key) |
|
|
logging.info(f"Image (transform): {vol_image.shape}; Label: {None if vol_label is None else vol_label.shape}") |
|
|
|
|
|
vol_image = np.moveaxis(vol_image, -1, 0) |
|
|
if vol_label is not None: |
|
|
vol_label = np.moveaxis(vol_label, -1, 0) |
|
|
logging.info(f"Image (final): {vol_image.shape}; Label: {None if vol_label is None else vol_label.shape}") |
|
|
|
|
|
if dimension == 2: |
|
|
data = _save_data_2d( |
|
|
vol_idx=idx, |
|
|
vol_image=vol_image, |
|
|
vol_label=vol_label, |
|
|
dataset_dir=output_dir, |
|
|
relative_path=relative_path, |
|
|
) |
|
|
else: |
|
|
data = _save_data_3d( |
|
|
vol_idx=idx, |
|
|
vol_image=vol_image, |
|
|
vol_label=vol_label, |
|
|
dataset_dir=output_dir, |
|
|
relative_path=relative_path, |
|
|
) |
|
|
new_datalist.extend(data) |
|
|
return new_datalist |
|
|
|
|
|
|
|
|
def _default_transforms(image_key, label_key, pixdim): |
|
|
keys = [image_key] if label_key is None else [image_key, label_key] |
|
|
mode = [GridSampleMode.BILINEAR, GridSampleMode.NEAREST] if len(keys) == 2 else [GridSampleMode.BILINEAR] |
|
|
return Compose( |
|
|
[ |
|
|
LoadImaged(keys=keys), |
|
|
EnsureChannelFirstd(keys=keys), |
|
|
Orientationd(keys=keys, axcodes="RAS"), |
|
|
Spacingd(keys=keys, pixdim=pixdim, mode=mode), |
|
|
SqueezeDimd(keys=keys), |
|
|
] |
|
|
) |
|
|
|
|
|
|
|
|
def _save_data_2d(vol_idx, vol_image, vol_label, dataset_dir, relative_path): |
|
|
data_list: list[dict[str, str | int]] = [] |
|
|
|
|
|
image_count = 0 |
|
|
label_count = 0 |
|
|
unique_labels_count = 0 |
|
|
for sid in range(vol_image.shape[0]): |
|
|
image = vol_image[sid, ...] |
|
|
label = vol_label[sid, ...] if vol_label is not None else None |
|
|
|
|
|
if vol_label is not None and np.sum(label) == 0: |
|
|
continue |
|
|
|
|
|
image_file_prefix = f"vol_idx_{vol_idx:0>4d}_slice_{sid:0>3d}" |
|
|
image_file = os.path.join(dataset_dir, "images", image_file_prefix) |
|
|
image_file += ".npy" |
|
|
|
|
|
os.makedirs(os.path.join(dataset_dir, "images"), exist_ok=True) |
|
|
np.save(image_file, image) |
|
|
image_count += 1 |
|
|
|
|
|
|
|
|
if vol_label is None: |
|
|
data_list.append( |
|
|
{"image": image_file.replace(dataset_dir + os.pathsep, "") if relative_path else image_file} |
|
|
) |
|
|
continue |
|
|
|
|
|
|
|
|
unique_labels = np.unique(label.flatten()) |
|
|
unique_labels = unique_labels[unique_labels != 0] |
|
|
unique_labels_count = max(unique_labels_count, len(unique_labels)) |
|
|
|
|
|
for idx in unique_labels: |
|
|
label_file_prefix = f"{image_file_prefix}_region_{int(idx):0>2d}" |
|
|
label_file = os.path.join(dataset_dir, "labels", label_file_prefix) |
|
|
label_file += ".npy" |
|
|
|
|
|
os.makedirs(os.path.join(dataset_dir, "labels"), exist_ok=True) |
|
|
curr_label = (label == idx).astype(np.float32) |
|
|
np.save(label_file, curr_label) |
|
|
|
|
|
label_count += 1 |
|
|
data_list.append( |
|
|
{ |
|
|
"image": image_file.replace(dataset_dir + os.pathsep, "") if relative_path else image_file, |
|
|
"label": label_file.replace(dataset_dir + os.pathsep, "") if relative_path else label_file, |
|
|
"region": int(idx), |
|
|
} |
|
|
) |
|
|
|
|
|
if unique_labels_count >= 20: |
|
|
logging.warning(f"Unique labels {unique_labels_count} exceeds 20. Please check if this is correct.") |
|
|
|
|
|
logging.info( |
|
|
"{} => Image Shape: {} => {}; Label Shape: {} => {}; Unique Labels: {}".format( |
|
|
vol_idx, |
|
|
vol_image.shape, |
|
|
image_count, |
|
|
vol_label.shape if vol_label is not None else None, |
|
|
label_count, |
|
|
unique_labels_count, |
|
|
) |
|
|
) |
|
|
return data_list |
|
|
|
|
|
|
|
|
def _save_data_3d(vol_idx, vol_image, vol_label, dataset_dir, relative_path): |
|
|
data_list: list[dict[str, str | int]] = [] |
|
|
|
|
|
image_count = 0 |
|
|
label_count = 0 |
|
|
unique_labels_count = 0 |
|
|
|
|
|
image_file_prefix = f"vol_idx_{vol_idx:0>4d}" |
|
|
image_file = os.path.join(dataset_dir, "images", image_file_prefix) |
|
|
image_file += ".npy" |
|
|
|
|
|
os.makedirs(os.path.join(dataset_dir, "images"), exist_ok=True) |
|
|
np.save(image_file, vol_image) |
|
|
image_count += 1 |
|
|
|
|
|
|
|
|
if vol_label is None: |
|
|
data_list.append({"image": image_file.replace(dataset_dir + os.pathsep, "") if relative_path else image_file}) |
|
|
else: |
|
|
|
|
|
unique_labels = np.unique(vol_label.flatten()) |
|
|
unique_labels = unique_labels[unique_labels != 0] |
|
|
unique_labels_count = max(unique_labels_count, len(unique_labels)) |
|
|
|
|
|
for idx in unique_labels: |
|
|
label_file_prefix = f"{image_file_prefix}_region_{int(idx):0>2d}" |
|
|
label_file = os.path.join(dataset_dir, "labels", label_file_prefix) |
|
|
label_file += ".npy" |
|
|
|
|
|
curr_label = (vol_label == idx).astype(np.float32) |
|
|
os.makedirs(os.path.join(dataset_dir, "labels"), exist_ok=True) |
|
|
np.save(label_file, curr_label) |
|
|
|
|
|
label_count += 1 |
|
|
data_list.append( |
|
|
{ |
|
|
"image": image_file.replace(dataset_dir + os.pathsep, "") if relative_path else image_file, |
|
|
"label": label_file.replace(dataset_dir + os.pathsep, "") if relative_path else label_file, |
|
|
"region": int(idx), |
|
|
} |
|
|
) |
|
|
|
|
|
if unique_labels_count >= 20: |
|
|
logging.warning(f"Unique labels {unique_labels_count} exceeds 20. Please check if this is correct.") |
|
|
|
|
|
logging.info( |
|
|
"{} => Image Shape: {} => {}; Label Shape: {} => {}; Unique Labels: {}".format( |
|
|
vol_idx, |
|
|
vol_image.shape, |
|
|
image_count, |
|
|
vol_label.shape if vol_label is not None else None, |
|
|
label_count, |
|
|
unique_labels_count, |
|
|
) |
|
|
) |
|
|
return data_list |
|
|
|