|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
from __future__ import annotations |
|
|
|
|
|
import warnings |
|
|
from os import path |
|
|
from typing import Any, cast |
|
|
|
|
|
import numpy as np |
|
|
import torch |
|
|
from torch.multiprocessing import get_context |
|
|
|
|
|
from monai.apps.auto3dseg.transforms import EnsureSameShaped |
|
|
from monai.apps.utils import get_logger |
|
|
from monai.auto3dseg import SegSummarizer |
|
|
from monai.auto3dseg.utils import datafold_read |
|
|
from monai.bundle import config_parser |
|
|
from monai.bundle.config_parser import ConfigParser |
|
|
from monai.data import DataLoader, Dataset, partition_dataset |
|
|
from monai.data.utils import no_collation |
|
|
from monai.transforms import Compose, EnsureTyped, LoadImaged, Orientationd |
|
|
from monai.utils import ImageMetaKey, StrEnum, min_version, optional_import |
|
|
from monai.utils.enums import DataStatsKeys, ImageStatsKeys |
|
|
|
|
|
|
|
|
def strenum_representer(dumper, data): |
|
|
return dumper.represent_scalar("tag:yaml.org,2002:str", data.value) |
|
|
|
|
|
|
|
|
if optional_import("yaml")[1]: |
|
|
config_parser.yaml.SafeDumper.add_multi_representer(StrEnum, strenum_representer) |
|
|
|
|
|
tqdm, has_tqdm = optional_import("tqdm", "4.47.0", min_version, "tqdm") |
|
|
logger = get_logger(module_name=__name__) |
|
|
|
|
|
__all__ = ["DataAnalyzer"] |
|
|
|
|
|
|
|
|
class DataAnalyzer: |
|
|
""" |
|
|
The DataAnalyzer automatically analyzes given medical image dataset and reports the statistics. |
|
|
The module expects file paths to the image data and utilizes the LoadImaged transform to read the |
|
|
files, which supports nii, nii.gz, png, jpg, bmp, npz, npy, and dcm formats. Currently, only |
|
|
segmentation task is supported, so the user needs to provide paths to the image and label files |
|
|
(if have). Also, label data format is preferred to be (1,H,W,D), with the label index in the |
|
|
first dimension. If it is in onehot format, it will be converted to the preferred format. |
|
|
|
|
|
Args: |
|
|
datalist: a Python dictionary storing group, fold, and other information of the medical |
|
|
image dataset, or a string to the JSON file storing the dictionary. |
|
|
dataroot: user's local directory containing the datasets. |
|
|
output_path: path to save the analysis result. |
|
|
average: whether to average the statistical value across different image modalities. |
|
|
do_ccp: apply the connected component algorithm to process the labels/images |
|
|
device: a string specifying hardware (CUDA/CPU) utilized for the operations. |
|
|
worker: number of workers to use for loading datasets in each GPU/CPU sub-process. |
|
|
image_key: a string that user specify for the image. The DataAnalyzer will look it up in the |
|
|
datalist to locate the image files of the dataset. |
|
|
label_key: a string that user specify for the label. The DataAnalyzer will look it up in the |
|
|
datalist to locate the label files of the dataset. If label_key is NoneType or "None", |
|
|
the DataAnalyzer will skip looking for labels and all label-related operations. |
|
|
hist_bins: bins to compute histogram for each image channel. |
|
|
hist_range: ranges to compute histogram for each image channel. |
|
|
fmt: format used to save the analysis results. Currently support ``"json"`` and ``"yaml"``, defaults to "yaml". |
|
|
histogram_only: whether to only compute histograms. Defaults to False. |
|
|
extra_params: other optional arguments. Currently supported arguments are : |
|
|
'allowed_shape_difference' (default 5) can be used to change the default tolerance of |
|
|
the allowed shape differences between the image and label items. In case of shape mismatch below |
|
|
the tolerance, the label image will be resized to match the image using nearest interpolation. |
|
|
|
|
|
|
|
|
Examples: |
|
|
.. code-block:: python |
|
|
|
|
|
from monai.apps.auto3dseg.data_analyzer import DataAnalyzer |
|
|
|
|
|
datalist = { |
|
|
"testing": [{"image": "image_003.nii.gz"}], |
|
|
"training": [ |
|
|
{"fold": 0, "image": "image_001.nii.gz", "label": "label_001.nii.gz"}, |
|
|
{"fold": 0, "image": "image_002.nii.gz", "label": "label_002.nii.gz"}, |
|
|
{"fold": 1, "image": "image_001.nii.gz", "label": "label_001.nii.gz"}, |
|
|
{"fold": 1, "image": "image_004.nii.gz", "label": "label_004.nii.gz"}, |
|
|
], |
|
|
} |
|
|
|
|
|
dataroot = '/datasets' # the directory where you have the image files (nii.gz) |
|
|
DataAnalyzer(datalist, dataroot) |
|
|
|
|
|
Notes: |
|
|
The module can also be called from the command line interface (CLI). |
|
|
|
|
|
For example: |
|
|
|
|
|
.. code-block:: bash |
|
|
|
|
|
python -m monai.apps.auto3dseg \\ |
|
|
DataAnalyzer \\ |
|
|
get_all_case_stats \\ |
|
|
--datalist="my_datalist.json" \\ |
|
|
--dataroot="my_dataroot_dir" |
|
|
|
|
|
""" |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
datalist: str | dict, |
|
|
dataroot: str = "", |
|
|
output_path: str = "./datastats.yaml", |
|
|
average: bool = True, |
|
|
do_ccp: bool = False, |
|
|
device: str | torch.device = "cuda", |
|
|
worker: int = 4, |
|
|
image_key: str = "image", |
|
|
label_key: str | None = "label", |
|
|
hist_bins: list | int | None = 0, |
|
|
hist_range: list | None = None, |
|
|
fmt: str = "yaml", |
|
|
histogram_only: bool = False, |
|
|
**extra_params: Any, |
|
|
): |
|
|
if path.isfile(output_path): |
|
|
warnings.warn(f"File {output_path} already exists and will be overwritten.") |
|
|
logger.debug(f"{output_path} will be overwritten by a new datastat.") |
|
|
|
|
|
self.datalist = datalist |
|
|
self.dataroot = dataroot |
|
|
self.output_path = output_path |
|
|
self.average = average |
|
|
self.do_ccp = do_ccp |
|
|
self.device = torch.device(device) |
|
|
self.worker = worker |
|
|
self.image_key = image_key |
|
|
self.label_key = None if label_key == "None" else label_key |
|
|
self.hist_bins = hist_bins |
|
|
self.hist_range: list = [-500, 500] if hist_range is None else hist_range |
|
|
self.fmt = fmt |
|
|
self.histogram_only = histogram_only |
|
|
self.extra_params = extra_params |
|
|
|
|
|
@staticmethod |
|
|
def _check_data_uniformity(keys: list[str], result: dict) -> bool: |
|
|
""" |
|
|
Check data uniformity since DataAnalyzer provides no support to multi-modal images with different |
|
|
affine matrices/spacings due to monai transforms. |
|
|
|
|
|
Args: |
|
|
keys: a list of string-type keys under image_stats dictionary. |
|
|
|
|
|
Returns: |
|
|
False if one of the selected key values is not constant across the dataset images. |
|
|
|
|
|
""" |
|
|
|
|
|
if DataStatsKeys.SUMMARY not in result or DataStatsKeys.IMAGE_STATS not in result[DataStatsKeys.SUMMARY]: |
|
|
return True |
|
|
constant_props = [result[DataStatsKeys.SUMMARY][DataStatsKeys.IMAGE_STATS][key] for key in keys] |
|
|
for prop in constant_props: |
|
|
if "stdev" in prop and np.any(prop["stdev"]): |
|
|
logger.debug(f"summary image_stats {prop} has non-zero stdev {prop['stdev']}.") |
|
|
return False |
|
|
|
|
|
return True |
|
|
|
|
|
def get_all_case_stats(self, key="training", transform_list=None): |
|
|
""" |
|
|
Get all case stats. Caller of the DataAnalyser class. The function initiates multiple GPU or CPU processes of the internal |
|
|
_get_all_case_stats functions, which iterates datalist and call SegSummarizer to generate stats for each case. |
|
|
After all case stats are generated, SegSummarizer is called to combine results. |
|
|
|
|
|
Args: |
|
|
key: dataset key |
|
|
transform_list: option list of transforms before SegSummarizer |
|
|
|
|
|
Returns: |
|
|
A data statistics dictionary containing |
|
|
"stats_summary" (summary statistics of the entire datasets). Within stats_summary |
|
|
there are "image_stats" (summarizing info of shape, channel, spacing, and etc |
|
|
using operations_summary), "image_foreground_stats" (info of the intensity for the |
|
|
non-zero labeled voxels), and "label_stats" (info of the labels, pixel percentage, |
|
|
image_intensity, and each individual label in a list) |
|
|
"stats_by_cases" (List type value. Each element of the list is statistics of |
|
|
an image-label info. Within each element, there are: "image" (value is the |
|
|
path to an image), "label" (value is the path to the corresponding label), "image_stats" |
|
|
(summarizing info of shape, channel, spacing, and etc using operations), |
|
|
"image_foreground_stats" (similar to the previous one but one foreground image), and |
|
|
"label_stats" (stats of the individual labels ) |
|
|
|
|
|
Notes: |
|
|
Since the backend of the statistics computation are torch/numpy, nan/inf value |
|
|
may be generated and carried over in the computation. In such cases, the output |
|
|
dictionary will include .nan/.inf in the statistics. |
|
|
|
|
|
""" |
|
|
result: dict[DataStatsKeys, Any] = {DataStatsKeys.SUMMARY: {}, DataStatsKeys.BY_CASE: []} |
|
|
result_bycase: dict[DataStatsKeys, Any] = {DataStatsKeys.SUMMARY: {}, DataStatsKeys.BY_CASE: []} |
|
|
if self.device.type == "cpu": |
|
|
nprocs = 1 |
|
|
logger.info("Using CPU for data analyzing!") |
|
|
else: |
|
|
nprocs = torch.cuda.device_count() |
|
|
logger.info(f"Found {nprocs} GPUs for data analyzing!") |
|
|
if nprocs > 1: |
|
|
tmp_ctx: Any = get_context("forkserver") |
|
|
with tmp_ctx.Manager() as manager: |
|
|
manager_list = manager.list() |
|
|
processes = [] |
|
|
for rank in range(nprocs): |
|
|
p = tmp_ctx.Process( |
|
|
target=self._get_all_case_stats, args=(rank, nprocs, manager_list, key, transform_list) |
|
|
) |
|
|
processes.append(p) |
|
|
for p in processes: |
|
|
p.start() |
|
|
for p in processes: |
|
|
p.join() |
|
|
|
|
|
for _ in manager_list: |
|
|
result_bycase[DataStatsKeys.BY_CASE].extend(_[DataStatsKeys.BY_CASE]) |
|
|
else: |
|
|
result_bycase = self._get_all_case_stats(0, 1, None, key, transform_list) |
|
|
|
|
|
summarizer = SegSummarizer( |
|
|
self.image_key, |
|
|
self.label_key, |
|
|
average=self.average, |
|
|
do_ccp=self.do_ccp, |
|
|
hist_bins=self.hist_bins, |
|
|
hist_range=self.hist_range, |
|
|
histogram_only=self.histogram_only, |
|
|
) |
|
|
n_cases = len(result_bycase[DataStatsKeys.BY_CASE]) |
|
|
result[DataStatsKeys.SUMMARY] = summarizer.summarize(cast(list, result_bycase[DataStatsKeys.BY_CASE])) |
|
|
result[DataStatsKeys.SUMMARY]["n_cases"] = n_cases |
|
|
result_bycase[DataStatsKeys.SUMMARY] = result[DataStatsKeys.SUMMARY] |
|
|
if not self._check_data_uniformity([ImageStatsKeys.SPACING], result): |
|
|
logger.info("Data spacing is not completely uniform. MONAI transforms may provide unexpected result") |
|
|
if self.output_path: |
|
|
logger.info(f"Writing data stats to {self.output_path}.") |
|
|
ConfigParser.export_config_file( |
|
|
result, self.output_path, fmt=self.fmt, default_flow_style=None, sort_keys=False |
|
|
) |
|
|
by_case_path = self.output_path.replace(f".{self.fmt}", f"_by_case.{self.fmt}") |
|
|
if by_case_path == self.output_path: |
|
|
by_case_path += f".by_case.{self.fmt}" |
|
|
logger.info(f"Writing by-case data stats to {by_case_path}, this may take a while.") |
|
|
ConfigParser.export_config_file( |
|
|
result_bycase, by_case_path, fmt=self.fmt, default_flow_style=None, sort_keys=False |
|
|
) |
|
|
|
|
|
if self.device.type == "cuda": |
|
|
|
|
|
|
|
|
torch.cuda.empty_cache() |
|
|
result[DataStatsKeys.BY_CASE] = result_bycase[DataStatsKeys.BY_CASE] |
|
|
return result |
|
|
|
|
|
def _get_all_case_stats( |
|
|
self, |
|
|
rank: int = 0, |
|
|
world_size: int = 1, |
|
|
manager_list: list | None = None, |
|
|
key: str = "training", |
|
|
transform_list: list | None = None, |
|
|
) -> Any: |
|
|
""" |
|
|
Get all case stats from a partitioned datalist. The function can only be called internally by get_all_case_stats. |
|
|
Args: |
|
|
rank: GPU process rank, 0 for CPU process |
|
|
world_size: total number of GPUs, 1 for CPU process |
|
|
manager_list: multiprocessing manager list object, if using multi-GPU. |
|
|
key: dataset key |
|
|
transform_list: option list of transforms before SegSummarizer |
|
|
""" |
|
|
summarizer = SegSummarizer( |
|
|
self.image_key, |
|
|
self.label_key, |
|
|
average=self.average, |
|
|
do_ccp=self.do_ccp, |
|
|
hist_bins=self.hist_bins, |
|
|
hist_range=self.hist_range, |
|
|
histogram_only=self.histogram_only, |
|
|
) |
|
|
keys = list(filter(None, [self.image_key, self.label_key])) |
|
|
if transform_list is None: |
|
|
transform_list = [ |
|
|
LoadImaged(keys=keys, ensure_channel_first=True, image_only=True), |
|
|
EnsureTyped(keys=keys, data_type="tensor", dtype=torch.float), |
|
|
Orientationd(keys=keys, axcodes="RAS"), |
|
|
] |
|
|
if self.label_key is not None: |
|
|
allowed_shape_difference = self.extra_params.pop("allowed_shape_difference", 5) |
|
|
transform_list.append( |
|
|
EnsureSameShaped( |
|
|
keys=self.label_key, |
|
|
source_key=self.image_key, |
|
|
allowed_shape_difference=allowed_shape_difference, |
|
|
) |
|
|
) |
|
|
|
|
|
transform = Compose(transform_list) |
|
|
files, _ = datafold_read(datalist=self.datalist, basedir=self.dataroot, fold=-1, key=key) |
|
|
if world_size <= len(files): |
|
|
files = partition_dataset(data=files, num_partitions=world_size)[rank] |
|
|
else: |
|
|
files = partition_dataset(data=files, num_partitions=len(files))[rank] if rank < len(files) else [] |
|
|
dataset = Dataset(data=files, transform=transform) |
|
|
dataloader = DataLoader( |
|
|
dataset, |
|
|
batch_size=1, |
|
|
shuffle=False, |
|
|
num_workers=self.worker, |
|
|
collate_fn=no_collation, |
|
|
pin_memory=self.device.type == "cuda", |
|
|
) |
|
|
result_bycase: dict[DataStatsKeys, Any] = {DataStatsKeys.SUMMARY: {}, DataStatsKeys.BY_CASE: []} |
|
|
device = self.device if self.device.type == "cpu" else torch.device("cuda", rank) |
|
|
if device.type == "cuda" and not (torch.cuda.is_available() and torch.cuda.device_count() > 0): |
|
|
logger.info(f"device={device} but CUDA device is not available, using CPU instead.") |
|
|
device = torch.device("cpu") |
|
|
if not has_tqdm: |
|
|
warnings.warn("tqdm is not installed. not displaying the caching progress.") |
|
|
|
|
|
for batch_data in tqdm(dataloader) if (has_tqdm and rank == 0) else dataloader: |
|
|
batch_data = batch_data[0] |
|
|
try: |
|
|
batch_data[self.image_key] = batch_data[self.image_key].to(device) |
|
|
_label_argmax = False |
|
|
if self.label_key is not None: |
|
|
label = batch_data[self.label_key] |
|
|
label = torch.argmax(label, dim=0) if label.shape[0] > 1 else label[0] |
|
|
_label_argmax = True |
|
|
batch_data[self.label_key] = label.to(device) |
|
|
d = summarizer(batch_data) |
|
|
except BaseException as err: |
|
|
if "image_meta_dict" in batch_data.keys(): |
|
|
filename = batch_data["image_meta_dict"][ImageMetaKey.FILENAME_OR_OBJ] |
|
|
else: |
|
|
filename = batch_data[self.image_key].meta[ImageMetaKey.FILENAME_OR_OBJ] |
|
|
logger.info(f"Unable to process data {filename} on {device}. {err}") |
|
|
if self.device.type == "cuda": |
|
|
logger.info("DataAnalyzer `device` set to GPU execution hit an exception. Falling back to `cpu`.") |
|
|
try: |
|
|
batch_data[self.image_key] = batch_data[self.image_key].to("cpu") |
|
|
if self.label_key is not None: |
|
|
label = batch_data[self.label_key] |
|
|
if not _label_argmax: |
|
|
label = torch.argmax(label, dim=0) if label.shape[0] > 1 else label[0] |
|
|
batch_data[self.label_key] = label.to("cpu") |
|
|
d = summarizer(batch_data) |
|
|
except BaseException as err: |
|
|
logger.info(f"Unable to process data {filename} on {device}. {err}") |
|
|
continue |
|
|
else: |
|
|
continue |
|
|
|
|
|
stats_by_cases = { |
|
|
DataStatsKeys.BY_CASE_IMAGE_PATH: d[DataStatsKeys.BY_CASE_IMAGE_PATH], |
|
|
DataStatsKeys.BY_CASE_LABEL_PATH: d[DataStatsKeys.BY_CASE_LABEL_PATH], |
|
|
} |
|
|
if not self.histogram_only: |
|
|
stats_by_cases[DataStatsKeys.IMAGE_STATS] = d[DataStatsKeys.IMAGE_STATS] |
|
|
if self.hist_bins != 0: |
|
|
stats_by_cases[DataStatsKeys.IMAGE_HISTOGRAM] = d[DataStatsKeys.IMAGE_HISTOGRAM] |
|
|
|
|
|
if self.label_key is not None: |
|
|
stats_by_cases.update( |
|
|
{ |
|
|
DataStatsKeys.FG_IMAGE_STATS: d[DataStatsKeys.FG_IMAGE_STATS], |
|
|
DataStatsKeys.LABEL_STATS: d[DataStatsKeys.LABEL_STATS], |
|
|
} |
|
|
) |
|
|
result_bycase[DataStatsKeys.BY_CASE].append(stats_by_cases) |
|
|
if manager_list is None: |
|
|
return result_bycase |
|
|
else: |
|
|
manager_list.append(result_bycase) |
|
|
|