File size: 6,761 Bytes
34a4bcb
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
# Copyright (c) MONAI Consortium
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#     http://www.apache.org/licenses/LICENSE-2.0
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

from __future__ import annotations

import copy
import os

import numpy as np

import monai
from monai.bundle import ConfigParser
from monai.utils import StrEnum, ensure_tuple, optional_import

tqdm, has_tqdm = optional_import("tqdm", name="tqdm")
nib, _ = optional_import("nibabel")

logger = monai.apps.utils.get_logger(__name__)

__all__ = ["analyze_data", "create_new_data_copy", "create_new_dataset_json", "NNUNETMode"]


class NNUNETMode(StrEnum):
    N_2D = "2d"
    N_3D_FULLRES = "3d_fullres"
    N_3D_LOWRES = "3d_lowres"
    N_3D_CASCADE_FULLRES = "3d_cascade_fullres"


def analyze_data(datalist_json: dict, data_dir: str) -> tuple[int, int]:
    """
    Analyze (training) data

    Args:
        datalist_json: original data list .json (required by most monai tutorials).
        data_dir: raw data directory.
    """
    img = monai.transforms.LoadImage(image_only=True, ensure_channel_first=True, simple_keys=True)(
        os.path.join(data_dir, datalist_json["training"][0]["image"])
    )
    num_input_channels = img.size()[0] if img.dim() == 4 else 1
    logger.info(f"num_input_channels: {num_input_channels}")

    num_foreground_classes = 0
    for _i in range(len(datalist_json["training"])):
        seg = monai.transforms.LoadImage(image_only=True, ensure_channel_first=True, simple_keys=True)(
            os.path.join(data_dir, datalist_json["training"][_i]["label"])
        )
        num_foreground_classes = max(num_foreground_classes, int(seg.max()))
    logger.info(f"num_foreground_classes: {num_foreground_classes}")

    return num_input_channels, num_foreground_classes


def create_new_data_copy(
    test_key: str, datalist_json: dict, data_dir: str, num_input_channels: int, output_datafolder: str
) -> None:
    """
    Create and organize a new copy of data to meet the requirements of nnU-Net V2

    Args:
        test_key: key for test data in the data list .json.
        datalist_json: original data list .json (required by most monai tutorials).
        data_dir: raw data directory.
        num_input_channels: number of input (image) channels.
        output_datafolder: output folder.
    """
    _index = 0
    new_datalist_json: dict = {"training": [], test_key: []}

    for _key, _folder, _label_folder in list(
        zip(["training", test_key], ["imagesTr", "imagesTs"], ["labelsTr", "labelsTs"])
    ):
        if _key is None:
            continue

        logger.info(f"converting data section: {_key}...")
        for _k in tqdm(range(len(datalist_json[_key]))) if has_tqdm else range(len(datalist_json[_key])):
            orig_img_name = (
                datalist_json[_key][_k]["image"]
                if isinstance(datalist_json[_key][_k], dict)
                else datalist_json[_key][_k]
            )
            img_name = f"case_{_index}"
            _index += 1

            # copy image
            nda = monai.transforms.LoadImage(image_only=True, ensure_channel_first=True, simple_keys=True)(
                os.path.join(data_dir, orig_img_name)
            )
            affine = nda.meta["original_affine"]
            nda = nda.numpy()
            for _l in range(num_input_channels):
                outimg = nib.Nifti1Image(nda[_l, ...], affine)
                index = "_" + str(_l + 10000)[-4:]
                nib.save(outimg, os.path.join(output_datafolder, _folder, img_name + index + ".nii.gz"))

            # copy label
            if isinstance(datalist_json[_key][_k], dict) and "label" in datalist_json[_key][_k]:
                nda = monai.transforms.LoadImage(image_only=True, ensure_channel_first=True, simple_keys=True)(
                    os.path.join(data_dir, datalist_json[_key][_k]["label"])
                )
                affine = nda.meta["original_affine"]
                nda = nda.numpy().astype(np.uint8)
                nda = nda[0, ...] if nda.ndim == 4 and nda.shape[0] == 1 else nda
                nib.save(
                    nib.Nifti1Image(nda, affine), os.path.join(output_datafolder, _label_folder, img_name + ".nii.gz")
                )

            if isinstance(datalist_json[_key][_k], dict):
                _val = copy.deepcopy(datalist_json[_key][_k])
                _val["new_name"] = img_name
                new_datalist_json[_key].append(_val)
            else:
                new_datalist_json[_key].append({"image": datalist_json[_key][_k], "new_name": img_name})

            ConfigParser.export_config_file(
                config=new_datalist_json,
                filepath=os.path.join(output_datafolder, "datalist.json"),
                fmt="json",
                sort_keys=True,
                indent=4,
                ensure_ascii=False,
            )

    return


def create_new_dataset_json(
    modality: str, num_foreground_classes: int, num_input_channels: int, num_training_data: int, output_filepath: str
) -> None:
    """
    Create a new copy of dataset .json to meet the requirements of nnU-Net V2

    Args:
        modality: image modality, could a string or a list of strings.
        num_foreground_classes: number of foreground classes.
        num_input_channels: number of input (image) channels.
        num_training_data: number of training data.
        output_filepath: output file path/name.
    """
    new_json_data: dict = {}

    # modality = self.input_info.pop("modality")
    modality = ensure_tuple(modality)  # type: ignore

    new_json_data["channel_names"] = {}
    for _j in range(num_input_channels):
        new_json_data["channel_names"][str(_j)] = modality[_j]

    new_json_data["labels"] = {}
    new_json_data["labels"]["background"] = 0
    for _j in range(num_foreground_classes):
        new_json_data["labels"][f"class{_j + 1}"] = _j + 1

    # new_json_data["numTraining"] = len(datalist_json["training"])
    new_json_data["numTraining"] = num_training_data
    new_json_data["file_ending"] = ".nii.gz"

    ConfigParser.export_config_file(
        config=new_json_data,
        # filepath=os.path.join(raw_data_foldername, "dataset.json"),
        filepath=output_filepath,
        fmt="json",
        sort_keys=True,
        indent=4,
        ensure_ascii=False,
    )

    return