| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | import hashlib |
| | import json |
| | import math |
| | import sys |
| | import threading |
| | import time |
| | from multiprocessing.pool import ThreadPool |
| | from pathlib import Path |
| | from typing import Any, Callable, List, Optional, Sequence, Tuple, Union |
| |
|
| | import numpy as np |
| | import torch |
| | from torch.utils.data import Dataset as _TorchDataset |
| |
|
| | from monai.transforms import Compose, Randomizable, Transform, apply_transform |
| | from monai.utils import get_seed, progress_bar |
| |
|
| |
|
| | class Dataset(_TorchDataset): |
| | """ |
| | A generic dataset with a length property and an optional callable data transform |
| | when fetching a data sample. |
| | For example, typical input data can be a list of dictionaries:: |
| | |
| | [{ { { |
| | 'img': 'image1.nii.gz', 'img': 'image2.nii.gz', 'img': 'image3.nii.gz', |
| | 'seg': 'label1.nii.gz', 'seg': 'label2.nii.gz', 'seg': 'label3.nii.gz', |
| | 'extra': 123 'extra': 456 'extra': 789 |
| | }, }, }] |
| | """ |
| |
|
| | def __init__(self, data: Sequence, transform: Optional[Callable] = None) -> None: |
| | """ |
| | Args: |
| | data: input data to load and transform to generate dataset for model. |
| | transform: a callable data transform on input data. |
| | """ |
| | self.data = data |
| | self.transform = transform |
| |
|
| | def __len__(self) -> int: |
| | return len(self.data) |
| |
|
| | def __getitem__(self, index: int): |
| | data = self.data[index] |
| | if self.transform is not None: |
| | data = apply_transform(self.transform, data) |
| |
|
| | return data |
| |
|
| |
|
| | class PersistentDataset(Dataset): |
| | """ |
| | Persistent storage of pre-computed values to efficiently manage larger than memory dictionary format data, |
| | it can operate transforms for specific fields. Results from the non-random transform components are computed |
| | when first used, and stored in the `cache_dir` for rapid retrieval on subsequent uses. |
| | |
| | For example, typical input data can be a list of dictionaries:: |
| | |
| | [{ { { |
| | 'img': 'image1.nii.gz', 'img': 'image2.nii.gz', 'img': 'image3.nii.gz', |
| | 'seg': 'label1.nii.gz', 'seg': 'label2.nii.gz', 'seg': 'label3.nii.gz', |
| | 'extra': 123 'extra': 456 'extra': 789 |
| | }, }, }] |
| | |
| | For a composite transform like |
| | |
| | .. code-block:: python |
| | |
| | [ LoadNiftid(keys=['image', 'label']), |
| | Orientationd(keys=['image', 'label'], axcodes='RAS'), |
| | ScaleIntensityRanged(keys=['image'], a_min=-57, a_max=164, b_min=0.0, b_max=1.0, clip=True), |
| | RandCropByPosNegLabeld(keys=['image', 'label'], label_key='label', spatial_size=(96, 96, 96), |
| | pos=1, neg=1, num_samples=4, image_key='image', image_threshold=0), |
| | ToTensord(keys=['image', 'label'])] |
| | |
| | Upon first use a filename based dataset will be processed by the transform for the |
| | [LoadNiftid, Orientationd, ScaleIntensityRanged] and the resulting tensor written to |
| | the `cache_dir` before applying the remaining random dependant transforms |
| | [RandCropByPosNegLabeld, ToTensord] elements for use in the analysis. |
| | |
| | Subsequent uses of a dataset directly read pre-processed results from `cache_dir` |
| | followed by applying the random dependant parts of transform processing. |
| | |
| | Note: |
| | The input data must be a list of file paths and will hash them as cache keys. |
| | |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | data: Sequence[str], |
| | transform: Union[Sequence[Callable], Callable], |
| | cache_dir: Optional[Union[Path, str]] = None, |
| | ) -> None: |
| | """ |
| | Args: |
| | data: input data file paths to load and transform to generate dataset for model. |
| | `PersistentDataset` expects input data to be a list of file paths and hashes them as cache keys. |
| | transform: transforms to execute operations on input data. |
| | cache_dir: If specified, this is the location for persistent storage |
| | of pre-computed transformed data tensors. The cache_dir is computed once, and |
| | persists on disk until explicitly removed. Different runs, programs, experiments |
| | may share a common cache dir provided that the transforms pre-processing is consistent. |
| | If the cache_dir doesn't exist, will automatically create it. |
| | """ |
| | if not isinstance(transform, Compose): |
| | transform = Compose(transform) |
| | super().__init__(data=data, transform=transform) |
| | self.cache_dir = Path(cache_dir) if cache_dir is not None else None |
| | if self.cache_dir is not None: |
| | if not self.cache_dir.exists(): |
| | self.cache_dir.mkdir(parents=True) |
| | if not self.cache_dir.is_dir(): |
| | raise ValueError("cache_dir must be a directory.") |
| |
|
| | def _pre_first_random_transform(self, item_transformed): |
| | """ |
| | Process the data from original state up to the first random element. |
| | |
| | Args: |
| | item_transformed: The data to be transformed |
| | |
| | Returns: |
| | the transformed element up to the first identified |
| | random transform object |
| | """ |
| | for _transform in self.transform.transforms: |
| | |
| | if isinstance(_transform, Randomizable) or not isinstance(_transform, Transform): |
| | break |
| | item_transformed = apply_transform(_transform, item_transformed) |
| | return item_transformed |
| |
|
| | def _first_random_and_beyond_transform(self, item_transformed): |
| | """ |
| | Process the data from before the first random transform to the final state ready for evaluation. |
| | |
| | Args: |
| | item_transformed: The data to be transformed (already processed up to the first random transform) |
| | |
| | Returns: |
| | the transformed element through the random transforms |
| | """ |
| | start_post_randomize_run = False |
| | for _transform in self.transform.transforms: |
| | if ( |
| | start_post_randomize_run |
| | or isinstance(_transform, Randomizable) |
| | or not isinstance(_transform, Transform) |
| | ): |
| | start_post_randomize_run = True |
| | item_transformed = apply_transform(_transform, item_transformed) |
| | return item_transformed |
| |
|
| | def _pre_first_random_cachecheck(self, item_transformed): |
| | """ |
| | A function to cache the expensive input data transform operations |
| | so that huge data sets (larger than computer memory) can be processed |
| | on the fly as needed, and intermediate results written to disk for |
| | future use. |
| | |
| | Args: |
| | item_transformed: The current data element to be mutated into transformed representation |
| | |
| | Returns: |
| | The transformed data_element, either from cache, or explicitly computing it. |
| | |
| | Warning: |
| | The current implementation does not encode transform information as part of the |
| | hashing mechanism used for generating cache names. If the transforms applied are |
| | changed in any way, the objects in the cache dir will be invalid. The hash for the |
| | cache is ONLY dependant on the input filename paths. |
| | """ |
| | if item_transformed.get("cached", False) is False: |
| | hashfile = None |
| | if self.cache_dir is not None: |
| | |
| | data_item_md5 = hashlib.md5(json.dumps(item_transformed, sort_keys=True).encode("utf-8")).hexdigest() |
| | hashfile = self.cache_dir / f"{data_item_md5}.pt" |
| |
|
| | if hashfile is not None and hashfile.is_file(): |
| | item_transformed = torch.load(hashfile) |
| | else: |
| | item_transformed = self._pre_first_random_transform(item_transformed) |
| | if hashfile is not None: |
| | |
| | item_transformed["cached"] = True |
| | |
| | |
| | |
| | temp_hash_file = hashfile.with_suffix(".temp_write_cache") |
| | torch.save(item_transformed, temp_hash_file) |
| | temp_hash_file.rename(hashfile) |
| |
|
| | return item_transformed |
| |
|
| | def __getitem__(self, index: int): |
| | pre_random_item = self._pre_first_random_cachecheck(self.data[index]) |
| | post_random_item = self._first_random_and_beyond_transform(pre_random_item) |
| | return post_random_item |
| |
|
| |
|
| | class CacheDataset(Dataset): |
| | """ |
| | Dataset with cache mechanism that can load data and cache deterministic transforms' result during training. |
| | |
| | By caching the results of non-random preprocessing transforms, it accelerates the training data pipeline. |
| | If the requested data is not in the cache, all transforms will run normally |
| | (see also :py:class:`monai.data.dataset.Dataset`). |
| | |
| | Users can set the cache rate or number of items to cache. |
| | It is recommended to experiment with different `cache_num` or `cache_rate` to identify the best training speed. |
| | |
| | To improve the caching efficiency, please always put as many as possible non-random transforms |
| | before the randomized ones when composing the chain of transforms. |
| | |
| | For example, if the transform is a `Compose` of:: |
| | |
| | transforms = Compose([ |
| | LoadNiftid(), |
| | AddChanneld(), |
| | Spacingd(), |
| | Orientationd(), |
| | ScaleIntensityRanged(), |
| | RandCropByPosNegLabeld(), |
| | ToTensord() |
| | ]) |
| | |
| | when `transforms` is used in a multi-epoch training pipeline, before the first training epoch, |
| | this dataset will cache the results up to ``ScaleIntensityRanged``, as |
| | all non-random transforms `LoadNiftid`, `AddChanneld`, `Spacingd`, `Orientationd`, `ScaleIntensityRanged` |
| | can be cached. During training, the dataset will load the cached results and run |
| | ``RandCropByPosNegLabeld`` and ``ToTensord``, as ``RandCropByPosNegLabeld`` is a randomized transform |
| | and the outcome not cached. |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | data: Sequence, |
| | transform: Union[Sequence[Callable], Callable], |
| | cache_num: int = sys.maxsize, |
| | cache_rate: float = 1.0, |
| | num_workers: int = 0, |
| | ) -> None: |
| | """ |
| | Args: |
| | data: input data to load and transform to generate dataset for model. |
| | transform: transforms to execute operations on input data. |
| | cache_num: number of items to be cached. Default is `sys.maxsize`. |
| | will take the minimum of (cache_num, data_length x cache_rate, data_length). |
| | cache_rate: percentage of cached data in total, default is 1.0 (cache all). |
| | will take the minimum of (cache_num, data_length x cache_rate, data_length). |
| | num_workers: the number of worker threads to use. |
| | If 0 a single thread will be used. Default is 0. |
| | """ |
| | if not isinstance(transform, Compose): |
| | transform = Compose(transform) |
| | super().__init__(data, transform) |
| | self.cache_num = min(cache_num, int(len(data) * cache_rate), len(data)) |
| | if self.cache_num > 0: |
| | self._cache = [None] * self.cache_num |
| | if num_workers > 0: |
| | self._item_processed = 0 |
| | self._thread_lock = threading.Lock() |
| | with ThreadPool(num_workers) as p: |
| | p.map( |
| | self._load_cache_item_thread, |
| | [(i, data[i], transform.transforms) for i in range(self.cache_num)], |
| | ) |
| | else: |
| | for i in range(self.cache_num): |
| | self._cache[i] = self._load_cache_item(data[i], transform.transforms) |
| | progress_bar(i + 1, self.cache_num, "Load and cache transformed data: ") |
| |
|
| | def _load_cache_item(self, item: Any, transforms: Sequence[Callable]): |
| | """ |
| | Args: |
| | item: input item to load and transform to generate dataset for model. |
| | transforms: transforms to execute operations on input item. |
| | """ |
| | for _transform in transforms: |
| | |
| | if isinstance(_transform, Randomizable) or not isinstance(_transform, Transform): |
| | break |
| | item = apply_transform(_transform, item) |
| | return item |
| |
|
| | def _load_cache_item_thread(self, args: Tuple[int, Any, Sequence[Callable]]) -> None: |
| | """ |
| | Args: |
| | args: tuple with contents (i, item, transforms). |
| | i: the index to load the cached item to. |
| | item: input item to load and transform to generate dataset for model. |
| | transforms: transforms to execute operations on input item. |
| | """ |
| | i, item, transforms = args |
| | self._cache[i] = self._load_cache_item(item, transforms) |
| | with self._thread_lock: |
| | self._item_processed += 1 |
| | progress_bar(self._item_processed, self.cache_num, "Load and cache transformed data: ") |
| |
|
| | def __getitem__(self, index): |
| | if index < self.cache_num: |
| | |
| | start_run = False |
| | data = self._cache[index] |
| | for _transform in self.transform.transforms: |
| | if not start_run and not isinstance(_transform, Randomizable) and isinstance(_transform, Transform): |
| | continue |
| | else: |
| | start_run = True |
| | data = apply_transform(_transform, data) |
| | else: |
| | |
| | data = super(CacheDataset, self).__getitem__(index) |
| | return data |
| |
|
| |
|
| | class SmartCacheDataset(CacheDataset): |
| | """ |
| | Re-implementation of the SmartCache mechanism in NVIDIA Clara-train SDK. |
| | At any time, the cache pool only keeps a subset of the whole dataset. In each epoch, only the items |
| | in the cache are used for training. This ensures that data needed for training is readily available, |
| | keeping GPU resources busy. Note that cached items may still have to go through a non-deterministic |
| | transform sequence before being fed to GPU. At the same time, another thread is preparing replacement |
| | items by applying the transform sequence to items not in cache. Once one epoch is completed, Smart |
| | Cache replaces the same number of items with replacement items. |
| | Smart Cache uses a simple `running window` algorithm to determine the cache content and replacement items. |
| | Let N be the configured number of objects in cache; and R be the number of replacement objects (R = ceil(N * r), |
| | where r is the configured replace rate). |
| | For more details, please refer to: |
| | https://docs.nvidia.com/clara/tlt-mi/clara-train-sdk-v3.0/nvmidl/additional_features/smart_cache.html#smart-cache |
| | |
| | For example, if we have 5 images: `[image1, image2, image3, image4, image5]`, and `cache_num=4`, `replace_rate=0.25`. |
| | so the actual training images cached and replaced for every epoch are as below: |
| | epoch 1: [image1, image2, image3, image4] |
| | epoch 2: [image2, image3, image4, image5] |
| | epoch 3: [image3, image4, image5, image1] |
| | epoch 3: [image4, image5, image1, image2] |
| | epoch N: [image[N % 5] ...] |
| | |
| | The usage of `SmartCacheDataset` contains 4 steps: |
| | 1. Initialize `SmartCacheDataset` object and cache for the first epoch. |
| | 2. Call `start()` to run replacement thread in background. |
| | 3. Call `update_cache()` before every epoch to replace training items. |
| | 4. Call `shutdown()` when training ends. |
| | |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | data: Sequence, |
| | transform: Union[Sequence[Callable], Callable], |
| | replace_rate: float, |
| | cache_num: int = sys.maxsize, |
| | cache_rate: float = 1.0, |
| | num_init_workers: int = 0, |
| | num_replace_workers: int = 0, |
| | ) -> None: |
| | """ |
| | Args: |
| | data: input data to load and transform to generate dataset for model. |
| | transform: transforms to execute operations on input data. |
| | replace_rate: percentage of the cached items to be replaced in every epoch. |
| | cache_num: number of items to be cached. Default is `sys.maxsize`. |
| | will take the minimum of (cache_num, data_length x cache_rate, data_length). |
| | cache_rate: percentage of cached data in total, default is 1.0 (cache all). |
| | will take the minimum of (cache_num, data_length x cache_rate, data_length). |
| | num_init_workers: the number of worker threads to initialize the cache for first epoch. |
| | if 0, run in main thread, no separate thread will open. |
| | num_replace_workers: the number of worker threads to prepare the replacement cache for every epoch. |
| | if 0, run in main thread, no separate thread will open. |
| | """ |
| | super().__init__(data, transform, cache_num, cache_rate, num_init_workers) |
| | if self.cache_num >= len(data): |
| | raise ValueError("cache_num must be smaller than dataset length to support replacement.") |
| | if replace_rate <= 0: |
| | raise ValueError("replace_rate must be greater than 0, otherwise, please use CacheDataset.") |
| | self.num_replace_workers: int = num_replace_workers |
| |
|
| | self._total_num: int = len(data) |
| | self._replace_num: int = min(math.ceil(self.cache_num * replace_rate), len(data) - self.cache_num) |
| | self._replacements: List[Any] = [None for _ in range(self._replace_num)] |
| | self._replace_data_idx: List[int] = list(range(self._replace_num)) |
| |
|
| | self._start_pos: int = 0 |
| | self._update_lock: threading.Lock = threading.Lock() |
| | self._round: int = 1 |
| | self._replace_done: bool = False |
| | self._replace_mgr: Optional[threading.Thread] = None |
| |
|
| | self._compute_data_idx() |
| |
|
| | def _compute_data_idx(self): |
| | """ |
| | Update the replacement data position in the total data. |
| | |
| | """ |
| | for i in range(self._replace_num): |
| | pos: int = self._start_pos + self.cache_num + i |
| | if pos >= self._total_num: |
| | pos -= self._total_num |
| | self._replace_data_idx[i] = pos |
| |
|
| | def is_started(self): |
| | """ |
| | Check whether the replacement thread is already started. |
| | |
| | """ |
| | if self._replace_mgr is None: |
| | return False |
| | return self._replace_mgr.isAlive() |
| |
|
| | def start(self): |
| | """ |
| | Start the background thread to replace training items for every epoch. |
| | |
| | """ |
| | if self._replace_mgr is None or not self.is_started(): |
| | self._restart() |
| |
|
| | def _restart(self): |
| | """ |
| | Restart background thread if killed for some reason. |
| | |
| | """ |
| | self._round = 1 |
| | self._replace_mgr = threading.Thread(target=self.manage_replacement, daemon=True) |
| | self._replace_mgr.start() |
| |
|
| | def _try_update_cache(self): |
| | """ |
| | Update the cache items with new replacement for current epoch. |
| | |
| | """ |
| | with self._update_lock: |
| | if self._replace_done: |
| | remain_num: int = self.cache_num - self._replace_num |
| | for i in range(remain_num): |
| | self._cache[i] = self._cache[i + self._replace_num] |
| | for i in range(self._replace_num): |
| | self._cache[remain_num + i] = self._replacements[i] |
| |
|
| | self._start_pos += self._replace_num |
| | if self._start_pos >= self._total_num: |
| | self._start_pos -= self._total_num |
| |
|
| | self._compute_data_idx() |
| |
|
| | |
| | self._round += 1 |
| | self._replace_done = False |
| | return True |
| | else: |
| | return False |
| |
|
| | def update_cache(self): |
| | """ |
| | Update cache items for current epoch, need to call this function before every epoch. |
| | If the cache has been shutdown before, need to restart the `_replace_mgr` thread. |
| | |
| | """ |
| | if not self._replace_mgr.isAlive(): |
| | self._restart() |
| |
|
| | |
| | while not self._try_update_cache(): |
| | time.sleep(0.01) |
| |
|
| | def _try_shutdown(self): |
| | """ |
| | Wait for thread lock to shut down the background thread. |
| | |
| | """ |
| | with self._update_lock: |
| | if self._replace_done: |
| | self._round = 0 |
| | self._replace_done = False |
| | return True |
| | else: |
| | return False |
| |
|
| | def shutdown(self): |
| | """ |
| | Shut down the background thread for replacement. |
| | |
| | """ |
| | if not self.is_started(): |
| | return |
| |
|
| | |
| | while not self._try_shutdown(): |
| | time.sleep(0.01) |
| | self._replace_mgr.join() |
| |
|
| | def _replace_cache_thread(self, index: int): |
| | """ |
| | Execute deterministic transforms on the new data for replacement. |
| | |
| | """ |
| | pos: int = self._replace_data_idx[index] |
| | self._replacements[index] = self._load_cache_item(self.data[pos], self.transform.transforms) |
| |
|
| | def _compute_replacements(self): |
| | """ |
| | Compute expected items for the replacement of next epoch, execute deterministic transforms. |
| | It can support multi-threads to accelerate the computation progress. |
| | |
| | """ |
| | if self.num_replace_workers > 0: |
| | with ThreadPool(self.num_replace_workers) as p: |
| | p.map(self._replace_cache_thread, list(range(self._replace_num))) |
| | else: |
| | for i in range(self._replace_num): |
| | self._replace_cache_thread(i) |
| | self._replace_done = True |
| |
|
| | def _try_manage_replacement(self, check_round): |
| | """ |
| | Wait thread lock and replace training items in the background thread. |
| | |
| | """ |
| | with self._update_lock: |
| | if self._round <= 0: |
| | |
| | self._replace_done = True |
| | return True, -1 |
| |
|
| | if self._round != check_round: |
| | self._compute_replacements() |
| | return False, self._round |
| |
|
| | def manage_replacement(self): |
| | """ |
| | Background thread for replacement. |
| | |
| | """ |
| | check_round: int = -1 |
| | done = False |
| | while not done: |
| | done, check_round = self._try_manage_replacement(check_round) |
| | time.sleep(0.01) |
| |
|
| | def __len__(self): |
| | """ |
| | The dataset length is given by cache_num instead of len(data). |
| | |
| | """ |
| | return self.cache_num |
| |
|
| |
|
| | class ZipDataset(Dataset): |
| | """ |
| | Zip several PyTorch datasets and output data(with the same index) together in a tuple. |
| | If the output of single dataset is already a tuple, flatten it and extend to the result. |
| | For example: if datasetA returns (img, imgmeta), datasetB returns (seg, segmeta), |
| | finally return (img, imgmeta, seg, segmeta). |
| | And if the datasets don't have same length, use the minimum length of them as the length |
| | of ZipDataset. |
| | |
| | Examples:: |
| | |
| | >>> zip_data = ZipDataset([[1, 2, 3], [4, 5]]) |
| | >>> print(len(zip_data)) |
| | 2 |
| | >>> for item in zip_data: |
| | >>> print(item) |
| | [1, 4] |
| | [2, 5] |
| | |
| | """ |
| |
|
| | def __init__(self, datasets: Sequence, transform: Optional[Callable] = None) -> None: |
| | """ |
| | Args: |
| | datasets: list of datasets to zip together. |
| | transform: a callable data transform operates on the zipped item from `datasets`. |
| | """ |
| | super().__init__(list(datasets), transform=transform) |
| |
|
| | def __len__(self) -> int: |
| | return min((len(dataset) for dataset in self.data)) |
| |
|
| | def __getitem__(self, index: int): |
| | def to_list(x): |
| | return list(x) if isinstance(x, (tuple, list)) else [x] |
| |
|
| | data = list() |
| | for dataset in self.data: |
| | data.extend(to_list(dataset[index])) |
| | if self.transform is not None: |
| | data = apply_transform(self.transform, data, map_items=False) |
| | |
| | return tuple(data) |
| |
|
| |
|
| | class ArrayDataset(Randomizable, _TorchDataset): |
| | """ |
| | Dataset for segmentation and classification tasks based on array format input data and transforms. |
| | It ensures the same random seeds in the randomized transforms defined for image, segmentation and label. |
| | The `transform` can be :py:class:`monai.transforms.Compose` or any other callable object. |
| | For example: |
| | If train based on Nifti format images without metadata, all transforms can be composed:: |
| | |
| | img_transform = Compose( |
| | [ |
| | LoadNifti(image_only=True), |
| | AddChannel(), |
| | RandAdjustContrast() |
| | ] |
| | ) |
| | ArrayDataset(img_file_list, img_transform=img_transform) |
| | |
| | If training based on images and the metadata, the array transforms can not be composed |
| | because several transforms receives multiple parameters or return multiple values. Then Users need |
| | to define their own callable method to parse metadata from `LoadNifti` or set `affine` matrix |
| | to `Spacing` transform:: |
| | |
| | class TestCompose(Compose): |
| | def __call__(self, input_): |
| | img, metadata = self.transforms[0](input_) |
| | img = self.transforms[1](img) |
| | img, _, _ = self.transforms[2](img, metadata["affine"]) |
| | return self.transforms[3](img), metadata |
| | img_transform = TestCompose( |
| | [ |
| | LoadNifti(image_only=False), |
| | AddChannel(), |
| | Spacing(pixdim=(1.5, 1.5, 3.0)), |
| | RandAdjustContrast() |
| | ] |
| | ) |
| | ArrayDataset(img_file_list, img_transform=img_transform) |
| | |
| | Examples:: |
| | |
| | >>> ds = ArrayDataset([1, 2, 3, 4], lambda x: x + 0.1) |
| | >>> print(ds[0]) |
| | 1.1 |
| | |
| | >>> ds = ArrayDataset(img=[1, 2, 3, 4], seg=[5, 6, 7, 8]) |
| | >>> print(ds[0]) |
| | [1, 5] |
| | |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | img: Sequence, |
| | img_transform: Optional[Callable] = None, |
| | seg: Optional[Sequence] = None, |
| | seg_transform: Optional[Callable] = None, |
| | labels: Optional[Sequence] = None, |
| | label_transform: Optional[Callable] = None, |
| | ) -> None: |
| | """ |
| | Initializes the dataset with the filename lists. The transform `img_transform` is applied |
| | to the images and `seg_transform` to the segmentations. |
| | |
| | Args: |
| | img: sequence of images. |
| | img_transform: transform to apply to each element in `img`. |
| | seg: sequence of segmentations. |
| | seg_transform: transform to apply to each element in `seg`. |
| | labels: sequence of labels. |
| | label_transform: transform to apply to each element in `labels`. |
| | |
| | """ |
| | items = [(img, img_transform), (seg, seg_transform), (labels, label_transform)] |
| | self.set_random_state(seed=get_seed()) |
| | datasets = [Dataset(x[0], x[1]) for x in items if x[0] is not None] |
| | self.dataset = datasets[0] if len(datasets) == 1 else ZipDataset(datasets) |
| |
|
| | self._seed = 0 |
| |
|
| | def __len__(self) -> int: |
| | return len(self.dataset) |
| |
|
| | def randomize(self, data: Optional[Any] = None) -> None: |
| | self._seed = self.R.randint(np.iinfo(np.int32).max) |
| |
|
| | def __getitem__(self, index: int): |
| | self.randomize() |
| | if isinstance(self.dataset, ZipDataset): |
| | |
| | for dataset in self.dataset.data: |
| | transform = getattr(dataset, "transform", None) |
| | if isinstance(transform, Randomizable): |
| | transform.set_random_state(seed=self._seed) |
| | transform = getattr(self.dataset, "transform", None) |
| | if isinstance(transform, Randomizable): |
| | transform.set_random_state(seed=self._seed) |
| | return self.dataset[index] |
| |
|