|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
from __future__ import annotations |
|
|
|
|
|
from collections.abc import Callable, Sequence |
|
|
|
|
|
import torch |
|
|
|
|
|
from monai.data import decollate_batch, list_data_collate |
|
|
from monai.engines import SupervisedEvaluator, SupervisedTrainer |
|
|
from monai.engines.utils import IterationEvents |
|
|
from monai.transforms import Compose |
|
|
from monai.utils.enums import CommonKeys |
|
|
|
|
|
|
|
|
class Interaction: |
|
|
""" |
|
|
Ignite process_function used to introduce interactions (simulation of clicks) for Deepgrow Training/Evaluation. |
|
|
For more details please refer to: https://pytorch.org/ignite/generated/ignite.engine.engine.Engine.html. |
|
|
This implementation is based on: |
|
|
|
|
|
Sakinis et al., Interactive segmentation of medical images through |
|
|
fully convolutional neural networks. (2019) https://arxiv.org/abs/1903.08205 |
|
|
|
|
|
Args: |
|
|
transforms: execute additional transformation during every iteration (before train). |
|
|
Typically, several Tensor based transforms composed by `Compose`. |
|
|
max_interactions: maximum number of interactions per iteration |
|
|
train: training or evaluation |
|
|
key_probability: field name to fill probability for every interaction |
|
|
""" |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
transforms: Sequence[Callable] | Callable, |
|
|
max_interactions: int, |
|
|
train: bool, |
|
|
key_probability: str = "probability", |
|
|
) -> None: |
|
|
if not isinstance(transforms, Compose): |
|
|
transforms = Compose(transforms) |
|
|
|
|
|
self.transforms: Compose = transforms |
|
|
self.max_interactions = max_interactions |
|
|
self.train = train |
|
|
self.key_probability = key_probability |
|
|
|
|
|
def __call__(self, engine: SupervisedTrainer | SupervisedEvaluator, batchdata: dict[str, torch.Tensor]) -> dict: |
|
|
if batchdata is None: |
|
|
raise ValueError("Must provide batch data for current iteration.") |
|
|
|
|
|
for j in range(self.max_interactions): |
|
|
inputs, _ = engine.prepare_batch(batchdata) |
|
|
inputs = inputs.to(engine.state.device) |
|
|
|
|
|
engine.fire_event(IterationEvents.INNER_ITERATION_STARTED) |
|
|
|
|
|
engine.network.eval() |
|
|
with torch.no_grad(): |
|
|
if engine.amp: |
|
|
with torch.cuda.amp.autocast(): |
|
|
predictions = engine.inferer(inputs, engine.network) |
|
|
else: |
|
|
predictions = engine.inferer(inputs, engine.network) |
|
|
|
|
|
engine.fire_event(IterationEvents.INNER_ITERATION_COMPLETED) |
|
|
|
|
|
batchdata.update({CommonKeys.PRED: predictions}) |
|
|
|
|
|
|
|
|
batchdata_list = decollate_batch(batchdata, detach=True) |
|
|
for i in range(len(batchdata_list)): |
|
|
batchdata_list[i][self.key_probability] = ( |
|
|
(1.0 - ((1.0 / self.max_interactions) * j)) if self.train else 1.0 |
|
|
) |
|
|
batchdata_list[i] = self.transforms(batchdata_list[i]) |
|
|
|
|
|
|
|
|
batchdata = list_data_collate(batchdata_list) |
|
|
|
|
|
return engine._iteration(engine, batchdata) |
|
|
|