|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
from __future__ import annotations |
|
|
|
|
|
import logging |
|
|
import os |
|
|
import warnings |
|
|
from collections.abc import Mapping |
|
|
from typing import TYPE_CHECKING, Any |
|
|
|
|
|
from monai.config import IgniteInfo |
|
|
from monai.utils import is_scalar, min_version, optional_import |
|
|
|
|
|
Events, _ = optional_import("ignite.engine", IgniteInfo.OPT_IMPORT_VERSION, min_version, "Events") |
|
|
|
|
|
if TYPE_CHECKING: |
|
|
from ignite.engine import Engine |
|
|
from ignite.handlers import Checkpoint, DiskSaver |
|
|
else: |
|
|
Engine, _ = optional_import("ignite.engine", IgniteInfo.OPT_IMPORT_VERSION, min_version, "Engine") |
|
|
DiskSaver, _ = optional_import("ignite.handlers", IgniteInfo.OPT_IMPORT_VERSION, min_version, "DiskSaver") |
|
|
Checkpoint, _ = optional_import("ignite.handlers", IgniteInfo.OPT_IMPORT_VERSION, min_version, "Checkpoint") |
|
|
|
|
|
|
|
|
class CheckpointSaver: |
|
|
""" |
|
|
CheckpointSaver acts as an Ignite handler to save checkpoint data into files. |
|
|
It supports to save according to metrics result, epoch number, iteration number |
|
|
and last model or exception. |
|
|
|
|
|
Args: |
|
|
save_dir: the target directory to save the checkpoints. |
|
|
save_dict: source objects that save to the checkpoint. examples:: |
|
|
|
|
|
{'network': net, 'optimizer': optimizer, 'lr_scheduler': lr_scheduler} |
|
|
|
|
|
name: identifier of logging.logger to use, if None, defaulting to ``engine.logger``. |
|
|
file_prefix: prefix for the filenames to which objects will be saved. |
|
|
save_final: whether to save checkpoint or session at final iteration or exception. |
|
|
If checkpoints are to be saved when an exception is raised, put this handler before |
|
|
`StatsHandler` in the handler list, because the logic with Ignite can only trigger |
|
|
the first attached handler for `EXCEPTION_RAISED` event. |
|
|
final_filename: set a fixed filename to save the final model if `save_final=True`. |
|
|
If None, default to `checkpoint_final_iteration=N.pt`. |
|
|
save_key_metric: whether to save checkpoint or session when the value of key_metric is |
|
|
higher than all the previous values during training.keep 4 decimal places of metric, |
|
|
checkpoint name is: {file_prefix}_key_metric=0.XXXX.pth. |
|
|
key_metric_name: the name of key_metric in ignite metrics dictionary. |
|
|
If None, use `engine.state.key_metric` instead. |
|
|
key_metric_n_saved: save top N checkpoints or sessions, sorted by the value of key |
|
|
metric in descending order. |
|
|
key_metric_filename: set a fixed filename to set the best metric model, if not None, |
|
|
`key_metric_n_saved` should be 1 and only keep the best metric model. |
|
|
key_metric_save_state: whether to save the tracking list of key metric in the checkpoint file. |
|
|
if `True`, then will save an object in the checkpoint file with key `checkpointer` to be |
|
|
consistent with the `include_self` arg of `Checkpoint` in ignite: |
|
|
https://pytorch.org/ignite/v0.4.5/generated/ignite.handlers.checkpoint.Checkpoint.html. |
|
|
typically, it's used to resume training and compare current metric with previous N values. |
|
|
key_metric_greater_or_equal: if `True`, the latest equally scored model is stored. Otherwise, |
|
|
save the first equally scored model. default to `False`. |
|
|
key_metric_negative_sign: whether adding a negative sign to the metric score to compare metrics, |
|
|
because for error-like metrics, smaller is better(objects with larger score are retained). |
|
|
default to `False`. |
|
|
epoch_level: save checkpoint during training for every N epochs or every N iterations. |
|
|
`True` is epoch level, `False` is iteration level. |
|
|
save_interval: save checkpoint every N epochs, default is 0 to save no checkpoint. |
|
|
n_saved: save latest N checkpoints of epoch level or iteration level, 'None' is to save all. |
|
|
|
|
|
Note: |
|
|
CheckpointHandler can be used during training, validation or evaluation. |
|
|
example of saved files: |
|
|
|
|
|
- checkpoint_iteration=400.pt |
|
|
- checkpoint_iteration=800.pt |
|
|
- checkpoint_epoch=1.pt |
|
|
- checkpoint_final_iteration=1000.pt |
|
|
- checkpoint_key_metric=0.9387.pt |
|
|
|
|
|
""" |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
save_dir: str, |
|
|
save_dict: dict, |
|
|
name: str | None = None, |
|
|
file_prefix: str = "", |
|
|
save_final: bool = False, |
|
|
final_filename: str | None = None, |
|
|
save_key_metric: bool = False, |
|
|
key_metric_name: str | None = None, |
|
|
key_metric_n_saved: int = 1, |
|
|
key_metric_filename: str | None = None, |
|
|
key_metric_save_state: bool = False, |
|
|
key_metric_greater_or_equal: bool = False, |
|
|
key_metric_negative_sign: bool = False, |
|
|
epoch_level: bool = True, |
|
|
save_interval: int = 0, |
|
|
n_saved: int | None = None, |
|
|
) -> None: |
|
|
if save_dir is None: |
|
|
raise AssertionError("must provide directory to save the checkpoints.") |
|
|
self.save_dir = save_dir |
|
|
if not (save_dict is not None and len(save_dict) > 0): |
|
|
raise AssertionError("must provide source objects to save.") |
|
|
self.save_dict = save_dict |
|
|
self.logger = logging.getLogger(name) |
|
|
self.epoch_level = epoch_level |
|
|
self.save_interval = save_interval |
|
|
self._final_checkpoint: Checkpoint | None = None |
|
|
self._key_metric_checkpoint: Checkpoint | None = None |
|
|
self._interval_checkpoint: Checkpoint | None = None |
|
|
self._name = name |
|
|
self._final_filename = final_filename |
|
|
|
|
|
class _DiskSaver(DiskSaver): |
|
|
""" |
|
|
Enhance the DiskSaver to support fixed filename. |
|
|
|
|
|
""" |
|
|
|
|
|
def __init__(self, dirname: str, filename: str | None = None): |
|
|
|
|
|
|
|
|
super().__init__(dirname=dirname, require_empty=False, atomic=False) |
|
|
self.filename = filename |
|
|
|
|
|
def __call__(self, checkpoint: Mapping, filename: str, metadata: Mapping | None = None) -> None: |
|
|
if self.filename is not None: |
|
|
filename = self.filename |
|
|
super().__call__(checkpoint=checkpoint, filename=filename, metadata=metadata) |
|
|
|
|
|
def remove(self, filename: str) -> None: |
|
|
if self.filename is not None: |
|
|
filename = self.filename |
|
|
super().remove(filename=filename) |
|
|
|
|
|
if save_final: |
|
|
|
|
|
def _final_func(engine: Engine) -> Any: |
|
|
return engine.state.iteration |
|
|
|
|
|
self._final_checkpoint = Checkpoint( |
|
|
to_save=self.save_dict, |
|
|
save_handler=_DiskSaver(dirname=self.save_dir, filename=self._final_filename), |
|
|
filename_prefix=file_prefix, |
|
|
score_function=_final_func, |
|
|
score_name="final_iteration", |
|
|
) |
|
|
|
|
|
if save_key_metric: |
|
|
|
|
|
def _score_func(engine: Engine) -> Any: |
|
|
if isinstance(key_metric_name, str): |
|
|
metric_name = key_metric_name |
|
|
elif hasattr(engine.state, "key_metric_name"): |
|
|
metric_name = engine.state.key_metric_name |
|
|
else: |
|
|
raise ValueError( |
|
|
f"Incompatible values: save_key_metric=True and key_metric_name={key_metric_name}." |
|
|
) |
|
|
metric = engine.state.metrics[metric_name] |
|
|
if not is_scalar(metric): |
|
|
warnings.warn( |
|
|
"key metric is not a scalar value, skip metric comparison and don't save a model." |
|
|
"please use other metrics as key metric, or change the `reduction` mode to 'mean'." |
|
|
f"got metric: {metric_name}={metric}." |
|
|
) |
|
|
return -1 |
|
|
return (-1 if key_metric_negative_sign else 1) * metric |
|
|
|
|
|
if key_metric_filename is not None and key_metric_n_saved > 1: |
|
|
raise ValueError("if using fixed filename to save the best metric model, we should only save 1 model.") |
|
|
|
|
|
self._key_metric_checkpoint = Checkpoint( |
|
|
to_save=self.save_dict, |
|
|
save_handler=_DiskSaver(dirname=self.save_dir, filename=key_metric_filename), |
|
|
filename_prefix=file_prefix, |
|
|
score_function=_score_func, |
|
|
score_name="key_metric", |
|
|
n_saved=key_metric_n_saved, |
|
|
include_self=key_metric_save_state, |
|
|
greater_or_equal=key_metric_greater_or_equal, |
|
|
) |
|
|
|
|
|
if save_interval > 0: |
|
|
|
|
|
def _interval_func(engine: Engine) -> Any: |
|
|
return engine.state.epoch if self.epoch_level else engine.state.iteration |
|
|
|
|
|
self._interval_checkpoint = Checkpoint( |
|
|
to_save=self.save_dict, |
|
|
save_handler=_DiskSaver(dirname=self.save_dir), |
|
|
filename_prefix=file_prefix, |
|
|
score_function=_interval_func, |
|
|
score_name="epoch" if self.epoch_level else "iteration", |
|
|
n_saved=n_saved, |
|
|
) |
|
|
|
|
|
def load_state_dict(self, state_dict: dict) -> None: |
|
|
""" |
|
|
Utility to resume the internal state of key metric tracking list if configured to save |
|
|
checkpoints based on the key metric value. |
|
|
Note to set `key_metric_save_state=True` when saving the previous checkpoint. |
|
|
|
|
|
Example:: |
|
|
|
|
|
CheckpointSaver( |
|
|
... |
|
|
save_key_metric=True, |
|
|
key_metric_save_state=True, # config to also save the state of this saver |
|
|
).attach(engine) |
|
|
engine.run(...) |
|
|
|
|
|
# resumed training with a new CheckpointSaver |
|
|
saver = CheckpointSaver(save_key_metric=True, ...) |
|
|
# load the previous key metric tracking list into saver |
|
|
CheckpointLoader("/test/model.pt"), {"checkpointer": saver}).attach(engine) |
|
|
|
|
|
""" |
|
|
if self._key_metric_checkpoint is not None: |
|
|
self._key_metric_checkpoint.load_state_dict(state_dict) |
|
|
else: |
|
|
warnings.warn("no key metric checkpoint saver to resume the key metric tracking list.") |
|
|
|
|
|
def attach(self, engine: Engine) -> None: |
|
|
""" |
|
|
Args: |
|
|
engine: Ignite Engine, it can be a trainer, validator or evaluator. |
|
|
""" |
|
|
if self._name is None: |
|
|
self.logger = engine.logger |
|
|
if self._final_checkpoint is not None: |
|
|
engine.add_event_handler(Events.COMPLETED, self.completed) |
|
|
engine.add_event_handler(Events.EXCEPTION_RAISED, self.exception_raised) |
|
|
if self._key_metric_checkpoint is not None: |
|
|
engine.add_event_handler(Events.EPOCH_COMPLETED, self.metrics_completed) |
|
|
if self._interval_checkpoint is not None: |
|
|
if self.epoch_level: |
|
|
engine.add_event_handler(Events.EPOCH_COMPLETED(every=self.save_interval), self.interval_completed) |
|
|
else: |
|
|
engine.add_event_handler(Events.ITERATION_COMPLETED(every=self.save_interval), self.interval_completed) |
|
|
|
|
|
def _delete_previous_final_ckpt(self): |
|
|
if self._final_checkpoint is not None: |
|
|
saved = self._final_checkpoint._saved |
|
|
if len(saved) > 0: |
|
|
item = saved.pop(0) |
|
|
self._final_checkpoint.save_handler.remove(item.filename) |
|
|
self.logger.info(f"Deleted previous saved final checkpoint: {item.filename}") |
|
|
|
|
|
def completed(self, engine: Engine) -> None: |
|
|
"""Callback for train or validation/evaluation completed Event. |
|
|
Save final checkpoint if configure save_final is True. |
|
|
|
|
|
Args: |
|
|
engine: Ignite Engine, it can be a trainer, validator or evaluator. |
|
|
""" |
|
|
if not callable(self._final_checkpoint): |
|
|
raise AssertionError("Error: _final_checkpoint function not specified.") |
|
|
|
|
|
self._delete_previous_final_ckpt() |
|
|
self._final_checkpoint(engine) |
|
|
if self.logger is None: |
|
|
raise AssertionError |
|
|
if not hasattr(self.logger, "info"): |
|
|
raise AssertionError("Error, provided logger has not info attribute.") |
|
|
if self._final_filename is not None: |
|
|
_final_checkpoint_path = os.path.join(self.save_dir, self._final_filename) |
|
|
else: |
|
|
_final_checkpoint_path = self._final_checkpoint.last_checkpoint |
|
|
self.logger.info(f"Train completed, saved final checkpoint: {_final_checkpoint_path}") |
|
|
|
|
|
def exception_raised(self, engine: Engine, e: Exception) -> None: |
|
|
"""Callback for train or validation/evaluation exception raised Event. |
|
|
Save current data as final checkpoint if configure save_final is True. This callback may be skipped |
|
|
because the logic with Ignite can only trigger the first attached handler for `EXCEPTION_RAISED` event. |
|
|
|
|
|
Args: |
|
|
engine: Ignite Engine, it can be a trainer, validator or evaluator. |
|
|
e: the exception caught in Ignite during engine.run(). |
|
|
""" |
|
|
if not callable(self._final_checkpoint): |
|
|
raise AssertionError("Error: _final_checkpoint function not specified.") |
|
|
|
|
|
self._delete_previous_final_ckpt() |
|
|
self._final_checkpoint(engine) |
|
|
if self.logger is None: |
|
|
raise AssertionError |
|
|
if not hasattr(self.logger, "info"): |
|
|
raise AssertionError("Error, provided logger has not info attribute.") |
|
|
if self._final_filename is not None: |
|
|
_final_checkpoint_path = os.path.join(self.save_dir, self._final_filename) |
|
|
else: |
|
|
_final_checkpoint_path = self._final_checkpoint.last_checkpoint |
|
|
self.logger.info(f"Exception raised, saved the last checkpoint: {_final_checkpoint_path}") |
|
|
raise e |
|
|
|
|
|
def metrics_completed(self, engine: Engine) -> None: |
|
|
"""Callback to compare metrics and save models in train or validation when epoch completed. |
|
|
|
|
|
Args: |
|
|
engine: Ignite Engine, it can be a trainer, validator or evaluator. |
|
|
""" |
|
|
if not callable(self._key_metric_checkpoint): |
|
|
raise AssertionError("Error: _key_metric_checkpoint function not specified.") |
|
|
self._key_metric_checkpoint(engine) |
|
|
|
|
|
def interval_completed(self, engine: Engine) -> None: |
|
|
"""Callback for train epoch/iteration completed Event. |
|
|
Save checkpoint if configure save_interval = N |
|
|
|
|
|
Args: |
|
|
engine: Ignite Engine, it can be a trainer, validator or evaluator. |
|
|
""" |
|
|
if not callable(self._interval_checkpoint): |
|
|
raise AssertionError("Error: _interval_checkpoint function not specified.") |
|
|
self._interval_checkpoint(engine) |
|
|
if self.logger is None: |
|
|
raise AssertionError |
|
|
if not hasattr(self.logger, "info"): |
|
|
raise AssertionError("Error, provided logger has not info attribute.") |
|
|
if self.epoch_level: |
|
|
self.logger.info(f"Saved checkpoint at epoch: {engine.state.epoch}") |
|
|
else: |
|
|
self.logger.info(f"Saved checkpoint at iteration: {engine.state.iteration}") |
|
|
|