|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
from __future__ import annotations |
|
|
|
|
|
import logging |
|
|
import warnings |
|
|
from typing import TYPE_CHECKING |
|
|
|
|
|
import torch |
|
|
|
|
|
from monai.config import IgniteInfo |
|
|
from monai.networks.utils import copy_model_state |
|
|
from monai.utils import min_version, optional_import |
|
|
|
|
|
Events, _ = optional_import("ignite.engine", IgniteInfo.OPT_IMPORT_VERSION, min_version, "Events") |
|
|
Checkpoint, _ = optional_import("ignite.handlers", IgniteInfo.OPT_IMPORT_VERSION, min_version, "Checkpoint") |
|
|
if TYPE_CHECKING: |
|
|
from ignite.engine import Engine |
|
|
else: |
|
|
Engine, _ = optional_import("ignite.engine", IgniteInfo.OPT_IMPORT_VERSION, min_version, "Engine") |
|
|
|
|
|
|
|
|
class CheckpointLoader: |
|
|
""" |
|
|
CheckpointLoader acts as an Ignite handler to load checkpoint data from file. |
|
|
It can load variables for network, optimizer, lr_scheduler, etc. |
|
|
If saving checkpoint after `torch.nn.DataParallel`, need to save `model.module` instead |
|
|
as PyTorch recommended and then use this loader to load the model. |
|
|
|
|
|
Usage example:: |
|
|
|
|
|
trainer = SupervisedTrainer(...) |
|
|
save_dict = { |
|
|
"trainer": trainer, |
|
|
"net": network, |
|
|
"opt": optimizer, |
|
|
"lr": lr_scheduler, |
|
|
} |
|
|
|
|
|
map_location = "cuda:0" |
|
|
# checkpoint needs to have same save_dict for this to work |
|
|
handler = CheckpointLoader(load_path="/test/checkpoint.pt", load_dict=save_dict, map_location=map_location, strict=True) |
|
|
handler(trainer) |
|
|
# Trainer now has the same state as stored, including the number of epochs and iterations completed |
|
|
# so you can resume an interrupted training at the place where it left |
|
|
|
|
|
Args: |
|
|
load_path: the file path of checkpoint, it should be a PyTorch `pth` file. |
|
|
load_dict: target objects that load checkpoint to. examples:: |
|
|
|
|
|
{'network': net, 'optimizer': optimizer, 'lr_scheduler': lr_scheduler} |
|
|
|
|
|
name: identifier of logging.logger to use, if None, defaulting to ``engine.logger``. |
|
|
map_location: when loading the module for distributed training/evaluation, |
|
|
need to provide an appropriate map_location argument to prevent a process |
|
|
to step into others’ devices. If map_location is missing, torch.load will |
|
|
first load the module to CPU and then copy each parameter to where it was |
|
|
saved, which would result in all processes on the same machine using the |
|
|
same set of devices. |
|
|
strict: whether to strictly enforce that the keys and data shape in the `state_dict` of every item |
|
|
of `load_dict` match the `state_dict` of the corresponding items of checkpoint, default to `True`. |
|
|
strict_shape: whether to enforce the data shape of the matched layers in the checkpoint, |
|
|
`if `False`, it will skip the layers that have different data shape with checkpoint content, |
|
|
and ignore the `strict` arg. this can be useful advanced feature for transfer learning. |
|
|
users should totally understand which layers will have different shape. default to `True`. |
|
|
|
|
|
Note: if `strict_shape=False`, will only load checkpoint for `torch.nn.Module` and skip other |
|
|
items in the `load_dict`. For example, if the shape of some layers in current model can't |
|
|
match the checkpoint, the `parameter_group` of current optimizer may also can't match the |
|
|
checkpoint, so skip loading checkpoint for optimizer. |
|
|
|
|
|
For more details about loading checkpoint, please refer to: |
|
|
https://pytorch.org/ignite/v0.4.5/generated/ignite.handlers.checkpoint.Checkpoint.html |
|
|
#ignite.handlers.checkpoint.Checkpoint.load_objects. |
|
|
https://pytorch.org/docs/stable/generated/torch.nn.Module.html#torch.nn.Module.load_state_dict. |
|
|
|
|
|
""" |
|
|
|
|
|
def __init__( |
|
|
self, |
|
|
load_path: str, |
|
|
load_dict: dict, |
|
|
name: str | None = None, |
|
|
map_location: dict | None = None, |
|
|
strict: bool = True, |
|
|
strict_shape: bool = True, |
|
|
) -> None: |
|
|
if load_path is None: |
|
|
raise AssertionError("must provide clear path to load checkpoint.") |
|
|
self.load_path = load_path |
|
|
if load_dict is None or len(load_dict) <= 0: |
|
|
raise AssertionError("must provide target objects to load.") |
|
|
self.logger = logging.getLogger(name) |
|
|
self.load_dict = load_dict |
|
|
self._name = name |
|
|
self.map_location = map_location |
|
|
if strict and not strict_shape: |
|
|
warnings.warn("as `strict_shape` is already False, change `strict` to False.") |
|
|
strict = False |
|
|
self.strict = strict |
|
|
self.strict_shape = strict_shape |
|
|
|
|
|
def attach(self, engine: Engine) -> None: |
|
|
""" |
|
|
Args: |
|
|
engine: Ignite Engine, it can be a trainer, validator or evaluator. |
|
|
""" |
|
|
if self._name is None: |
|
|
self.logger = engine.logger |
|
|
engine.add_event_handler(Events.STARTED, self) |
|
|
|
|
|
def __call__(self, engine: Engine) -> None: |
|
|
""" |
|
|
Args: |
|
|
engine: Ignite Engine, it can be a trainer, validator or evaluator. |
|
|
""" |
|
|
checkpoint = torch.load(self.load_path, map_location=self.map_location) |
|
|
|
|
|
k, _ = list(self.load_dict.items())[0] |
|
|
|
|
|
if len(self.load_dict) == 1 and k not in checkpoint: |
|
|
checkpoint = {k: checkpoint} |
|
|
|
|
|
if not self.strict_shape: |
|
|
pop_items: list[str] = [] |
|
|
for k, obj in self.load_dict.items(): |
|
|
if isinstance(obj, torch.nn.Module): |
|
|
|
|
|
checkpoint[k] = copy_model_state(obj, checkpoint, inplace=False)[0] |
|
|
else: |
|
|
warnings.warn("`strict_shape` is False, load checkpoint for model, skip others in `load_dict`.") |
|
|
pop_items.append(k) |
|
|
for i in pop_items: |
|
|
self.load_dict.pop(i) |
|
|
|
|
|
|
|
|
prior_max_epochs = engine.state.max_epochs |
|
|
Checkpoint.load_objects(to_load=self.load_dict, checkpoint=checkpoint, strict=self.strict) |
|
|
if prior_max_epochs is not None and engine.state.epoch > prior_max_epochs: |
|
|
raise ValueError( |
|
|
f"Epoch count ({engine.state.epoch}) in checkpoint is larger than " |
|
|
f"the `engine.state.max_epochs` ({prior_max_epochs}) of engine. To further train from checkpoint, " |
|
|
"construct trainer with `max_epochs` larger than checkpoint's epoch count. " |
|
|
"To use checkpoint for inference, no need to load state_dict for the engine." |
|
|
) |
|
|
engine.state.max_epochs = prior_max_epochs |
|
|
|
|
|
self.logger.info(f"Restored all variables from {self.load_path}") |
|
|
|