| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| from __future__ import annotations |
|
|
| import logging |
| import warnings |
| from typing import TYPE_CHECKING |
|
|
| import torch |
|
|
| from monai.config import IgniteInfo |
| from monai.networks.utils import copy_model_state |
| from monai.utils import min_version, optional_import |
|
|
| Events, _ = optional_import("ignite.engine", IgniteInfo.OPT_IMPORT_VERSION, min_version, "Events") |
| Checkpoint, _ = optional_import("ignite.handlers", IgniteInfo.OPT_IMPORT_VERSION, min_version, "Checkpoint") |
| if TYPE_CHECKING: |
| from ignite.engine import Engine |
| else: |
| Engine, _ = optional_import("ignite.engine", IgniteInfo.OPT_IMPORT_VERSION, min_version, "Engine") |
|
|
|
|
| class CheckpointLoader: |
| """ |
| CheckpointLoader acts as an Ignite handler to load checkpoint data from file. |
| It can load variables for network, optimizer, lr_scheduler, etc. |
| If saving checkpoint after `torch.nn.DataParallel`, need to save `model.module` instead |
| as PyTorch recommended and then use this loader to load the model. |
| |
| Usage example:: |
| |
| trainer = SupervisedTrainer(...) |
| save_dict = { |
| "trainer": trainer, |
| "net": network, |
| "opt": optimizer, |
| "lr": lr_scheduler, |
| } |
| |
| map_location = "cuda:0" |
| # checkpoint needs to have same save_dict for this to work |
| handler = CheckpointLoader(load_path="/test/checkpoint.pt", load_dict=save_dict, map_location=map_location, strict=True) |
| handler(trainer) |
| # Trainer now has the same state as stored, including the number of epochs and iterations completed |
| # so you can resume an interrupted training at the place where it left |
| |
| Args: |
| load_path: the file path of checkpoint, it should be a PyTorch `pth` file. |
| load_dict: target objects that load checkpoint to. examples:: |
| |
| {'network': net, 'optimizer': optimizer, 'lr_scheduler': lr_scheduler} |
| |
| name: identifier of logging.logger to use, if None, defaulting to ``engine.logger``. |
| map_location: when loading the module for distributed training/evaluation, |
| need to provide an appropriate map_location argument to prevent a process |
| to step into others’ devices. If map_location is missing, torch.load will |
| first load the module to CPU and then copy each parameter to where it was |
| saved, which would result in all processes on the same machine using the |
| same set of devices. |
| strict: whether to strictly enforce that the keys and data shape in the `state_dict` of every item |
| of `load_dict` match the `state_dict` of the corresponding items of checkpoint, default to `True`. |
| strict_shape: whether to enforce the data shape of the matched layers in the checkpoint, |
| `if `False`, it will skip the layers that have different data shape with checkpoint content, |
| and ignore the `strict` arg. this can be useful advanced feature for transfer learning. |
| users should totally understand which layers will have different shape. default to `True`. |
| |
| Note: if `strict_shape=False`, will only load checkpoint for `torch.nn.Module` and skip other |
| items in the `load_dict`. For example, if the shape of some layers in current model can't |
| match the checkpoint, the `parameter_group` of current optimizer may also can't match the |
| checkpoint, so skip loading checkpoint for optimizer. |
| |
| For more details about loading checkpoint, please refer to: |
| https://pytorch.org/ignite/v0.4.5/generated/ignite.handlers.checkpoint.Checkpoint.html |
| #ignite.handlers.checkpoint.Checkpoint.load_objects. |
| https://pytorch.org/docs/stable/generated/torch.nn.Module.html#torch.nn.Module.load_state_dict. |
| |
| """ |
|
|
| def __init__( |
| self, |
| load_path: str, |
| load_dict: dict, |
| name: str | None = None, |
| map_location: dict | None = None, |
| strict: bool = True, |
| strict_shape: bool = True, |
| ) -> None: |
| if load_path is None: |
| raise AssertionError("must provide clear path to load checkpoint.") |
| self.load_path = load_path |
| if load_dict is None or len(load_dict) <= 0: |
| raise AssertionError("must provide target objects to load.") |
| self.logger = logging.getLogger(name) |
| self.load_dict = load_dict |
| self._name = name |
| self.map_location = map_location |
| if strict and not strict_shape: |
| warnings.warn("as `strict_shape` is already False, change `strict` to False.") |
| strict = False |
| self.strict = strict |
| self.strict_shape = strict_shape |
|
|
| def attach(self, engine: Engine) -> None: |
| """ |
| Args: |
| engine: Ignite Engine, it can be a trainer, validator or evaluator. |
| """ |
| if self._name is None: |
| self.logger = engine.logger |
| engine.add_event_handler(Events.STARTED, self) |
|
|
| def __call__(self, engine: Engine) -> None: |
| """ |
| Args: |
| engine: Ignite Engine, it can be a trainer, validator or evaluator. |
| """ |
| checkpoint = torch.load(self.load_path, map_location=self.map_location) |
|
|
| k, _ = list(self.load_dict.items())[0] |
| |
| if len(self.load_dict) == 1 and k not in checkpoint: |
| checkpoint = {k: checkpoint} |
|
|
| if not self.strict_shape: |
| pop_items: list[str] = [] |
| for k, obj in self.load_dict.items(): |
| if isinstance(obj, torch.nn.Module): |
| |
| checkpoint[k] = copy_model_state(obj, checkpoint, inplace=False)[0] |
| else: |
| warnings.warn("`strict_shape` is False, load checkpoint for model, skip others in `load_dict`.") |
| pop_items.append(k) |
| for i in pop_items: |
| self.load_dict.pop(i) |
|
|
| |
| prior_max_epochs = engine.state.max_epochs |
| Checkpoint.load_objects(to_load=self.load_dict, checkpoint=checkpoint, strict=self.strict) |
| if prior_max_epochs is not None and engine.state.epoch > prior_max_epochs: |
| raise ValueError( |
| f"Epoch count ({engine.state.epoch}) in checkpoint is larger than " |
| f"the `engine.state.max_epochs` ({prior_max_epochs}) of engine. To further train from checkpoint, " |
| "construct trainer with `max_epochs` larger than checkpoint's epoch count. " |
| "To use checkpoint for inference, no need to load state_dict for the engine." |
| ) |
| engine.state.max_epochs = prior_max_epochs |
|
|
| self.logger.info(f"Restored all variables from {self.load_path}") |
|
|