| | |
| | from contextlib import contextmanager |
| |
|
| | import torch |
| | import torch.nn as nn |
| |
|
| | from mmengine.device import is_cuda_available, is_npu_available |
| | from mmengine.registry import OPTIM_WRAPPERS |
| | from mmengine.utils import digit_version |
| | from mmengine.utils.dl_utils import TORCH_VERSION |
| | from .optimizer_wrapper import OptimWrapper |
| |
|
| | if is_npu_available(): |
| | from torch.npu.amp import GradScaler |
| | else: |
| | from torch.cuda.amp import GradScaler |
| |
|
| |
|
| | @OPTIM_WRAPPERS.register_module() |
| | class AmpOptimWrapper(OptimWrapper): |
| | """A subclass of :class:`OptimWrapper` that supports automatic mixed |
| | precision training based on torch.cuda.amp. |
| | |
| | ``AmpOptimWrapper`` provides a unified interface with |
| | ``OptimWrapper``, so ``AmpOptimWrapper`` can be used in the same way |
| | as ``OptimWrapper``. |
| | |
| | Warnings: |
| | ``AmpOptimWrapper`` requires PyTorch >= 1.6. |
| | |
| | Args: |
| | loss_scale (float or str or dict): The initial configuration of |
| | `torch.cuda.amp.GradScaler`. See more specific arguments |
| | introduction at `PyTorch AMP <https://pytorch.org/docs/stable/amp.html?highlight=gradscalertorch.cuda.amp.GradScaler>`_ # noqa: E501 |
| | Defaults to ``dynamic``. |
| | |
| | - "dynamic": Initialize GradScale without any arguments. |
| | - float: Initialize GradScaler with ``init_scale``. |
| | - dict: Initialize GradScaler with more detail configuration. |
| | |
| | **kwargs: Keyword arguments passed to OptimWrapper. |
| | |
| | Note: |
| | If you use ``IterBasedRunner`` and enable gradient accumulation, |
| | the original `max_iters` should be multiplied by |
| | ``accumulative_counts``. |
| | """ |
| |
|
| | def __init__(self, loss_scale='dynamic', **kwargs): |
| | assert digit_version(TORCH_VERSION) >= digit_version('1.6.0'), ( |
| | '`torch.cuda.amp` is only available when pytorch version >= 1.6') |
| | assert is_cuda_available() or is_npu_available(), ( |
| | '``AmpOptimizerWrapper`` is only available training on gpu or npu') |
| | super().__init__(**kwargs) |
| | self._scale_update_param = None |
| | if loss_scale == 'dynamic': |
| | |
| | |
| | self.loss_scaler = GradScaler() |
| | elif isinstance(loss_scale, float): |
| | |
| | self._scale_update_param = loss_scale |
| | self.loss_scaler = GradScaler(init_scale=loss_scale) |
| | elif isinstance(loss_scale, dict): |
| | |
| | self.loss_scaler = GradScaler(**loss_scale) |
| | else: |
| | raise TypeError('loss_scale must be of type float, dict, or ' |
| | f'"dynamic", but got {loss_scale}') |
| |
|
| | def backward(self, loss: torch.Tensor, **kwargs): |
| | """Perform gradient back propagation with :attr:`loss_scaler`. |
| | |
| | Args: |
| | loss (torch.Tensor): The loss of current iteration. |
| | kwargs: Keyword arguments passed to :meth:`torch.Tensor.backward` |
| | """ |
| | self.loss_scaler.scale(loss).backward(**kwargs) |
| | self._inner_count += 1 |
| |
|
| | def step(self, **kwargs): |
| | """Update parameters with :attr:`loss_scaler`. |
| | |
| | Args: |
| | kwargs: Keyword arguments passed to |
| | :meth:`torch.optim.Optimizer.step`. |
| | """ |
| | if self.clip_grad_kwargs: |
| | self.loss_scaler.unscale_(self.optimizer) |
| | self._clip_grad() |
| | self.loss_scaler.step(self.optimizer, **kwargs) |
| | self.loss_scaler.update(self._scale_update_param) |
| |
|
| | def state_dict(self) -> dict: |
| | """Get the state dictionary of :attr:`optimizer` and |
| | :attr:`loss_scaler`. |
| | |
| | Based on the state dictionary of the optimizer, the returned state |
| | dictionary will add a key named "loss_scaler". |
| | |
| | Returns: |
| | dict: The merged state dict of :attr:`loss_scaler` and |
| | :attr:`optimizer`. |
| | """ |
| | |
| | state_dict = self.optimizer.state_dict() |
| | state_dict['loss_scaler'] = self.loss_scaler.state_dict() |
| | return state_dict |
| |
|
| | def load_state_dict(self, state_dict: dict): |
| | """Load and parse the state dictionary of :attr:`optimizer` and |
| | :attr:`loss_scaler`. |
| | |
| | If state_dict contains "loss_scaler.", the :attr:`loss_scaler` will |
| | load the corresponding keys. Otherwise, only the :attr:`optimizer` |
| | will load the state dictionary. |
| | |
| | Args: |
| | state_dict (dict): The state dict of :attr:`optimizer` and |
| | :attr:`loss_scaler` |
| | """ |
| | if 'loss_scaler' in state_dict: |
| | self.loss_scaler.load_state_dict(state_dict.pop('loss_scaler')) |
| | self.optimizer.load_state_dict(state_dict) |
| |
|
| | @contextmanager |
| | def optim_context(self, model: nn.Module): |
| | """Enables the context for mixed precision training, and enables the |
| | context for disabling gradient synchronization during gradient |
| | accumulation context. |
| | |
| | Args: |
| | model (nn.Module): The training model. |
| | """ |
| | from mmengine.runner.amp import autocast |
| | with super().optim_context(model), autocast(): |
| | yield |
| |
|