| | |
| | from contextlib import contextmanager |
| | from typing import Dict, Iterator, List, Optional, Tuple |
| |
|
| | import torch |
| | import torch.nn as nn |
| |
|
| | from .optimizer_wrapper import OptimWrapper |
| |
|
| |
|
| | class OptimWrapperDict(OptimWrapper): |
| | """A dictionary container of :obj:`OptimWrapper`. |
| | |
| | If runner is training with multiple optimizers, all optimizer wrappers |
| | should be managed by :obj:`OptimWrapperDict` which is built by |
| | ``CustomOptimWrapperConstructor``. ``OptimWrapperDict`` will load and save |
| | the state dictionary of all optimizer wrappers. |
| | |
| | Consider the semantic ambiguity of calling :meth:``update_params``, |
| | :meth:`backward` of all optimizer wrappers, ``OptimWrapperDict`` will not |
| | implement these methods. |
| | |
| | Examples: |
| | >>> import torch.nn as nn |
| | >>> from torch.optim import SGD |
| | >>> from mmengine.optim import OptimWrapperDict, OptimWrapper |
| | >>> model1 = nn.Linear(1, 1) |
| | >>> model2 = nn.Linear(1, 1) |
| | >>> optim_wrapper1 = OptimWrapper(SGD(model1.parameters(), lr=0.1)) |
| | >>> optim_wrapper2 = OptimWrapper(SGD(model2.parameters(), lr=0.1)) |
| | >>> optim_wrapper_dict = OptimWrapperDict(model1=optim_wrapper1, |
| | >>> model2=optim_wrapper2) |
| | |
| | Note: |
| | The optimizer wrapper contained in ``OptimWrapperDict`` can be accessed |
| | in the same way as `dict`. |
| | |
| | Args: |
| | **optim_wrappers: A dictionary of ``OptimWrapper`` instance. |
| | """ |
| |
|
| | def __init__(self, **optim_wrapper_dict: OptimWrapper): |
| | for key, value in optim_wrapper_dict.items(): |
| | assert isinstance(value, OptimWrapper), ( |
| | '`OptimWrapperDict` only accept OptimWrapper instance, ' |
| | f'but got {key}: {type(value)}') |
| | self.optim_wrappers = optim_wrapper_dict |
| |
|
| | def update_params(self, |
| | loss: torch.Tensor, |
| | step_kwargs: Optional[Dict] = None, |
| | zero_kwargs: Optional[Dict] = None) -> None: |
| | """Update all optimizer wrappers would lead to a duplicate backward |
| | errors, and OptimWrapperDict does not know which optimizer wrapper |
| | should be updated. |
| | |
| | Therefore, this method is not implemented. The optimizer wrapper of |
| | OptimWrapperDict should be accessed and call its `update_params. |
| | """ |
| | raise NotImplementedError('`update_params` should be called by each ' |
| | 'optimizer separately`') |
| |
|
| | def backward(self, loss: torch.Tensor, **kwargs) -> None: |
| | """Since OptimWrapperDict doesn't know which optimizer wrapper's |
| | backward method should be called (``loss_scaler`` maybe different in |
| | different :obj:AmpOptimWrapper), this method is not implemented. |
| | |
| | The optimizer wrapper of OptimWrapperDict should be accessed and call |
| | its `backward. |
| | """ |
| | raise NotImplementedError('`backward` should be called by each ' |
| | 'optimizer separately`') |
| |
|
| | def step(self, **kwargs) -> None: |
| | """Since the backward method is not implemented, the step should not be |
| | implemented either.""" |
| | raise NotImplementedError('`step` should be called by each ' |
| | 'optimizer separately`') |
| |
|
| | def zero_grad(self, **kwargs) -> None: |
| | """Set the gradients of all optimizer wrappers to zero.""" |
| | for optim_wrapper in self.optim_wrappers.values(): |
| | optim_wrapper.zero_grad() |
| |
|
| | @contextmanager |
| | def optim_context(self, model: nn.Module): |
| | """``optim_context`` should be called by each optimizer separately.""" |
| | raise NotImplementedError( |
| | '`optim_context` should be called by each optimizer separately') |
| |
|
| | def initialize_count_status(self, model: nn.Module, cur_iter, |
| | max_iters) -> None: |
| | """Do nothing but provide unified interface for :obj:`OptimWrapper` |
| | |
| | Since ``OptimWrapperDict`` does not know the correspondence between |
| | model and optimizer wrapper. ``initialize_iter_status`` will do nothing |
| | and each optimizer wrapper should call ``initialize_iter_status`` |
| | separately. |
| | """ |
| | return |
| |
|
| | @property |
| | def param_groups(self): |
| | """Returns the parameter groups of each OptimWrapper.""" |
| | param_groups = dict() |
| | for key, value in self.optim_wrappers.items(): |
| | param_groups[key] = value.param_groups |
| | return param_groups |
| |
|
| | def get_lr(self) -> Dict[str, List[float]]: |
| | """Get the learning rate of all optimizers. |
| | |
| | Returns: |
| | Dict[str, List[float]]: Learning rate of all optimizers. |
| | """ |
| | lr_dict = dict() |
| | for name, optim_wrapper in self.optim_wrappers.items(): |
| | lr_dict[f'{name}.lr'] = optim_wrapper.get_lr()['lr'] |
| | return lr_dict |
| |
|
| | def get_momentum(self) -> Dict[str, List[float]]: |
| | """Get the momentum of all optimizers. |
| | |
| | Returns: |
| | Dict[str, List[float]]: momentum of all optimizers. |
| | """ |
| | momentum_dict = dict() |
| | for name, optim_wrapper in self.optim_wrappers.items(): |
| | momentum_dict[f'{name}.momentum'] = optim_wrapper.get_momentum( |
| | )['momentum'] |
| | return momentum_dict |
| |
|
| | def state_dict(self) -> dict: |
| | """Get the state dictionary of all optimizer wrappers. |
| | |
| | Returns: |
| | dict: Each key-value pair in the dictionary represents the name |
| | and state dictionary of corresponding :obj:`OptimWrapper`. |
| | """ |
| | state_dict = dict() |
| | for name, optim_wrapper in self.optim_wrappers.items(): |
| | state_dict[name] = optim_wrapper.state_dict() |
| | return state_dict |
| |
|
| | def load_state_dict(self, state_dict: dict) -> None: |
| | """Load the state dictionary from the ``state_dict``. |
| | |
| | Args: |
| | state_dict (dict): Each key-value pair in `state_dict` represents |
| | the name and the state dictionary of corresponding |
| | :obj:`OptimWrapper`. |
| | """ |
| | for name, _state_dict in state_dict.items(): |
| | assert name in self.optim_wrappers, ( |
| | f'Mismatched `state_dict`! cannot found {name} in ' |
| | 'OptimWrapperDict') |
| | self.optim_wrappers[name].load_state_dict(_state_dict) |
| |
|
| | def items(self) -> Iterator[Tuple[str, OptimWrapper]]: |
| | """A generator to get the name and corresponding |
| | :obj:`OptimWrapper`""" |
| | yield from self.optim_wrappers.items() |
| |
|
| | def values(self) -> Iterator[OptimWrapper]: |
| | """A generator to get :obj:`OptimWrapper`""" |
| | yield from self.optim_wrappers.values() |
| |
|
| | def keys(self) -> Iterator[str]: |
| | """A generator to get the name of :obj:`OptimWrapper`""" |
| | yield from self.optim_wrappers.keys() |
| |
|
| | def __getitem__(self, key: str) -> OptimWrapper: |
| | assert key in self.optim_wrappers, ( |
| | f'Cannot find {key} in OptimWrapperDict, please check ' |
| | 'your optimizer constructor.') |
| | return self.optim_wrappers[key] |
| |
|
| | def __contains__(self, key: str) -> bool: |
| | return key in self.optim_wrappers |
| |
|
| | def __len__(self) -> int: |
| | return len(self.optim_wrappers) |
| |
|
| | def __repr__(self) -> str: |
| | desc = '' |
| | for name, optim_wrapper in self.optim_wrappers.items(): |
| | desc += f'name: {name}\n' |
| | desc += repr(optim_wrapper) |
| | return desc |
| |
|