| |
| from unittest.mock import patch |
|
|
| import torch |
| import torch.nn as nn |
|
|
| from mmengine.model import BaseModel |
| from mmengine.registry import MODELS |
|
|
|
|
| @MODELS.register_module() |
| class ToyModel(BaseModel): |
|
|
| def __init__(self, *args, **kwargs): |
| super().__init__() |
| self.conv = nn.Conv2d(1, 1, 1) |
|
|
| def forward(self, *args, **kwargs): |
| return {'loss': torch.tensor(0.0)} |
|
|
|
|
| def update_params_step(self, loss): |
| pass |
|
|
|
|
| def runtimeinfo_step(self, runner, batch_idx, data_batch=None): |
| runner.message_hub.update_info('iter', runner.iter) |
| lr_dict = runner.optim_wrapper.get_lr() |
| for name, lr in lr_dict.items(): |
| runner.message_hub.update_scalar(f'train/{name}', lr[0]) |
|
|
| momentum_dict = runner.optim_wrapper.get_momentum() |
| for name, momentum in momentum_dict.items(): |
| runner.message_hub.update_scalar(f'train/{name}', momentum[0]) |
|
|
|
|
| @patch('mmengine.optim.optimizer.OptimWrapper.update_params', |
| update_params_step) |
| @patch('mmengine.hooks.RuntimeInfoHook.before_train_iter', runtimeinfo_step) |
| def fake_run(cfg): |
| from mmengine.runner import Runner |
| cfg.pop('model') |
| cfg.pop('visualizer') |
| cfg.pop('val_dataloader') |
| cfg.pop('val_evaluator') |
| cfg.pop('val_cfg') |
| cfg.pop('test_dataloader') |
| cfg.pop('test_evaluator') |
| cfg.pop('test_cfg') |
| extra_cfg = dict( |
| model=dict(type='ToyModel'), |
| visualizer=dict( |
| type='Visualizer', |
| vis_backends=[ |
| dict(type='TensorboardVisBackend', save_dir='temp_dir') |
| ]), |
| ) |
| cfg.merge_from_dict(extra_cfg) |
| |
| runner = Runner.from_cfg(cfg) |
|
|
| |
| runner.train() |
|
|