| | |
| | import os |
| |
|
| | import pytest |
| | import torch |
| | import torch.nn as nn |
| | from torch.distributed import destroy_process_group, init_process_group |
| | from torch.nn.parallel import DataParallel, DistributedDataParallel |
| |
|
| | from mmengine.model import (MMDistributedDataParallel, |
| | MMSeparateDistributedDataParallel, |
| | convert_sync_batchnorm, is_model_wrapper, |
| | revert_sync_batchnorm) |
| | from mmengine.registry import MODEL_WRAPPERS, Registry |
| | from mmengine.utils import is_installed |
| |
|
| |
|
| | class ToyModule(nn.Module): |
| |
|
| | def __init__(self): |
| | super().__init__() |
| | self.layer1 = nn.Linear(1, 1) |
| |
|
| | def add_module(self, name, module): |
| | raise ValueError() |
| |
|
| |
|
| | @pytest.mark.skipif( |
| | torch.__version__ == 'parrots', reason='not supported in parrots now') |
| | def test_revert_syncbn(): |
| | |
| | conv = nn.Sequential(nn.Conv2d(3, 8, 2), nn.SyncBatchNorm(8)) |
| | x = torch.randn(1, 3, 10, 10) |
| | |
| | with pytest.raises(ValueError): |
| | y = conv(x) |
| | conv = revert_sync_batchnorm(conv) |
| | y = conv(x) |
| | assert y.shape == (1, 8, 9, 9) |
| |
|
| | |
| | |
| | |
| | conv = nn.Sequential(ToyModule(), nn.SyncBatchNorm(8)) |
| | revert_sync_batchnorm(conv) |
| |
|
| |
|
| | @pytest.mark.skipif( |
| | torch.__version__ == 'parrots', reason='not supported in parrots now') |
| | def test_convert_syncbn(): |
| | |
| | conv = nn.Sequential(nn.Conv2d(3, 8, 2), nn.BatchNorm2d(8)) |
| | x = torch.randn(1, 3, 10, 10) |
| | y = conv(x) |
| | assert y.shape == (1, 8, 9, 9) |
| |
|
| | |
| | if is_installed('mmcv'): |
| | |
| | |
| | |
| | with pytest.raises((RuntimeError, AssertionError)): |
| | convert_sync_batchnorm(conv, implementation='mmcv') |
| |
|
| | |
| | |
| | converted_conv = convert_sync_batchnorm(conv) |
| | assert isinstance(converted_conv[1], torch.nn.SyncBatchNorm) |
| | with pytest.raises(ValueError): |
| | converted_conv(x) |
| |
|
| |
|
| | def test_is_model_wrapper(): |
| | |
| | os.environ['MASTER_ADDR'] = '127.0.0.1' |
| | os.environ['MASTER_PORT'] = '29510' |
| | os.environ['RANK'] = str(0) |
| | init_process_group(backend='gloo', rank=0, world_size=1) |
| | model = nn.Linear(1, 1) |
| |
|
| | for wrapper in [ |
| | DistributedDataParallel, MMDistributedDataParallel, |
| | MMSeparateDistributedDataParallel, DataParallel |
| | ]: |
| | wrapper_model = wrapper(model) |
| | assert is_model_wrapper(wrapper_model) |
| |
|
| | |
| | |
| | CHILD_REGISTRY = Registry('test_is_model_wrapper', parent=MODEL_WRAPPERS) |
| |
|
| | class CustomModelWrapper(nn.Module): |
| |
|
| | def __init__(self, model): |
| | super().__init__() |
| | self.module = model |
| |
|
| | pass |
| |
|
| | CHILD_REGISTRY.register_module(module=CustomModelWrapper, force=True) |
| |
|
| | for wrapper in [ |
| | DistributedDataParallel, MMDistributedDataParallel, |
| | MMSeparateDistributedDataParallel, DataParallel, CustomModelWrapper |
| | ]: |
| | wrapper_model = wrapper(model) |
| | assert is_model_wrapper(wrapper_model) |
| |
|
| | |
| | |
| | for wrapper in [ |
| | DistributedDataParallel, MMDistributedDataParallel, |
| | MMSeparateDistributedDataParallel, DataParallel |
| | ]: |
| | wrapper_model = wrapper(model) |
| | assert not is_model_wrapper(wrapper_model, registry=CHILD_REGISTRY) |
| |
|
| | wrapper_model = CustomModelWrapper(model) |
| | assert is_model_wrapper(wrapper_model, registry=CHILD_REGISTRY) |
| | destroy_process_group() |
| |
|