| | |
| | import logging |
| | from unittest import TestCase |
| |
|
| | import torch |
| | from torch import nn |
| |
|
| | from mmengine.logging.logger import MMLogger |
| | from mmengine.model import BaseModule, ModuleDict, ModuleList, Sequential |
| | from mmengine.registry import Registry, build_from_cfg |
| |
|
| | COMPONENTS = Registry('component') |
| | FOOMODELS = Registry('model') |
| |
|
| | Logger = MMLogger.get_current_instance() |
| |
|
| |
|
| | @COMPONENTS.register_module() |
| | class FooConv1d(BaseModule): |
| |
|
| | def __init__(self, init_cfg=None): |
| | super().__init__(init_cfg) |
| | self.conv1d = nn.Conv1d(4, 1, 4) |
| |
|
| | def forward(self, x): |
| | return self.conv1d(x) |
| |
|
| |
|
| | @COMPONENTS.register_module() |
| | class FooConv2d(BaseModule): |
| |
|
| | def __init__(self, init_cfg=None): |
| | super().__init__(init_cfg) |
| | self.conv2d = nn.Conv2d(3, 1, 3) |
| |
|
| | def forward(self, x): |
| | return self.conv2d(x) |
| |
|
| |
|
| | @COMPONENTS.register_module() |
| | class FooLinear(BaseModule): |
| |
|
| | def __init__(self, init_cfg=None): |
| | super().__init__(init_cfg) |
| | self.linear = nn.Linear(3, 4) |
| |
|
| | def forward(self, x): |
| | return self.linear(x) |
| |
|
| |
|
| | @COMPONENTS.register_module() |
| | class FooLinearConv1d(BaseModule): |
| |
|
| | def __init__(self, linear=None, conv1d=None, init_cfg=None): |
| | super().__init__(init_cfg) |
| | if linear is not None: |
| | self.linear = build_from_cfg(linear, COMPONENTS) |
| | if conv1d is not None: |
| | self.conv1d = build_from_cfg(conv1d, COMPONENTS) |
| |
|
| | def forward(self, x): |
| | x = self.linear(x) |
| | return self.conv1d(x) |
| |
|
| |
|
| | @FOOMODELS.register_module() |
| | class FooModel(BaseModule): |
| |
|
| | def __init__(self, |
| | component1=None, |
| | component2=None, |
| | component3=None, |
| | component4=None, |
| | init_cfg=None) -> None: |
| | super().__init__(init_cfg) |
| | if component1 is not None: |
| | self.component1 = build_from_cfg(component1, COMPONENTS) |
| | if component2 is not None: |
| | self.component2 = build_from_cfg(component2, COMPONENTS) |
| | if component3 is not None: |
| | self.component3 = build_from_cfg(component3, COMPONENTS) |
| | if component4 is not None: |
| | self.component4 = build_from_cfg(component4, COMPONENTS) |
| |
|
| | |
| | |
| | self.reg = nn.Linear(3, 4) |
| |
|
| |
|
| | class TestBaseModule(TestCase): |
| |
|
| | def setUp(self) -> None: |
| | self.BaseModule = BaseModule() |
| | self.model_cfg = dict( |
| | type='FooModel', |
| | init_cfg=[ |
| | dict(type='Constant', val=1, bias=2, layer='Linear'), |
| | dict(type='Constant', val=3, bias=4, layer='Conv1d'), |
| | dict(type='Constant', val=5, bias=6, layer='Conv2d') |
| | ], |
| | component1=dict(type='FooConv1d'), |
| | component2=dict(type='FooConv2d'), |
| | component3=dict(type='FooLinear'), |
| | component4=dict( |
| | type='FooLinearConv1d', |
| | linear=dict(type='FooLinear'), |
| | conv1d=dict(type='FooConv1d'))) |
| |
|
| | self.model = build_from_cfg(self.model_cfg, FOOMODELS) |
| | self.logger = MMLogger.get_instance(self._testMethodName) |
| |
|
| | def tearDown(self) -> None: |
| | logging.shutdown() |
| | MMLogger._instance_dict.clear() |
| | return super().tearDown() |
| |
|
| | def test_is_init(self): |
| | assert self.BaseModule.is_init is False |
| |
|
| | def test_init_weights(self): |
| | """ |
| | Config |
| | model (FooModel, Linear: weight=1, bias=2, Conv1d: weight=3, bias=4, |
| | Conv2d: weight=5, bias=6) |
| | ├──component1 (FooConv1d) |
| | ├──component2 (FooConv2d) |
| | ├──component3 (FooLinear) |
| | ├──component4 (FooLinearConv1d) |
| | ├──linear (FooLinear) |
| | ├──conv1d (FooConv1d) |
| | ├──reg (nn.Linear) |
| | Parameters after initialization |
| | model (FooModel) |
| | ├──component1 (FooConv1d, weight=3, bias=4) |
| | ├──component2 (FooConv2d, weight=5, bias=6) |
| | ├──component3 (FooLinear, weight=1, bias=2) |
| | ├──component4 (FooLinearConv1d) |
| | ├──linear (FooLinear, weight=1, bias=2) |
| | ├──conv1d (FooConv1d, weight=3, bias=4) |
| | ├──reg (nn.Linear, weight=1, bias=2) |
| | """ |
| |
|
| | self.model.init_weights() |
| |
|
| | assert torch.equal( |
| | self.model.component1.conv1d.weight, |
| | torch.full(self.model.component1.conv1d.weight.shape, 3.0)) |
| | assert torch.equal( |
| | self.model.component1.conv1d.bias, |
| | torch.full(self.model.component1.conv1d.bias.shape, 4.0)) |
| | assert torch.equal( |
| | self.model.component2.conv2d.weight, |
| | torch.full(self.model.component2.conv2d.weight.shape, 5.0)) |
| | assert torch.equal( |
| | self.model.component2.conv2d.bias, |
| | torch.full(self.model.component2.conv2d.bias.shape, 6.0)) |
| | assert torch.equal( |
| | self.model.component3.linear.weight, |
| | torch.full(self.model.component3.linear.weight.shape, 1.0)) |
| | assert torch.equal( |
| | self.model.component3.linear.bias, |
| | torch.full(self.model.component3.linear.bias.shape, 2.0)) |
| | assert torch.equal( |
| | self.model.component4.linear.linear.weight, |
| | torch.full(self.model.component4.linear.linear.weight.shape, 1.0)) |
| | assert torch.equal( |
| | self.model.component4.linear.linear.bias, |
| | torch.full(self.model.component4.linear.linear.bias.shape, 2.0)) |
| | assert torch.equal( |
| | self.model.component4.conv1d.conv1d.weight, |
| | torch.full(self.model.component4.conv1d.conv1d.weight.shape, 3.0)) |
| | assert torch.equal( |
| | self.model.component4.conv1d.conv1d.bias, |
| | torch.full(self.model.component4.conv1d.conv1d.bias.shape, 4.0)) |
| | assert torch.equal(self.model.reg.weight, |
| | torch.full(self.model.reg.weight.shape, 1.0)) |
| | assert torch.equal(self.model.reg.bias, |
| | torch.full(self.model.reg.bias.shape, 2.0)) |
| |
|
| | def test_dump_init_info(self): |
| | import os |
| | import shutil |
| | dump_dir = 'tests/test_model/test_dump_info' |
| | if not (os.path.exists(dump_dir) and os.path.isdir(dump_dir)): |
| | os.makedirs(dump_dir) |
| | for filename in os.listdir(dump_dir): |
| | file_path = os.path.join(dump_dir, filename) |
| | if os.path.isfile(file_path) or os.path.islink(file_path): |
| | os.unlink(file_path) |
| | elif os.path.isdir(file_path): |
| | shutil.rmtree(file_path) |
| |
|
| | MMLogger.get_instance('logger1') |
| | model1 = build_from_cfg(self.model_cfg, FOOMODELS) |
| | model1.init_weights() |
| | assert len(os.listdir(dump_dir)) == 0 |
| | log_path = os.path.join(dump_dir, 'out.log') |
| | MMLogger.get_instance( |
| | 'logger2', log_file=log_path) |
| | model2 = build_from_cfg(self.model_cfg, FOOMODELS) |
| | model2.init_weights() |
| | assert len(os.listdir(dump_dir)) == 1 |
| | assert os.stat(log_path).st_size != 0 |
| | |
| | |
| | logging.shutdown() |
| | MMLogger._instance_dict.clear() |
| | shutil.rmtree(dump_dir) |
| |
|
| |
|
| | class TestModuleList(TestCase): |
| |
|
| | def test_modulelist_weight_init(self): |
| | models_cfg = [ |
| | dict( |
| | type='FooConv1d', |
| | init_cfg=dict( |
| | type='Constant', layer='Conv1d', val=0., bias=1.)), |
| | dict( |
| | type='FooConv2d', |
| | init_cfg=dict( |
| | type='Constant', layer='Conv2d', val=2., bias=3.)), |
| | ] |
| | layers = [build_from_cfg(cfg, COMPONENTS) for cfg in models_cfg] |
| | modellist = ModuleList(layers) |
| | modellist.init_weights() |
| | self.assertTrue( |
| | torch.equal(modellist[0].conv1d.weight, |
| | torch.full(modellist[0].conv1d.weight.shape, 0.))) |
| | self.assertTrue( |
| | torch.equal(modellist[0].conv1d.bias, |
| | torch.full(modellist[0].conv1d.bias.shape, 1.))) |
| | self.assertTrue( |
| | torch.equal(modellist[1].conv2d.weight, |
| | torch.full(modellist[1].conv2d.weight.shape, 2.))) |
| | self.assertTrue( |
| | torch.equal(modellist[1].conv2d.bias, |
| | torch.full(modellist[1].conv2d.bias.shape, 3.))) |
| | |
| | layers = [build_from_cfg(cfg, COMPONENTS) for cfg in models_cfg] |
| | modellist = ModuleList( |
| | layers, |
| | init_cfg=dict( |
| | type='Constant', layer=['Conv1d', 'Conv2d'], val=4., bias=5.)) |
| | modellist.init_weights() |
| | self.assertTrue( |
| | torch.equal(modellist[0].conv1d.weight, |
| | torch.full(modellist[0].conv1d.weight.shape, 0.))) |
| | self.assertTrue( |
| | torch.equal(modellist[0].conv1d.bias, |
| | torch.full(modellist[0].conv1d.bias.shape, 1.))) |
| | self.assertTrue( |
| | torch.equal(modellist[1].conv2d.weight, |
| | torch.full(modellist[1].conv2d.weight.shape, 2.))) |
| | self.assertTrue( |
| | torch.equal(modellist[1].conv2d.bias, |
| | torch.full(modellist[1].conv2d.bias.shape, 3.))) |
| |
|
| |
|
| | class TestModuleDict(TestCase): |
| |
|
| | def test_moduledict_weight_init(self): |
| | models_cfg = dict( |
| | foo_conv_1d=dict( |
| | type='FooConv1d', |
| | init_cfg=dict( |
| | type='Constant', layer='Conv1d', val=0., bias=1.)), |
| | foo_conv_2d=dict( |
| | type='FooConv2d', |
| | init_cfg=dict( |
| | type='Constant', layer='Conv2d', val=2., bias=3.)), |
| | ) |
| | layers = { |
| | name: build_from_cfg(cfg, COMPONENTS) |
| | for name, cfg in models_cfg.items() |
| | } |
| | modeldict = ModuleDict(layers) |
| | modeldict.init_weights() |
| | self.assertTrue( |
| | torch.equal( |
| | modeldict['foo_conv_1d'].conv1d.weight, |
| | torch.full(modeldict['foo_conv_1d'].conv1d.weight.shape, 0.))) |
| | self.assertTrue( |
| | torch.equal( |
| | modeldict['foo_conv_1d'].conv1d.bias, |
| | torch.full(modeldict['foo_conv_1d'].conv1d.bias.shape, 1.))) |
| | self.assertTrue( |
| | torch.equal( |
| | modeldict['foo_conv_2d'].conv2d.weight, |
| | torch.full(modeldict['foo_conv_2d'].conv2d.weight.shape, 2.))) |
| | self.assertTrue( |
| | torch.equal( |
| | modeldict['foo_conv_2d'].conv2d.bias, |
| | torch.full(modeldict['foo_conv_2d'].conv2d.bias.shape, 3.))) |
| | |
| | layers = { |
| | name: build_from_cfg(cfg, COMPONENTS) |
| | for name, cfg in models_cfg.items() |
| | } |
| | modeldict = ModuleDict( |
| | layers, |
| | init_cfg=dict( |
| | type='Constant', layer=['Conv1d', 'Conv2d'], val=4., bias=5.)) |
| | modeldict.init_weights() |
| | self.assertTrue( |
| | torch.equal( |
| | modeldict['foo_conv_1d'].conv1d.weight, |
| | torch.full(modeldict['foo_conv_1d'].conv1d.weight.shape, 0.))) |
| | self.assertTrue( |
| | torch.equal( |
| | modeldict['foo_conv_1d'].conv1d.bias, |
| | torch.full(modeldict['foo_conv_1d'].conv1d.bias.shape, 1.))) |
| | self.assertTrue( |
| | torch.equal( |
| | modeldict['foo_conv_2d'].conv2d.weight, |
| | torch.full(modeldict['foo_conv_2d'].conv2d.weight.shape, 2.))) |
| | self.assertTrue( |
| | torch.equal( |
| | modeldict['foo_conv_2d'].conv2d.bias, |
| | torch.full(modeldict['foo_conv_2d'].conv2d.bias.shape, 3.))) |
| |
|
| |
|
| | class TestSequential(TestCase): |
| |
|
| | def test_sequential_model_weight_init(self): |
| | seq_model_cfg = [ |
| | dict( |
| | type='FooConv1d', |
| | init_cfg=dict( |
| | type='Constant', layer='Conv1d', val=0., bias=1.)), |
| | dict( |
| | type='FooConv2d', |
| | init_cfg=dict( |
| | type='Constant', layer='Conv2d', val=2., bias=3.)), |
| | ] |
| | layers = [build_from_cfg(cfg, COMPONENTS) for cfg in seq_model_cfg] |
| | seq_model = Sequential(*layers) |
| | seq_model.init_weights() |
| | self.assertTrue( |
| | torch.equal(seq_model[0].conv1d.weight, |
| | torch.full(seq_model[0].conv1d.weight.shape, 0.))) |
| | self.assertTrue( |
| | torch.equal(seq_model[0].conv1d.bias, |
| | torch.full(seq_model[0].conv1d.bias.shape, 1.))) |
| | self.assertTrue( |
| | torch.equal(seq_model[1].conv2d.weight, |
| | torch.full(seq_model[1].conv2d.weight.shape, 2.))) |
| | self.assertTrue( |
| | torch.equal(seq_model[1].conv2d.bias, |
| | torch.full(seq_model[1].conv2d.bias.shape, 3.))) |
| | |
| | layers = [build_from_cfg(cfg, COMPONENTS) for cfg in seq_model_cfg] |
| | seq_model = Sequential( |
| | *layers, |
| | init_cfg=dict( |
| | type='Constant', layer=['Conv1d', 'Conv2d'], val=4., bias=5.)) |
| | seq_model.init_weights() |
| | self.assertTrue( |
| | torch.equal(seq_model[0].conv1d.weight, |
| | torch.full(seq_model[0].conv1d.weight.shape, 0.))) |
| | self.assertTrue( |
| | torch.equal(seq_model[0].conv1d.bias, |
| | torch.full(seq_model[0].conv1d.bias.shape, 1.))) |
| | self.assertTrue( |
| | torch.equal(seq_model[1].conv2d.weight, |
| | torch.full(seq_model[1].conv2d.weight.shape, 2.))) |
| | self.assertTrue( |
| | torch.equal(seq_model[1].conv2d.bias, |
| | torch.full(seq_model[1].conv2d.bias.shape, 3.))) |
| |
|