| |
| import pytest |
| import torch |
|
|
| from mmpose.models.backbones.hrformer import (HRFomerModule, HRFormer, |
| HRFormerBlock) |
|
|
|
|
| def test_hrformer_module(): |
| norm_cfg = dict(type='BN') |
| block = HRFormerBlock |
| |
| num_channles = (32, 64) |
| num_inchannels = [c * block.expansion for c in num_channles] |
| hrmodule = HRFomerModule( |
| num_branches=2, |
| block=block, |
| num_blocks=(2, 2), |
| num_inchannels=num_inchannels, |
| num_channels=num_channles, |
| num_heads=(1, 2), |
| num_window_sizes=(7, 7), |
| num_mlp_ratios=(4, 4), |
| drop_paths=(0., 0.), |
| norm_cfg=norm_cfg) |
|
|
| feats = [ |
| torch.randn(1, num_inchannels[0], 64, 64), |
| torch.randn(1, num_inchannels[1], 32, 32) |
| ] |
| feats = hrmodule(feats) |
|
|
| assert len(str(hrmodule)) > 0 |
| assert len(feats) == 2 |
| assert feats[0].shape == torch.Size([1, num_inchannels[0], 64, 64]) |
| assert feats[1].shape == torch.Size([1, num_inchannels[1], 32, 32]) |
|
|
| |
| num_channles = (32, 64) |
| in_channels = [c * block.expansion for c in num_channles] |
| hrmodule = HRFomerModule( |
| num_branches=2, |
| block=block, |
| num_blocks=(2, 2), |
| num_inchannels=num_inchannels, |
| num_channels=num_channles, |
| num_heads=(1, 2), |
| num_window_sizes=(7, 7), |
| num_mlp_ratios=(4, 4), |
| drop_paths=(0., 0.), |
| norm_cfg=norm_cfg, |
| multiscale_output=False, |
| ) |
|
|
| feats = [ |
| torch.randn(1, in_channels[0], 64, 64), |
| torch.randn(1, in_channels[1], 32, 32) |
| ] |
| feats = hrmodule(feats) |
|
|
| assert len(feats) == 1 |
| assert feats[0].shape == torch.Size([1, in_channels[0], 64, 64]) |
|
|
| |
| hrmodule = HRFomerModule( |
| num_branches=1, |
| block=block, |
| num_blocks=(1, ), |
| num_inchannels=[num_inchannels[0]], |
| num_channels=[num_channles[0]], |
| num_heads=(1, ), |
| num_window_sizes=(7, ), |
| num_mlp_ratios=(4, ), |
| drop_paths=(0.1, ), |
| norm_cfg=norm_cfg, |
| ) |
|
|
| feats = [ |
| torch.randn(1, in_channels[0], 64, 64), |
| ] |
| feats = hrmodule(feats) |
|
|
| assert len(feats) == 1 |
| assert feats[0].shape == torch.Size([1, in_channels[0], 64, 64]) |
|
|
| |
| kwargs = dict( |
| num_branches=2, |
| block=block, |
| num_blocks=(2, 2), |
| num_inchannels=num_inchannels, |
| num_channels=num_channles, |
| num_heads=(1, 2), |
| num_window_sizes=(7, 7), |
| num_mlp_ratios=(4, 4), |
| drop_paths=(0.1, 0.1), |
| norm_cfg=norm_cfg, |
| ) |
|
|
| with pytest.raises(ValueError): |
| |
| kwargs['num_blocks'] = [2, 2, 2] |
| HRFomerModule(**kwargs) |
| kwargs['num_blocks'] = [2, 2] |
|
|
| with pytest.raises(ValueError): |
| |
| kwargs['num_channels'] = [2] |
| HRFomerModule(**kwargs) |
| kwargs['num_channels'] = [2, 2] |
|
|
| with pytest.raises(ValueError): |
| |
| kwargs['num_inchannels'] = [2] |
| HRFomerModule(**kwargs) |
| kwargs['num_inchannels'] = [2, 2] |
|
|
|
|
| def test_hrformer_backbone(): |
| norm_cfg = dict(type='BN') |
| |
| extra = dict( |
| drop_path_rate=0.2, |
| stage1=dict( |
| num_modules=1, |
| num_branches=1, |
| block='BOTTLENECK', |
| num_blocks=(2, ), |
| num_channels=(64, )), |
| stage2=dict( |
| num_modules=1, |
| num_branches=2, |
| block='HRFORMERBLOCK', |
| window_sizes=(7, 7), |
| num_heads=(1, 2), |
| mlp_ratios=(4, 4), |
| num_blocks=(2, 2), |
| num_channels=(32, 64)), |
| stage3=dict( |
| num_modules=4, |
| num_branches=3, |
| block='HRFORMERBLOCK', |
| window_sizes=(7, 7, 7), |
| num_heads=(1, 2, 4), |
| mlp_ratios=(4, 4, 4), |
| num_blocks=(2, 2, 2), |
| num_channels=(32, 64, 128)), |
| stage4=dict( |
| num_modules=3, |
| num_branches=4, |
| block='HRFORMERBLOCK', |
| window_sizes=(7, 7, 7, 7), |
| num_heads=(1, 2, 4, 8), |
| mlp_ratios=(4, 4, 4, 4), |
| num_blocks=(2, 2, 2, 2), |
| num_channels=(32, 64, 128, 256), |
| multiscale_output=True)) |
|
|
| with pytest.raises(ValueError): |
| |
| extra['stage4']['num_branches'] = 3 |
| HRFormer(extra=extra) |
| extra['stage4']['num_branches'] = 4 |
|
|
| |
| model = HRFormer(extra=extra, norm_cfg=norm_cfg) |
| model.init_weights() |
| model.train() |
|
|
| imgs = torch.randn(1, 3, 64, 64) |
| feats = model(imgs) |
| assert len(feats) == 4 |
| assert feats[0].shape == torch.Size([1, 32, 16, 16]) |
| assert feats[3].shape == torch.Size([1, 256, 2, 2]) |
|
|
| |
| |
| extra['stage4']['multiscale_output'] = False |
| extra['with_rpe'] = False |
| model = HRFormer(extra=extra, norm_cfg=norm_cfg) |
| model.init_weights() |
| model.train() |
|
|
| imgs = torch.randn(1, 3, 64, 64) |
| feats = model(imgs) |
| assert len(feats) == 1 |
| assert feats[0].shape == torch.Size([1, 32, 16, 16]) |
|
|