| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | import os |
| |
|
| | import hydra |
| | from fvcore.nn import ActivationCountAnalysis, FlopCountAnalysis, flop_count_table |
| | from omegaconf import DictConfig |
| |
|
| | import torch |
| | from torch.utils import benchmark |
| |
|
| |
|
| | @torch.inference_mode() |
| | @hydra.main(config_path='configs', config_name='bench', version_base='1.2') |
| | def main(config: DictConfig): |
| | |
| | os.environ['CUBLAS_WORKSPACE_CONFIG'] = ':4096:8' |
| | torch.backends.cudnn.benchmark = False |
| | torch.use_deterministic_algorithms(True) |
| |
|
| | device = config.get('device', 'cuda') |
| |
|
| | h, w = config.data.img_size |
| | x = torch.rand(1, 3, h, w, device=device) |
| | model = hydra.utils.instantiate(config.model).eval().to(device) |
| |
|
| | if config.get('range', False): |
| | for i in range(1, 26, 4): |
| | timer = benchmark.Timer(stmt='model(x, len)', globals={'model': model, 'x': x, 'len': i}) |
| | print(timer.blocked_autorange(min_run_time=1)) |
| | else: |
| | timer = benchmark.Timer(stmt='model(x)', globals={'model': model, 'x': x}) |
| | flops = FlopCountAnalysis(model, x) |
| | acts = ActivationCountAnalysis(model, x) |
| | print(timer.blocked_autorange(min_run_time=1)) |
| | print(flop_count_table(flops, 1, acts, False)) |
| |
|
| |
|
| | if __name__ == '__main__': |
| | main() |
| |
|