File size: 468 Bytes
c94c8c9
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
import torch.optim as optim

from fvcore.common.registry import Registry
OPTIM_REGISTRY = Registry("loss")

from common.type_utils import cfg2dict


def get_optimizer(cfg, params):
  if getattr(optim, cfg.solver.optim.name, None) is not None:
    optimizer = getattr(optim, cfg.solver.optim.name)(params, **cfg2dict(cfg.solver.optim.args))
  else:
    optimizer = OPTIM_REGISTRY.get(cfg.solver.optim.name)(params, **cfg2dict(cfg.solver.optim.args))
  return optimizer