File size: 949 Bytes
c94c8c9 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 |
import math
from torch.optim.lr_scheduler import LambdaLR
def warmup_cosine(step, warmup_step, total_step, minimum_ratio=1e-5):
if step <= warmup_step and warmup_step > 0:
return step / warmup_step
return max(
0.5 * (1 + math.cos((step - warmup_step) / (total_step - warmup_step) * math.pi)),
minimum_ratio
)
def warmup_exp(step, warmup_step, total_step, **kwargs):
if step <= warmup_step and warmup_step > 0:
return step / warmup_step
return kwargs["gamma"] ** (step * 1. / (total_step - warmup_step))
def get_scheduler(cfg, optimizer, total_steps):
warmup_steps = cfg.solver.sched.args.warmup_steps * cfg.num_gpu
minimum_ratio = cfg.solver.sched.args.get("minimum_ratio", 1e-5)
lambda_func = lambda step: globals()[cfg.solver.sched.name](
step, warmup_steps, total_steps, minimum_ratio=minimum_ratio
)
return LambdaLR(optimizer=optimizer, lr_lambda=lambda_func)
|