DBNet / DB /training /learning_rate.py
fasdfsa's picture
add DB code
52a9452
from bisect import bisect_right
import numpy as np
import torch.optim.lr_scheduler as lr_scheduler
from concern.config import Configurable, State
from concern.signal_monitor import SignalMonitor
class ConstantLearningRate(Configurable):
lr = State(default=0.0001)
def __init__(self, **kwargs):
self.load_all(**kwargs)
def get_learning_rate(self, epoch, step):
return self.lr
class FileMonitorLearningRate(Configurable):
file_path = State()
def __init__(self, **kwargs):
self.load_all(**kwargs)
self.monitor = SignalMonitor(self.file_path)
def get_learning_rate(self, epoch, step):
signal = self.monitor.get_signal()
if signal is not None:
return float(signal)
return None
class PriorityLearningRate(Configurable):
learning_rates = State()
def __init__(self, **kwargs):
self.load_all(**kwargs)
def get_learning_rate(self, epoch, step):
for learning_rate in self.learning_rates:
lr = learning_rate.get_learning_rate(epoch, step)
if lr is not None:
return lr
return None
class MultiStepLR(Configurable):
lr = State()
milestones = State(default=[]) # milestones must be sorted
gamma = State(default=0.1)
def __init__(self, cmd={}, **kwargs):
self.load_all(**kwargs)
self.lr = cmd.get('lr', self.lr)
def get_learning_rate(self, epoch, step):
return self.lr * self.gamma ** bisect_right(self.milestones, epoch)
class WarmupLR(Configurable):
steps = State(default=4000)
warmup_lr = State(default=1e-5)
origin_lr = State()
def __init__(self, cmd={}, **kwargs):
self.load_all(**kwargs)
def get_learning_rate(self, epoch, step):
if epoch == 0 and step < self.steps:
return self.warmup_lr
return self.origin_lr.get_learning_rate(epoch, step)
class PiecewiseConstantLearningRate(Configurable):
boundaries = State(default=[10000, 20000])
values = State(default=[0.001, 0.0001, 0.00001])
def __init__(self, **kwargs):
self.load_all(**kwargs)
def get_learning_rate(self, epoch, step):
for boundary, value in zip(self.boundaries, self.values[:-1]):
if step < boundary:
return value
return self.values[-1]
class DecayLearningRate(Configurable):
lr = State(default=0.007)
epochs = State(default=1200)
factor = State(default=0.9)
def __init__(self, **kwargs):
self.load_all(**kwargs)
def get_learning_rate(self, epoch, step=None):
rate = np.power(1.0 - epoch / float(self.epochs + 1), self.factor)
return rate * self.lr
class BuitlinLearningRate(Configurable):
lr = State(default=0.001)
klass = State(default='StepLR')
args = State(default=[])
kwargs = State(default={})
def __init__(self, cmd={}, **kwargs):
self.load_all(**kwargs)
self.lr = cmd.get('lr', None) or self.lr
self.scheduler = None
def prepare(self, optimizer):
self.scheduler = getattr(lr_scheduler, self.klass)(
optimizer, *self.args, **self.kwargs)
def get_learning_rate(self, epoch, step=None):
if self.scheduler is None:
raise 'learning rate not ready(prepared with optimizer) '
self.scheduler.last_epoch = epoch
# return value of gt_lr is a list,
# where each element is the corresponding learning rate for a
# paramater group.
return self.scheduler.get_lr()[0]