| | |
| | import os, sys |
| | import os.path as osp |
| | import numpy as np |
| | import torch |
| | from torch import nn |
| | from torch.optim import Optimizer |
| | from functools import reduce |
| | from torch.optim import AdamW |
| |
|
| |
|
| | class MultiOptimizer: |
| | def __init__(self, optimizers={}, schedulers={}): |
| | self.optimizers = optimizers |
| | self.schedulers = schedulers |
| | self.keys = list(optimizers.keys()) |
| | self.param_groups = reduce( |
| | lambda x, y: x + y, [v.param_groups for v in self.optimizers.values()] |
| | ) |
| |
|
| | def state_dict(self): |
| | state_dicts = [(key, self.optimizers[key].state_dict()) for key in self.keys] |
| | return state_dicts |
| |
|
| | def load_state_dict(self, state_dict): |
| | for key, val in state_dict: |
| | try: |
| | self.optimizers[key].load_state_dict(val) |
| | except: |
| | print("Unloaded %s" % key) |
| |
|
| | def step(self, key=None, scaler=None): |
| | keys = [key] if key is not None else self.keys |
| | _ = [self._step(key, scaler) for key in keys] |
| |
|
| | def _step(self, key, scaler=None): |
| | if scaler is not None: |
| | scaler.step(self.optimizers[key]) |
| | scaler.update() |
| | else: |
| | self.optimizers[key].step() |
| |
|
| | def zero_grad(self, key=None): |
| | if key is not None: |
| | self.optimizers[key].zero_grad() |
| | else: |
| | _ = [self.optimizers[key].zero_grad() for key in self.keys] |
| |
|
| | def scheduler(self, *args, key=None): |
| | if key is not None: |
| | self.schedulers[key].step(*args) |
| | else: |
| | _ = [self.schedulers[key].step(*args) for key in self.keys] |
| |
|
| |
|
| | def define_scheduler(optimizer, params): |
| | scheduler = torch.optim.lr_scheduler.OneCycleLR( |
| | optimizer, |
| | max_lr=params.get("max_lr", 2e-4), |
| | epochs=params.get("epochs", 200), |
| | steps_per_epoch=params.get("steps_per_epoch", 1000), |
| | pct_start=params.get("pct_start", 0.0), |
| | div_factor=1, |
| | final_div_factor=1, |
| | ) |
| |
|
| | return scheduler |
| |
|
| |
|
| | def build_optimizer(parameters_dict, scheduler_params_dict, lr): |
| | optim = dict( |
| | [ |
| | (key, AdamW(params, lr=lr, weight_decay=1e-4, betas=(0.0, 0.99), eps=1e-9)) |
| | for key, params in parameters_dict.items() |
| | ] |
| | ) |
| |
|
| | schedulers = dict( |
| | [ |
| | (key, define_scheduler(opt, scheduler_params_dict[key])) |
| | for key, opt in optim.items() |
| | ] |
| | ) |
| |
|
| | multi_optim = MultiOptimizer(optim, schedulers) |
| | return multi_optim |
| |
|