|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| import abc
|
| import logging
|
| import math
|
| from dataclasses import asdict, dataclass
|
| from pathlib import Path
|
|
|
| import draccus
|
| from torch.optim import Optimizer
|
| from torch.optim.lr_scheduler import LambdaLR, LRScheduler
|
|
|
| from lerobot.datasets.utils import write_json
|
| from lerobot.utils.constants import SCHEDULER_STATE
|
| from lerobot.utils.io_utils import deserialize_json_into_object
|
|
|
|
|
| @dataclass
|
| class LRSchedulerConfig(draccus.ChoiceRegistry, abc.ABC):
|
| num_warmup_steps: int
|
|
|
| @property
|
| def type(self) -> str:
|
| return self.get_choice_name(self.__class__)
|
|
|
| @abc.abstractmethod
|
| def build(self, optimizer: Optimizer, num_training_steps: int) -> LRScheduler | None:
|
| raise NotImplementedError
|
|
|
|
|
| @LRSchedulerConfig.register_subclass("diffuser")
|
| @dataclass
|
| class DiffuserSchedulerConfig(LRSchedulerConfig):
|
| name: str = "cosine"
|
| num_warmup_steps: int | None = None
|
|
|
| def build(self, optimizer: Optimizer, num_training_steps: int) -> LambdaLR:
|
| from diffusers.optimization import get_scheduler
|
|
|
| kwargs = {**asdict(self), "num_training_steps": num_training_steps, "optimizer": optimizer}
|
| return get_scheduler(**kwargs)
|
|
|
|
|
| @LRSchedulerConfig.register_subclass("vqbet")
|
| @dataclass
|
| class VQBeTSchedulerConfig(LRSchedulerConfig):
|
| num_warmup_steps: int
|
| num_vqvae_training_steps: int
|
| num_cycles: float = 0.5
|
|
|
| def build(self, optimizer: Optimizer, num_training_steps: int) -> LambdaLR:
|
| def lr_lambda(current_step):
|
| if current_step < self.num_vqvae_training_steps:
|
| return float(1)
|
| else:
|
| adjusted_step = current_step - self.num_vqvae_training_steps
|
| if adjusted_step < self.num_warmup_steps:
|
| return float(adjusted_step) / float(max(1, self.num_warmup_steps))
|
| progress = float(adjusted_step - self.num_warmup_steps) / float(
|
| max(1, num_training_steps - self.num_warmup_steps)
|
| )
|
| return max(0.0, 0.5 * (1.0 + math.cos(math.pi * float(self.num_cycles) * 2.0 * progress)))
|
|
|
| return LambdaLR(optimizer, lr_lambda, -1)
|
|
|
|
|
| @LRSchedulerConfig.register_subclass("cosine_decay_with_warmup")
|
| @dataclass
|
| class CosineDecayWithWarmupSchedulerConfig(LRSchedulerConfig):
|
| """Used by Physical Intelligence to train Pi0.
|
|
|
| Automatically scales warmup and decay steps if num_training_steps < num_decay_steps.
|
| This ensures the learning rate schedule completes properly even with shorter training runs.
|
| """
|
|
|
| num_warmup_steps: int
|
| num_decay_steps: int
|
| peak_lr: float
|
| decay_lr: float
|
|
|
| def build(self, optimizer: Optimizer, num_training_steps: int) -> LambdaLR:
|
|
|
| actual_warmup_steps = self.num_warmup_steps
|
| actual_decay_steps = self.num_decay_steps
|
|
|
| if num_training_steps < self.num_decay_steps:
|
|
|
| scale_factor = num_training_steps / self.num_decay_steps
|
| actual_warmup_steps = int(self.num_warmup_steps * scale_factor)
|
| actual_decay_steps = num_training_steps
|
|
|
| logging.info(
|
| f"Auto-scaling LR scheduler: "
|
| f"num_training_steps ({num_training_steps}) < num_decay_steps ({self.num_decay_steps}). "
|
| f"Scaling warmup: {self.num_warmup_steps} → {actual_warmup_steps}, "
|
| f"decay: {self.num_decay_steps} → {actual_decay_steps} "
|
| f"(scale factor: {scale_factor:.3f})"
|
| )
|
|
|
| def lr_lambda(current_step):
|
| def linear_warmup_schedule(current_step):
|
| if current_step <= 0:
|
| return 1 / (actual_warmup_steps + 1)
|
| frac = 1 - current_step / actual_warmup_steps
|
| return (1 / (actual_warmup_steps + 1) - 1) * frac + 1
|
|
|
| def cosine_decay_schedule(current_step):
|
| step = min(current_step, actual_decay_steps)
|
| cosine_decay = 0.5 * (1 + math.cos(math.pi * step / actual_decay_steps))
|
| alpha = self.decay_lr / self.peak_lr
|
| decayed = (1 - alpha) * cosine_decay + alpha
|
| return decayed
|
|
|
| if current_step < actual_warmup_steps:
|
| return linear_warmup_schedule(current_step)
|
|
|
| return cosine_decay_schedule(current_step)
|
|
|
| return LambdaLR(optimizer, lr_lambda, -1)
|
|
|
|
|
| def save_scheduler_state(scheduler: LRScheduler, save_dir: Path) -> None:
|
| state_dict = scheduler.state_dict()
|
| write_json(state_dict, save_dir / SCHEDULER_STATE)
|
|
|
|
|
| def load_scheduler_state(scheduler: LRScheduler, save_dir: Path) -> LRScheduler:
|
| state_dict = deserialize_json_into_object(save_dir / SCHEDULER_STATE, scheduler.state_dict())
|
| scheduler.load_state_dict(state_dict)
|
| return scheduler
|
|
|