ZTWHHH's picture
Add files using upload-large-folder tool
70c471b verified
import math
from typing import Iterable, Literal, Optional, Tuple
import torch
import bitsandbytes.functional as F
from bitsandbytes.optim.optimizer import Optimizer2State
class _ReferenceAdEMAMix(torch.optim.Optimizer):
"""
Reference: https://hf.co/papers/2409.03137
"""
def __init__(
self,
params: Iterable[torch.nn.Parameter],
lr: float = 1e-3,
betas: Tuple[float, float, float] = (0.9, 0.999, 0.9999),
alpha: float = 5.0,
eps: float = 1e-8,
weight_decay: float = 1e-2, # default 0.0 or 1e-2?
t_beta3: Optional[int] = None,
t_alpha: Optional[int] = None,
):
defaults = dict(
lr=lr, betas=betas, alpha=alpha, eps=eps, weight_decay=weight_decay, t_beta3=t_beta3, t_alpha=t_alpha
)
super().__init__(params, defaults)
@torch.no_grad()
def step(self, closure=None):
loss = None
if closure is not None:
with torch.enable_grad():
loss = closure()
for group in self.param_groups:
if "step" in group:
group["step"] += 1
else:
group["step"] = 1
lr = group["lr"]
eps = group["eps"]
beta1, beta2, beta3 = group["betas"]
alpha = group["alpha"]
t_alpha = group["t_alpha"]
t_beta3 = group["t_beta3"]
weight_decay = group["weight_decay"]
for p in group["params"]:
if p.grad is None:
continue
grad = p.grad
state = self.state[p]
# State initialization
if len(state) == 0:
# For parity with bnb implementation we combine both fast
# and slow EMA stats into one stacked tensor.
state["m1_m2"] = p.new_zeros((2, *p.size()))
state["nu"] = torch.zeros_like(p) # second moment estimate
m1, m2, nu = state["m1_m2"][0], state["m1_m2"][1], state["nu"]
bias_correction1 = 1 - beta1 ** group["step"]
bias_correction2 = 1 - beta2 ** group["step"]
# Apply scheduler for alpha
if t_alpha is not None:
alpha = min(group["step"] * alpha / t_alpha, alpha)
# Apply scheduler for beta3
if t_beta3 is not None:
ln_beta1 = math.log(beta1)
ln_beta3 = math.log(beta3)
step_scale = group["step"] / t_beta3
beta3 = min(
math.exp((ln_beta1 * ln_beta3) / (((1 - step_scale) * ln_beta3) + (step_scale * ln_beta1))),
beta3,
)
# Update the EMAs
m1.mul_(beta1).add_(grad, alpha=1 - beta1)
m2.mul_(beta3).add_(grad, alpha=1 - beta3)
nu.mul_(beta2).addcmul_(grad, grad, value=1 - beta2)
# Compute step
denom = (nu.sqrt() / (bias_correction2**0.5)).add(eps)
update = (m1.div(bias_correction1) + alpha * m2) / denom
# Add weight decay
update.add_(p, alpha=weight_decay)
# Apply update scaled by learning rate
p.add_(-lr * update)
return loss
class AdEMAMix(Optimizer2State):
def __init__(
self,
params: Iterable[torch.nn.Parameter],
lr: float = 1e-3,
betas: Tuple[float, float, float] = (0.9, 0.999, 0.9999),
alpha: float = 5.0,
t_alpha: Optional[int] = None,
t_beta3: Optional[int] = None,
eps: float = 1e-8,
weight_decay: float = 1e-2,
optim_bits: Literal[8, 32] = 32,
min_8bit_size: int = 4096,
is_paged: bool = False,
):
super().__init__(
"ademamix",
params=params,
lr=lr,
betas=betas,
eps=eps,
weight_decay=weight_decay,
optim_bits=optim_bits,
args=None,
min_8bit_size=min_8bit_size,
percentile_clipping=100,
block_wise=True,
is_paged=is_paged,
alpha=alpha,
t_alpha=t_alpha,
t_beta3=t_beta3,
)
@torch.no_grad()
def init_state(self, group, p, gindex, pindex):
# In our AdEMAMix implementation, we use `state` to hold
# both the fast and slow EMAs. Here we override the base
# `Optimizer2State` to allocate a buffer twice as large.
# Additional consideration: we do not support block_wise=False,
# percentile clipping, or max_unorm.
config = self.get_config(gindex, pindex, group)
if config["optim_bits"] == 32:
dtype = torch.float32
elif config["optim_bits"] == 8:
dtype = torch.uint8
else:
raise NotImplementedError(f'Amount of optimizer bits not supported: {config["optim_bits"]}')
if p.numel() < config["min_8bit_size"]:
dtype = torch.float32
state = self.state[p]
state["step"] = 0
if dtype == torch.uint8:
if "dynamic" not in self.name2qmap:
self.fill_qmap()
self.name2qmap["dynamic"] = state["qmap1"] = self.name2qmap["dynamic"].to(p.device)
self.name2qmap["udynamic"] = state["qmap2"] = self.name2qmap["udynamic"].to(p.device)
n = p.numel()
blocks = (n // 256) + bool(n % 256)
state["absmax1"] = torch.zeros((2, blocks), dtype=torch.float32, device=p.device)
state["absmax2"] = torch.zeros((blocks,), dtype=torch.float32, device=p.device)
state["state1"] = self._get_state_double_buffer(p, dtype=dtype)
state["state2"] = self.get_state_buffer(p, dtype=dtype)
@torch.no_grad()
def update_step(self, group, p, gindex, pindex):
config = self.get_config(gindex, pindex, group)
if config["t_alpha"] is None and config["t_beta3"] is None:
# Not using alpha/beta3 scheduler; we can fall through.
super().update_step(group, p, gindex, pindex)
return
# Ensure contiguous memory layout
p.data = p.data.contiguous()
p.grad = p.grad.contiguous()
state = self.state[p]
grad = p.grad
state["step"] += 1
step = state["step"]
beta1, beta2, beta3 = config["betas"]
alpha = config["alpha"]
t_alpha = config["t_alpha"]
t_beta3 = config["t_beta3"]
# Apply scheduler for alpha
if t_alpha is not None:
alpha_t = min(step * alpha / t_alpha, alpha)
else:
alpha_t = alpha
# Apply scheduler for beta3
if t_beta3 is not None:
ln_beta1 = math.log(beta1)
ln_beta3 = math.log(beta3)
step_scale = step / t_beta3
beta3_t = min(
math.exp((ln_beta1 * ln_beta3) / (((1 - step_scale) * ln_beta3) + (step_scale * ln_beta1))), beta3
)
else:
beta3_t = beta3
# Apply updates
if state["state1"].dtype == torch.float32:
F.optimizer_update_32bit(
self.optimizer_name,
grad,
p,
state["state1"],
beta1,
config["eps"],
step,
config["lr"],
state["state2"],
beta2,
beta3_t,
alpha_t,
config["weight_decay"],
gnorm_scale=1.0,
unorm_vec=state["unorm_vec"] if config["max_unorm"] > 0.0 else None,
max_unorm=config["max_unorm"],
skip_zeros=config["skip_zeros"],
)
elif state["state1"].dtype == torch.uint8:
F.optimizer_update_8bit_blockwise(
self.optimizer_name,
grad,
p,
state["state1"],
state["state2"],
config["betas"][0],
config["betas"][1],
beta3_t,
alpha_t,
config["eps"],
step,
config["lr"],
state["qmap1"],
state["qmap2"],
state["absmax1"],
state["absmax2"],
config["weight_decay"],
gnorm_scale=1.0,
skip_zeros=config["skip_zeros"],
)
def _get_state_double_buffer(self, p, dtype=torch.float32):
if not self.is_paged or p.numel() < 1e5:
return torch.zeros((2, *p.size()), dtype=dtype, device=p.device)
else:
buff = F.get_paged(*(2, *p.size()), dtype=dtype, device=p.device)
F.fill(buff, 0)
self.page_mng.paged_tensors.append(buff)
return buff
class AdEMAMix8bit(AdEMAMix):
def __init__(
self,
params: Iterable[torch.nn.Parameter],
lr: float = 1e-3,
betas: Tuple[float, float, float] = (0.9, 0.999, 0.9999),
alpha: float = 5.0,
t_alpha: Optional[int] = None,
t_beta3: Optional[int] = None,
eps: float = 1e-8,
weight_decay: float = 1e-2,
min_8bit_size: int = 4096,
is_paged: bool = False,
):
super().__init__(
params,
lr=lr,
betas=betas,
alpha=alpha,
t_alpha=t_alpha,
t_beta3=t_beta3,
eps=eps,
weight_decay=weight_decay,
optim_bits=8,
min_8bit_size=min_8bit_size,
is_paged=is_paged,
)
class PagedAdEMAMix8bit(AdEMAMix8bit):
def __init__(
self,
params: Iterable[torch.nn.Parameter],
lr: float = 1e-3,
betas: Tuple[float, float, float] = (0.9, 0.999, 0.9999),
alpha: float = 5.0,
t_alpha: Optional[int] = None,
t_beta3: Optional[int] = None,
eps: float = 1e-8,
weight_decay: float = 1e-2,
min_8bit_size: int = 4096,
):
super().__init__(
params,
lr=lr,
betas=betas,
alpha=alpha,
t_alpha=t_alpha,
t_beta3=t_beta3,
eps=eps,
weight_decay=weight_decay,
min_8bit_size=min_8bit_size,
is_paged=True,
)
class PagedAdEMAMix(AdEMAMix):
def __init__(
self,
params: Iterable[torch.nn.Parameter],
lr: float = 1e-3,
betas: Tuple[float, float, float] = (0.9, 0.999, 0.9999),
alpha: float = 5.0,
t_alpha: Optional[int] = None,
t_beta3: Optional[int] = None,
eps: float = 1e-8,
weight_decay: float = 1e-2,
optim_bits: Literal[8, 32] = 32,
min_8bit_size: int = 4096,
):
super().__init__(
params,
lr=lr,
betas=betas,
alpha=alpha,
t_alpha=t_alpha,
t_beta3=t_beta3,
eps=eps,
weight_decay=weight_decay,
optim_bits=optim_bits,
min_8bit_size=min_8bit_size,
is_paged=True,
)
class AdEMAMix32bit(Optimizer2State):
def __init__(
self,
params: Iterable[torch.nn.Parameter],
lr: float = 1e-3,
betas: Tuple[float, float, float] = (0.9, 0.999, 0.9999),
alpha: float = 5.0,
t_alpha: Optional[int] = None,
t_beta3: Optional[int] = None,
eps: float = 1e-8,
weight_decay: float = 1e-2,
min_8bit_size: int = 4096,
is_paged: bool = False,
):
super().__init__(
"ademamix",
params=params,
lr=lr,
betas=betas,
eps=eps,
weight_decay=weight_decay,
optim_bits=32,
args=None,
min_8bit_size=min_8bit_size,
percentile_clipping=100,
block_wise=True,
is_paged=is_paged,
alpha=alpha,
t_alpha=t_alpha,
t_beta3=t_beta3,
)
class PagedAdEMAMix32bit(AdEMAMix32bit):
def __init__(
self,
params: Iterable[torch.nn.Parameter],
lr: float = 1e-3,
betas: Tuple[float, float, float] = (0.9, 0.999, 0.9999),
alpha: float = 5.0,
t_alpha: Optional[int] = None,
t_beta3: Optional[int] = None,
eps: float = 1e-8,
weight_decay: float = 1e-2,
min_8bit_size: int = 4096,
):
super().__init__(
params,
lr=lr,
betas=betas,
alpha=alpha,
t_alpha=t_alpha,
t_beta3=t_beta3,
eps=eps,
weight_decay=weight_decay,
min_8bit_size=min_8bit_size,
is_paged=True,
)