Buckets:
Overview
8-bit optimizers reduce the memory footprint of 32-bit optimizers without any performance degradation which means you can train large models with many parameters faster. At the core of 8-bit optimizers is block-wise quantization which enables quantization accuracy, computational efficiency, and stability.
bitsandbytes provides 8-bit optimizers through the base Optimizer8bit class, and additionally provides Optimizer2State and Optimizer1State for 2-state (for example, Adam) and 1-state (for example, Adagrad) optimizers respectively. To provide custom optimizer hyperparameters, use the GlobalOptimManager class to configure the optimizer.
Optimizer8bit[[bitsandbytes.optim.optimizer.Optimizer8bit]]
bitsandbytes.optim.optimizer.Optimizer8bit[[bitsandbytes.optim.optimizer.Optimizer8bit]]
__init__bitsandbytes.optim.optimizer.Optimizer8bit.__init__https://github.com/bitsandbytes-foundation/bitsandbytes/blob/main/bitsandbytes/optim/optimizer.py#L120[{"name": "params", "val": ""}, {"name": "defaults", "val": ""}, {"name": "optim_bits", "val": " = 32"}, {"name": "is_paged", "val": " = False"}]- params (torch.Tensor) --
The input parameters to optimize.
- optim_bits (
int, defaults to 32) -- The number of bits of the optimizer state. - is_paged (
bool, defaults toFalse) -- Whether the optimizer is a paged optimizer or not.0
Base 8-bit optimizer class.
Parameters:
params (torch.Tensor) : The input parameters to optimize.
optim_bits (int, defaults to 32) : The number of bits of the optimizer state.
is_paged (bool, defaults to False) : Whether the optimizer is a paged optimizer or not.
Optimizer2State[[bitsandbytes.optim.optimizer.Optimizer2State]]
bitsandbytes.optim.optimizer.Optimizer2State[[bitsandbytes.optim.optimizer.Optimizer2State]]
__init__bitsandbytes.optim.optimizer.Optimizer2State.__init__https://github.com/bitsandbytes-foundation/bitsandbytes/blob/main/bitsandbytes/optim/optimizer.py#L404[{"name": "optimizer_name", "val": ""}, {"name": "params", "val": ""}, {"name": "lr", "val": " = 0.001"}, {"name": "betas", "val": " = (0.9, 0.999)"}, {"name": "eps", "val": " = 1e-08"}, {"name": "weight_decay", "val": " = 0.0"}, {"name": "optim_bits", "val": " = 32"}, {"name": "args", "val": " = None"}, {"name": "min_8bit_size", "val": " = 4096"}, {"name": "max_unorm", "val": " = 0.0"}, {"name": "skip_zeros", "val": " = False"}, {"name": "is_paged", "val": " = False"}, {"name": "alpha", "val": " = 0.0"}, {"name": "t_alpha", "val": ": typing.Optional[int] = None"}, {"name": "t_beta3", "val": ": typing.Optional[int] = None"}]- optimizer_name (str) --
The name of the optimizer.
- params (
torch.Tensor) -- The input parameters to optimize. - lr (
float, defaults to 1e-3) -- The learning rate. - betas (
tuple, defaults to (0.9, 0.999)) -- The beta values for the optimizer. - eps (
float, defaults to 1e-8) -- The epsilon value for the optimizer. - weight_decay (
float, defaults to 0.0) -- The weight decay value for the optimizer. - optim_bits (
int, defaults to 32) -- The number of bits of the optimizer state. - args (
object, defaults toNone) -- An object with additional arguments. - min_8bit_size (
int, defaults to 4096) -- The minimum number of elements of the parameter tensors for 8-bit optimization. - max_unorm (
float, defaults to 0.0) -- The maximum value to normalize each block with. - skip_zeros (
bool, defaults toFalse) -- Whether to skip zero values for sparse gradients and models to ensure correct updates. - is_paged (
bool, defaults toFalse) -- Whether the optimizer is a paged optimizer or not. - alpha (
float, defaults to 0.0) -- The alpha value for the AdEMAMix optimizer. - t_alpha (
Optional[int], defaults toNone) -- Number of iterations for alpha scheduling with AdEMAMix. - t_beta3 (
Optional[int], defaults toNone) -- Number of iterations for beta scheduling with AdEMAMix.0
Base 2-state update optimizer class.
Parameters:
optimizer_name (str) : The name of the optimizer.
params (torch.Tensor) : The input parameters to optimize.
lr (float, defaults to 1e-3) : The learning rate.
betas (tuple, defaults to (0.9, 0.999)) : The beta values for the optimizer.
eps (float, defaults to 1e-8) : The epsilon value for the optimizer.
weight_decay (float, defaults to 0.0) : The weight decay value for the optimizer.
optim_bits (int, defaults to 32) : The number of bits of the optimizer state.
args (object, defaults to None) : An object with additional arguments.
min_8bit_size (int, defaults to 4096) : The minimum number of elements of the parameter tensors for 8-bit optimization.
max_unorm (float, defaults to 0.0) : The maximum value to normalize each block with.
skip_zeros (bool, defaults to False) : Whether to skip zero values for sparse gradients and models to ensure correct updates.
is_paged (bool, defaults to False) : Whether the optimizer is a paged optimizer or not.
alpha (float, defaults to 0.0) : The alpha value for the AdEMAMix optimizer.
t_alpha (Optional[int], defaults to None) : Number of iterations for alpha scheduling with AdEMAMix.
t_beta3 (Optional[int], defaults to None) : Number of iterations for beta scheduling with AdEMAMix.
Optimizer1State[[bitsandbytes.optim.optimizer.Optimizer1State]]
bitsandbytes.optim.optimizer.Optimizer1State[[bitsandbytes.optim.optimizer.Optimizer1State]]
__init__bitsandbytes.optim.optimizer.Optimizer1State.__init__https://github.com/bitsandbytes-foundation/bitsandbytes/blob/main/bitsandbytes/optim/optimizer.py#L594[{"name": "optimizer_name", "val": ""}, {"name": "params", "val": ""}, {"name": "lr", "val": " = 0.001"}, {"name": "betas", "val": " = (0.9, 0.0)"}, {"name": "eps", "val": " = 1e-08"}, {"name": "weight_decay", "val": " = 0.0"}, {"name": "optim_bits", "val": " = 32"}, {"name": "args", "val": " = None"}, {"name": "min_8bit_size", "val": " = 4096"}, {"name": "max_unorm", "val": " = 0.0"}, {"name": "skip_zeros", "val": " = False"}, {"name": "is_paged", "val": " = False"}]- optimizer_name (str) --
The name of the optimizer.
- params (
torch.Tensor) -- The input parameters to optimize. - lr (
float, defaults to 1e-3) -- The learning rate. - betas (
tuple, defaults to (0.9, 0.0)) -- The beta values for the optimizer. - eps (
float, defaults to 1e-8) -- The epsilon value for the optimizer. - weight_decay (
float, defaults to 0.0) -- The weight decay value for the optimizer. - optim_bits (
int, defaults to 32) -- The number of bits of the optimizer state. - args (
object, defaults toNone) -- An object with additional arguments. - min_8bit_size (
int, defaults to 4096) -- The minimum number of elements of the parameter tensors for 8-bit optimization. - max_unorm (
float, defaults to 0.0) -- The maximum value to normalize each block with. - skip_zeros (
bool, defaults toFalse) -- Whether to skip zero values for sparse gradients and models to ensure correct updates. - is_paged (
bool, defaults toFalse) -- Whether the optimizer is a paged optimizer or not.0
Base 1-state update optimizer class.
Parameters:
optimizer_name (str) : The name of the optimizer.
params (torch.Tensor) : The input parameters to optimize.
lr (float, defaults to 1e-3) : The learning rate.
betas (tuple, defaults to (0.9, 0.0)) : The beta values for the optimizer.
eps (float, defaults to 1e-8) : The epsilon value for the optimizer.
weight_decay (float, defaults to 0.0) : The weight decay value for the optimizer.
optim_bits (int, defaults to 32) : The number of bits of the optimizer state.
args (object, defaults to None) : An object with additional arguments.
min_8bit_size (int, defaults to 4096) : The minimum number of elements of the parameter tensors for 8-bit optimization.
max_unorm (float, defaults to 0.0) : The maximum value to normalize each block with.
skip_zeros (bool, defaults to False) : Whether to skip zero values for sparse gradients and models to ensure correct updates.
is_paged (bool, defaults to False) : Whether the optimizer is a paged optimizer or not.
Utilities[[bitsandbytes.optim.GlobalOptimManager]]
bitsandbytes.optim.GlobalOptimManager[[bitsandbytes.optim.GlobalOptimManager]]
A global optimizer manager for enabling custom optimizer configs.
override_configbitsandbytes.optim.GlobalOptimManager.override_confighttps://github.com/bitsandbytes-foundation/bitsandbytes/blob/main/bitsandbytes/optim/optimizer.py#L60[{"name": "parameters", "val": ""}, {"name": "key", "val": " = None"}, {"name": "value", "val": " = None"}, {"name": "key_value_dict", "val": " = None"}]- parameters (torch.Tensor or list(torch.Tensors)) --
The input parameters.
- key (
str) -- The hyperparameter to override. - value -- The hyperparameter value.
- key_value_dict (
dict) -- A dictionary with multiple key-values to override.0
Override initial optimizer config with specific hyperparameters.
The key-values of the optimizer config for the input parameters are overridden
This can be both, optimizer parameters like betas or lr, or it can be
8-bit specific parameters like optim_bits.
Example:
import torch
import bitsandbytes as bnb
mng = bnb.optim.GlobalOptimManager.get_instance()
model = MyModel()
mng.register_parameters(model.parameters()) # 1. register parameters while still on CPU
model = model.cuda()
# use 8-bit optimizer states for all parameters
adam = bnb.optim.Adam(model.parameters(), lr=0.001, optim_bits=8)
# 2. override: the parameter model.fc1.weight now uses 32-bit Adam
mng.override_config(model.fc1.weight, 'optim_bits', 32)
Parameters:
parameters (torch.Tensor or list(torch.Tensors)) : The input parameters.
key (str) : The hyperparameter to override.
value : The hyperparameter value.
key_value_dict (dict) : A dictionary with multiple key-values to override.
Xet Storage Details
- Size:
- 10.6 kB
- Xet hash:
- dc28329447503d505cfd8c01d456b2ea7394d98e80b6f4ad40e2bb4d48a4a938
Xet efficiently stores files, intelligently splitting them into unique chunks and accelerating uploads and downloads. More info.