Buckets:

|
download
raw
5.88 kB

4-bit quantization

QLoRA is a finetuning method that quantizes a model to 4-bits and adds a set of low-rank adaptation (LoRA) weights to the model and tuning them through the quantized weights. This method also introduces a new data type, 4-bit NormalFloat (LinearNF4) in addition to the standard Float4 data type (LinearFP4). LinearNF4 is a quantization data type for normally distributed data and can improve performance.

Linear4bit[[bitsandbytes.nn.Linear4bit]]

bitsandbytes.nn.Linear4bit[[bitsandbytes.nn.Linear4bit]]

Source

This class is the base module for the 4-bit quantization algorithm presented in QLoRA. QLoRA 4-bit linear layers uses blockwise k-bit quantization under the hood, with the possibility of selecting various compute datatypes such as FP4 and NF4.

In order to quantize a linear layer one should first load the original fp16 / bf16 weights into the Linear4bit module, then call quantized_module.to("cuda") to quantize the fp16 / bf16 weights.

Example:

import torch
import torch.nn as nn

import bitsandbytes as bnb
from bitsandbytes.nn import Linear4bit

fp16_model = nn.Sequential(
    nn.Linear(64, 64),
    nn.Linear(64, 64)
)

quantized_model = nn.Sequential(
    Linear4bit(64, 64),
    Linear4bit(64, 64)
)

quantized_model.load_state_dict(fp16_model.state_dict())
quantized_model = quantized_model.to(0) # Quantization happens here

__init__bitsandbytes.nn.Linear4bit.__init__https://github.com/bitsandbytes-foundation/bitsandbytes/blob/main/bitsandbytes/nn/modules.py#L537[{"name": "input_features", "val": ""}, {"name": "output_features", "val": ""}, {"name": "bias", "val": " = True"}, {"name": "compute_dtype", "val": " = None"}, {"name": "compress_statistics", "val": " = True"}, {"name": "quant_type", "val": " = 'fp4'"}, {"name": "quant_storage", "val": " = torch.uint8"}, {"name": "device", "val": " = None"}]- input_features (str) -- Number of input features of the linear layer.

  • output_features (str) -- Number of output features of the linear layer.
  • bias (bool, defaults to True) -- Whether the linear class uses the bias term as well.0

Initialize Linear4bit class.

Parameters:

input_features (str) : Number of input features of the linear layer.

output_features (str) : Number of output features of the linear layer.

bias (bool, defaults to True) : Whether the linear class uses the bias term as well.

LinearFP4[[bitsandbytes.nn.LinearFP4]]

bitsandbytes.nn.LinearFP4[[bitsandbytes.nn.LinearFP4]]

Source

Implements the FP4 data type.

__init__bitsandbytes.nn.LinearFP4.__init__https://github.com/bitsandbytes-foundation/bitsandbytes/blob/main/bitsandbytes/nn/modules.py#L645[{"name": "input_features", "val": ""}, {"name": "output_features", "val": ""}, {"name": "bias", "val": " = True"}, {"name": "compute_dtype", "val": " = None"}, {"name": "compress_statistics", "val": " = True"}, {"name": "quant_storage", "val": " = torch.uint8"}, {"name": "device", "val": " = None"}]- input_features (str) -- Number of input features of the linear layer.

  • output_features (str) -- Number of output features of the linear layer.
  • bias (bool, defaults to True) -- Whether the linear class uses the bias term as well.0

Parameters:

input_features (str) : Number of input features of the linear layer.

output_features (str) : Number of output features of the linear layer.

bias (bool, defaults to True) : Whether the linear class uses the bias term as well.

LinearNF4[[bitsandbytes.nn.LinearNF4]]

bitsandbytes.nn.LinearNF4[[bitsandbytes.nn.LinearNF4]]

Source

Implements the NF4 data type.

Constructs a quantization data type where each bin has equal area under a standard normal distribution N(0, 1) that is normalized into the range [-1, 1].

For more information read the paper: QLoRA: Efficient Finetuning of Quantized LLMs (https://arxiv.org/abs/2305.14314)

Implementation of the NF4 data type in bitsandbytes can be found in the create_normal_map function in the functional.py file: https://github.com/TimDettmers/bitsandbytes/blob/main/bitsandbytes/functional.py#L236.

__init__bitsandbytes.nn.LinearNF4.__init__https://github.com/bitsandbytes-foundation/bitsandbytes/blob/main/bitsandbytes/nn/modules.py#L688[{"name": "input_features", "val": ""}, {"name": "output_features", "val": ""}, {"name": "bias", "val": " = True"}, {"name": "compute_dtype", "val": " = None"}, {"name": "compress_statistics", "val": " = True"}, {"name": "quant_storage", "val": " = torch.uint8"}, {"name": "device", "val": " = None"}]- input_features (str) -- Number of input features of the linear layer.

  • output_features (str) -- Number of output features of the linear layer.
  • bias (bool, defaults to True) -- Whether the linear class uses the bias term as well.0

Parameters:

input_features (str) : Number of input features of the linear layer.

output_features (str) : Number of output features of the linear layer.

bias (bool, defaults to True) : Whether the linear class uses the bias term as well.

Params4bit[[bitsandbytes.nn.Params4bit]]

bitsandbytes.nn.Params4bit[[bitsandbytes.nn.Params4bit]]

Source

__init__bitsandbytes.nn.Params4bit.init[{"name": "*args", "val": ""}, {"name": "**kwargs", "val": ""}] Initialize self. See help(type(self)) for accurate signature.

Xet Storage Details

Size:
5.88 kB
·
Xet hash:
9d8c93c883b3de85e5603500e4edaf0204c78fb7710d14570adaa318ddb501ba

Xet efficiently stores files, intelligently splitting them into unique chunks and accelerating uploads and downloads. More info.