Harmony18090's picture
Add source batch 2/11
76f9669 verified
raw
history blame
9.95 kB
# Copyright (c) 2021 - present / Neuralmagic, Inc. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from collections import OrderedDict
from copy import deepcopy
from typing import Dict, List, Optional
from typing import OrderedDict as OrderedDictType
from typing import Union
import torch
from compressed_tensors.config import CompressionFormat
from compressed_tensors.modeling import (
initialize_hooked_attention,
initialize_hooked_kv_cache,
)
from compressed_tensors.quantization.lifecycle.initialize import (
initialize_module_for_quantization,
is_attention_module,
)
from compressed_tensors.quantization.quant_args import QuantizationArgs
from compressed_tensors.quantization.quant_config import (
QuantizationConfig,
QuantizationStatus,
)
from compressed_tensors.quantization.quant_scheme import QuantizationScheme
from compressed_tensors.utils.helpers import replace_module
from compressed_tensors.utils.match import (
is_narrow_match,
match_named_modules,
match_targets,
)
from compressed_tensors.utils.offload import update_parameter_data
from compressed_tensors.utils.safetensors_load import get_safetensors_folder
from loguru import logger
from safetensors import safe_open
from torch.nn import Module
__all__ = [
"load_pretrained_quantization_parameters",
"apply_quantization_config",
]
from compressed_tensors.quantization.utils.helpers import is_module_quantized
from compressed_tensors.utils.safetensors_load import (
get_quantization_parameter_to_path_mapping,
)
def load_pretrained_quantization_parameters(
model: Module,
model_name_or_path: Optional[str] = None,
load_weight_qparams: Optional[bool] = False,
):
"""
Loads the quantization parameters (scale and zero point) from model_name_or_path to
a model that has already been initialized with a quantization config.
NOTE: Will always load inputs/output parameters. Will conditioanlly load weight
parameters, if load_weight_qparams is set to True.
:param model: model to load pretrained quantization parameters to
:param model_name_or_path: Hugging Face stub or local folder containing a quantized
model, which is used to load quantization parameters
:param load_weight_qparams: whether or not the weight quantization parameters
should be loaded
"""
model_path = get_safetensors_folder(model_name_or_path)
mapping = get_quantization_parameter_to_path_mapping(model_path)
for name, submodule in model.named_modules():
if not is_module_quantized(submodule):
continue
if submodule.quantization_scheme.input_activations is not None:
base_name = "input"
_load_quant_args_from_mapping(
base_name=base_name,
module_name=name,
module=submodule,
mapping=mapping,
)
if submodule.quantization_scheme.output_activations is not None:
base_name = "output"
_load_quant_args_from_mapping(
base_name=base_name,
module_name=name,
module=submodule,
mapping=mapping,
)
if load_weight_qparams and submodule.quantization_scheme.weights:
base_name = "weight"
_load_quant_args_from_mapping(
base_name=base_name,
module_name=name,
module=submodule,
mapping=mapping,
)
def apply_quantization_config(
model: Module, config: Union[QuantizationConfig, None], run_compressed: bool = False
):
"""
Initializes the model for quantization in-place based on the given config.
Optionally coverts quantizable modules to compressed_linear modules
:param model: model to apply quantization config to
:param config: quantization config
:param run_compressed: Whether the model will be run in compressed mode or
decompressed fully on load
"""
from compressed_tensors.linear.compressed_linear import CompressedLinear
config = deepcopy(config)
if config is None: # see PR #180
return dict()
# force zero points during initialization
force_zero_point = config.quantization_status != QuantizationStatus.COMPRESSED
# apply and initialize kv cache quantization
if config.kv_cache_scheme is not None:
_apply_kv_cache_scheme(
model, config.kv_cache_scheme, config.quantization_status
)
# build mapping of targets to schemes for easier matching
# use ordered dict to preserve target ordering in config
target_to_scheme = OrderedDict()
for scheme in config.config_groups.values():
for target in scheme.targets:
target_to_scheme[target] = scheme
# mark appropriate layers for quantization by setting their quantization schemes
for name, submodule in match_named_modules(
model, target_to_scheme, config.ignore, warn_on_fail=True
):
# mark modules to be quantized by adding
# quant scheme to the matching layers
matched_targets = match_targets(name, submodule, target_to_scheme)
scheme = _scheme_from_targets(target_to_scheme, matched_targets, name)
# target matched - add layer and scheme to target list
submodule.quantization_scheme = scheme
# replace with run compressed if applicable
# FUTURE: move this to model compressor
if (
run_compressed
and isinstance(submodule, torch.nn.Linear)
and config.format != CompressionFormat.dense.value
):
# TODO: expand to more module types
compressed_linear = CompressedLinear.from_linear(
submodule,
quantization_scheme=scheme,
quantization_format=config.format,
)
replace_module(model, name, compressed_linear)
else:
if is_attention_module(submodule) and is_narrow_match(
model, scheme.targets, name
):
initialize_hooked_attention(model, submodule)
initialize_module_for_quantization(
submodule,
force_zero_point=force_zero_point,
)
submodule.quantization_status = config.quantization_status
def _apply_kv_cache_scheme(
model: torch.nn.Module,
kv_cache_scheme: QuantizationArgs,
status: QuantizationStatus,
):
if not kv_cache_scheme.symmetric:
raise logger.warning("vLLM does not support asymmetric kv cache quantization")
# applies and initializes kv cache quantization
# this step cannot come after attention apply/initialize
# otherwise it will override the attention qparams
scheme = QuantizationScheme(
targets=[".*self_attn$"], # is never read in practice
input_activations=kv_cache_scheme,
)
for submodule in model.modules():
if is_attention_module(submodule):
submodule.quantization_scheme = scheme
initialize_hooked_kv_cache(model, submodule)
initialize_module_for_quantization(submodule, force_zero_point=False)
submodule.quantization_status = status
def _load_quant_args_from_mapping(
base_name: str, module_name: str, module: Module, mapping: Dict
):
# TODO: skip update and just register here, don't do it in initialize
"""
Loads scale and zero point from a state_dict into the specified module
:param base_name: quantization target, one of: weights, input_activations or
output_activations
:param module_name: pytorch module name to look up in state_dict
:module: pytorch module associated with module_name
:mapping: mapping to search fetch paths on disk for a given parameter
"""
scale_name = f"{base_name}_scale"
zp_name = f"{base_name}_zero_point"
g_idx_name = f"{base_name}_g_idx"
state_dict_scale_path = mapping.get(f"{module_name}.{scale_name}", None)
state_dict_zp_path = mapping.get(f"{module_name}.{zp_name}", None)
state_dict_g_idx_path = mapping.get(f"{module_name}.{g_idx_name}", None)
if state_dict_g_idx_path is not None:
with safe_open(state_dict_g_idx_path, framework="pt", device="cpu") as f:
state_dict_g_idx = f.get_tensor(f"{module_name}.{g_idx_name}")
update_parameter_data(module, state_dict_g_idx, g_idx_name)
if state_dict_scale_path is not None:
# module is quantized
with safe_open(state_dict_scale_path, framework="pt", device="cpu") as f:
state_dict_scale = f.get_tensor(f"{module_name}.{scale_name}")
update_parameter_data(module, state_dict_scale, scale_name)
if state_dict_zp_path is None:
# fill in zero point for symmetric quantization
state_dict_zp = torch.zeros_like(state_dict_scale, device="cpu")
else:
with safe_open(state_dict_zp_path, framework="pt", device="cpu") as f:
state_dict_zp = f.get_tensor(f"{module_name}.{zp_name}")
update_parameter_data(module, state_dict_zp, zp_name)
def _scheme_from_targets(
target_to_scheme: OrderedDictType[str, QuantizationScheme],
targets: List[str],
name: str,
) -> QuantizationScheme:
# return the first scheme (the prioritized one,
# since the order of target_to_scheme matters)
return target_to_scheme[targets[0]]