File size: 16,187 Bytes
2b6c6b1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
"""NF4 (4-bit NormalFloat) quantization utilities for base model compression.

Parallel to fp8_optimization_utils.py — provides quantize/dequantize, monkey-patching,
and safetensors loading with NF4 quantization.

The NF4 codebook is the set of 16 values that are optimal for normally-distributed
weights (as derived in the QLoRA paper).  Weights are stored as packed uint8 (two
4-bit indices per byte) with per-block absmax scales.
"""

import os
from typing import List, Optional, Tuple, Union

import torch
import torch.nn as nn
import torch.nn.functional as F

import logging

from tqdm import tqdm

from musubi_tuner.utils.safetensors_utils import MemoryEfficientSafeOpen, TensorWeightAdapter, WeightTransformHooks
from musubi_tuner.utils.device_utils import clean_memory_on_device

logger = logging.getLogger(__name__)

# ---------------------------------------------------------------------------
# NF4 codebook (16 values, symmetric around 0, optimal for N(0,1) weights)
# ---------------------------------------------------------------------------
NF4_CODEBOOK = torch.tensor(
    [
        -1.0,
        -0.6961928009986877,
        -0.5250730514526367,
        -0.39491748809814453,
        -0.28444138169288635,
        -0.18477343022823334,
        -0.09105003625154495,
        0.0,
        0.07958029955625534,
        0.16093020141124725,
        0.24611230194568634,
        0.33791524171829224,
        0.44070982933044434,
        0.5626170039176941,
        0.7229568362236023,
        1.0,
    ],
    dtype=torch.float32,
)

DEFAULT_NF4_BLOCK_SIZE = 32


# ---------------------------------------------------------------------------
# Pack / unpack helpers
# ---------------------------------------------------------------------------

def pack_uint4(indices: torch.Tensor) -> torch.Tensor:
    """Pack pairs of 4-bit indices into uint8.  ``indices`` must have even length."""
    assert indices.numel() % 2 == 0, "Number of elements must be even for uint4 packing"
    flat = indices.view(-1)
    high = flat[0::2].to(torch.uint8)
    low = flat[1::2].to(torch.uint8)
    return (high << 4) | low


def unpack_uint4(packed: torch.Tensor, num_elements: int) -> torch.Tensor:
    """Unpack uint8 → pairs of 4-bit indices.  Returns ``[num_elements]`` uint8."""
    flat = packed.view(-1)
    high = (flat >> 4) & 0x0F
    low = flat & 0x0F
    # interleave: [h0, l0, h1, l1, ...]
    out = torch.stack([high, low], dim=1).view(-1)
    return out[:num_elements].to(torch.uint8)


# ---------------------------------------------------------------------------
# Quantize / dequantize
# ---------------------------------------------------------------------------

def quantize_nf4_block(
    tensor: torch.Tensor,
    block_size: int = DEFAULT_NF4_BLOCK_SIZE,
) -> Tuple[torch.Tensor, torch.Tensor]:
    """Quantize a 2-D weight tensor to NF4 with per-block absmax scaling.

    Args:
        tensor: ``[out_features, in_features]`` float weight.
        block_size: number of elements per quantization block.

    Returns:
        packed_weight: ``[out_features, in_features // 2]`` uint8 (two indices per byte).
        scale: ``[out_features, num_blocks, 1]`` float32 absmax per block.
    """
    assert tensor.ndim == 2, f"Expected 2-D tensor, got {tensor.ndim}-D"
    out_features, in_features = tensor.shape
    if in_features % block_size != 0:
        raise ValueError(
            f"in_features ({in_features}) must be divisible by block_size ({block_size})"
        )
    num_blocks = in_features // block_size

    # Reshape to [out, num_blocks, block_size]
    t = tensor.float().contiguous().view(out_features, num_blocks, block_size)

    # Per-block absmax scale
    scale = t.abs().amax(dim=2, keepdim=True).clamp_(min=1e-8)  # [out, num_blocks, 1]

    # Normalize to [-1, 1]
    normalized = t / scale  # [out, num_blocks, block_size]

    # Find nearest codebook entry using bucketize (O(N) memory, not O(N*16)).
    # The codebook is sorted, so we compute midpoints between consecutive values
    # and use bucketize to find which bin each normalized value falls into.
    codebook = NF4_CODEBOOK.to(normalized.device)  # [16]
    # Midpoints between consecutive codebook values → 15 boundaries
    midpoints = (codebook[:-1] + codebook[1:]) / 2.0  # [15]
    flat_norm = normalized.reshape(-1)  # [N]
    indices = torch.bucketize(flat_norm, midpoints)  # [N] int64, values in [0, 15]

    # Reshape indices back and pack
    indices = indices.view(out_features, in_features)  # [out, in]
    packed = pack_uint4(indices).view(out_features, in_features // 2)

    return packed, scale.to(torch.float32)


def dequantize_nf4_block(
    packed_weight: torch.Tensor,
    scale: torch.Tensor,
    out_features: int,
    in_features: int,
    block_size: int = DEFAULT_NF4_BLOCK_SIZE,
    dtype: torch.dtype = torch.bfloat16,
) -> torch.Tensor:
    """Dequantize NF4-packed weight back to a dense float tensor.

    Args:
        packed_weight: ``[out_features, in_features // 2]`` uint8.
        scale: ``[out_features, num_blocks, 1]`` float32.
        out_features, in_features: original weight shape.
        block_size: must match quantization block size.
        dtype: output dtype (e.g. bfloat16).

    Returns:
        ``[out_features, in_features]`` tensor in *dtype*.
    """
    num_blocks = in_features // block_size
    codebook = NF4_CODEBOOK.to(scale.device)  # [16]

    indices = unpack_uint4(packed_weight, out_features * in_features)
    indices = indices.to(scale.device).long()
    values = codebook[indices]  # float32

    values = values.view(out_features, num_blocks, block_size)
    values = values * scale  # broadcast [out, num_blocks, 1]
    values = values.view(out_features, in_features)
    return values.to(dtype)


# ---------------------------------------------------------------------------
# State-dict-level quantization (mirrors optimize_state_dict_with_fp8)
# ---------------------------------------------------------------------------

def optimize_state_dict_with_nf4(
    state_dict: dict,
    calc_device: Union[str, torch.device],
    target_layer_keys: Optional[List[str]] = None,
    exclude_layer_keys: Optional[List[str]] = None,
    block_size: int = DEFAULT_NF4_BLOCK_SIZE,
    move_to_device: bool = False,
) -> dict:
    """Quantize target Linear weights in *state_dict* to NF4 in-place.

    For each target ``.weight`` key the dict is updated with:
    * original key → packed uint8 weight ``[out, in//2]``
    * ``.scale_weight`` → per-block scale ``[out, num_blocks, 1]``
    * ``.nf4_shape`` → ``torch.tensor([out, in])`` (original shape)
    """
    optimized_count = 0

    target_keys_list = []
    for key in list(state_dict.keys()):
        is_target = (
            target_layer_keys is None or any(p in key for p in target_layer_keys)
        ) and key.endswith(".weight")
        is_excluded = exclude_layer_keys is not None and any(
            p in key for p in exclude_layer_keys
        )
        if is_target and not is_excluded and isinstance(state_dict[key], torch.Tensor):
            target_keys_list.append(key)

    for key in tqdm(target_keys_list, desc="NF4 quantizing"):
        value = state_dict[key]
        if value.ndim != 2:
            continue  # skip non-2D (e.g. bias, norm)
        out_features, in_features = value.shape
        if in_features % block_size != 0:
            logger.warning(
                "Skipping NF4 for %s: in_features=%d not divisible by block_size=%d",
                key,
                in_features,
                block_size,
            )
            continue

        original_device = value.device
        original_dtype = value.dtype
        if calc_device is not None:
            value = value.to(calc_device)

        packed, scale = quantize_nf4_block(value, block_size)

        target_device = calc_device if (calc_device is not None and move_to_device) else original_device
        state_dict[key] = packed.to(target_device)
        state_dict[key.replace(".weight", ".scale_weight")] = scale.to(
            dtype=original_dtype, device=target_device
        )
        state_dict[key.replace(".weight", ".nf4_shape")] = torch.tensor(
            [out_features, in_features], dtype=torch.int64
        )

        optimized_count += 1
        if calc_device is not None and optimized_count % 10 == 0:
            clean_memory_on_device(calc_device)

    logger.info(f"Number of NF4-quantized Linear layers: {optimized_count}")
    return state_dict


def load_safetensors_with_nf4_optimization(
    model_files: List[str],
    calc_device: Union[str, torch.device],
    target_layer_keys: Optional[List[str]] = None,
    exclude_layer_keys: Optional[List[str]] = None,
    block_size: int = DEFAULT_NF4_BLOCK_SIZE,
    move_to_device: bool = False,
    weight_hook=None,
    disable_numpy_memmap: bool = False,
    weight_transform_hooks: Optional[WeightTransformHooks] = None,
) -> dict:
    """Load safetensors and quantize target layers to NF4.

    Mirrors ``load_safetensors_with_fp8_optimization``.
    """

    def is_target_key(key):
        is_target = (
            target_layer_keys is None or any(p in key for p in target_layer_keys)
        ) and key.endswith(".weight")
        is_excluded = exclude_layer_keys is not None and any(
            p in key for p in exclude_layer_keys
        )
        return is_target and not is_excluded

    optimized_count = 0
    state_dict: dict = {}

    for model_file in model_files:
        with MemoryEfficientSafeOpen(model_file, disable_numpy_memmap=disable_numpy_memmap) as original_f:
            f = (
                TensorWeightAdapter(weight_transform_hooks, original_f)
                if weight_transform_hooks is not None
                else original_f
            )
            keys = f.keys()
            for key in tqdm(keys, desc=f"Loading {os.path.basename(model_file)}", unit="key"):
                value = f.get_tensor(key)
                original_device = value.device

                if weight_hook is not None:
                    value = weight_hook(key, value, keep_on_calc_device=(calc_device is not None))

                if not is_target_key(key) or value.ndim != 2:
                    target_device = calc_device if (calc_device is not None and move_to_device) else original_device
                    state_dict[key] = value.to(target_device)
                    continue

                out_features, in_features = value.shape
                if in_features % block_size != 0:
                    logger.warning(
                        "Skipping NF4 for %s: in_features=%d not divisible by block_size=%d",
                        key,
                        in_features,
                        block_size,
                    )
                    target_device = calc_device if (calc_device is not None and move_to_device) else original_device
                    state_dict[key] = value.to(target_device)
                    continue

                if calc_device is not None:
                    value = value.to(calc_device)

                original_dtype = value.dtype
                packed, scale = quantize_nf4_block(value, block_size)

                target_device = calc_device if (calc_device is not None and move_to_device) else original_device
                state_dict[key] = packed.to(target_device)
                state_dict[key.replace(".weight", ".scale_weight")] = scale.to(
                    dtype=original_dtype, device=target_device
                )
                state_dict[key.replace(".weight", ".nf4_shape")] = torch.tensor(
                    [out_features, in_features], dtype=torch.int64
                )

                optimized_count += 1
                if calc_device is not None and optimized_count % 10 == 0:
                    clean_memory_on_device(calc_device)

    logger.info(f"Number of NF4-quantized Linear layers: {optimized_count}")
    return state_dict


# ---------------------------------------------------------------------------
# Monkey-patched forward
# ---------------------------------------------------------------------------

def nf4_linear_forward_patch(self: nn.Linear, x: torch.Tensor) -> torch.Tensor:
    """Drop-in replacement forward for NF4-quantized Linear layers."""
    w = dequantize_nf4_block(
        self.weight,
        self.scale_weight,
        self.nf4_out_features,
        self.nf4_in_features,
        self.nf4_block_size,
        self.scale_weight.dtype,
    )
    if hasattr(self, "awq_scales"):
        # Un-scale columns: divide by the AWQ scales that were multiplied before quantization
        w = w / self.awq_scales.unsqueeze(0)  # [1, in] broadcast over [out, in]
        w = w.to(self.scale_weight.dtype)
    return F.linear(x, w, self.bias)


def apply_nf4_monkey_patch(
    model: nn.Module,
    optimized_state_dict: dict,
    block_size: int = DEFAULT_NF4_BLOCK_SIZE,
    awq_scales: Optional[dict] = None,
) -> nn.Module:
    """Register NF4 buffers and replace forward on quantized Linears.

    Mirrors ``apply_fp8_monkey_patch``.

    Args:
        awq_scales: Optional dict mapping module_name -> scale tensor [in_features].
            If provided, each patched module gets an ``awq_scales`` buffer for
            un-scaling during the forward pass.
    """
    # Identify NF4-quantized modules from the state dict
    shape_keys = [k for k in optimized_state_dict if k.endswith(".nf4_shape")]
    patched_module_paths: set = set()
    shape_info: dict = {}
    scale_shape_info: dict = {}
    for shape_key in shape_keys:
        module_path = shape_key.rsplit(".nf4_shape", 1)[0]
        patched_module_paths.add(module_path)
        shape_tensor = optimized_state_dict[shape_key]
        shape_info[module_path] = (int(shape_tensor[0].item()), int(shape_tensor[1].item()))
        scale_key = module_path + ".scale_weight"
        scale_shape_info[module_path] = optimized_state_dict[scale_key].shape

    patched_count = 0
    for name, module in model.named_modules():
        if name not in patched_module_paths:
            continue
        if not isinstance(module, nn.Linear):
            continue

        out_f, in_f = shape_info[name]
        scale_shape = scale_shape_info[name]
        packed_shape = (out_f, in_f // 2)

        # Replace weight parameter with a buffer of packed shape so
        # load_state_dict(assign=True) doesn't complain about shape mismatch.
        weight_dtype = module.weight.dtype if module.weight.dtype != torch.float32 else torch.uint8
        del module.weight
        module.register_buffer("weight", torch.zeros(packed_shape, dtype=torch.uint8))

        # Register buffers so load_state_dict can assign them
        module.register_buffer("scale_weight", torch.ones(scale_shape, dtype=weight_dtype))
        module.register_buffer("nf4_shape", torch.zeros(2, dtype=torch.int64))

        # Store metadata as plain attributes
        module.nf4_out_features = out_f
        module.nf4_in_features = in_f
        module.nf4_block_size = block_size
        module._nf4_quantized = True

        # Register AWQ scales buffer if available
        if awq_scales is not None:
            # AWQ scales are keyed by weight key (e.g. "layer.weight")
            weight_key = name + ".weight"
            if weight_key in awq_scales:
                module.register_buffer(
                    "awq_scales",
                    awq_scales[weight_key].to(dtype=torch.float32),
                )

        # Replace forward
        def new_forward(self, x):
            return nf4_linear_forward_patch(self, x)

        module.forward = new_forward.__get__(module, type(module))
        patched_count += 1

    logger.info(f"Number of NF4 monkey-patched Linear layers: {patched_count}")
    return model


# ---------------------------------------------------------------------------
# Detection helper
# ---------------------------------------------------------------------------

def is_nf4_module(module: nn.Module) -> bool:
    """Check whether *module* has been NF4-quantized (attribute-based, not dtype)."""
    return getattr(module, "_nf4_quantized", False)