File size: 1,734 Bytes
acc64cf | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 | === VPTQ Quantization Configuration === Date: 2026-02-22 19:40:24 UTC Target: ~2-bit quantization with Residual VQ Model: speakleash/Bielik-11B-v2.3-Instruct (MistralForCausalLM) Parameters: --model_name /workspace/models/bielik-11b-instruct --output_dir /workspace/variant-e/output/ --vector_lens -1 8 # Skip embeddings (-1), quantize in vectors of 8 --group_num 1 # Single group --num_centroids -1 65536 # 2^16 centroids = 16 bits per vector index --num_res_centroids -1 256 # 2^8 residual centroids --npercent 0 # No outlier channels --blocksize 128 # Block size for quantization --new_eval # Use new evaluation method --seq_len 4096 # Sequence length for calibration --kmeans_mode hessian # Hessian-weighted k-means clustering --num_gpus 1 # Single GPU --enable_perm # Channel permutation for better quant --enable_norm # Channel normalization --save_model # Save quantized model --save_packed_model # Save packed model for inference --hessian_path /workspace/hessians/quip-format/hessians --kiter 100 # K-means iterations --ktol 1e-5 # K-means convergence tolerance Effective bitwidth calculation: Primary: log2(65536) / 8 = 16/8 = 2.0 bits/weight Residual: log2(256) / 8 = 8/8 = 1.0 bits/weight overhead Total: ~2.0 bits/weight (residual adds codebook storage overhead, not per-weight) Hessian source: Jakubrd4/bielik-quip-e8p12 Format: QuIP# (flatH + mu + n + ct per layer) Calibration: CulturaX-PL, 512 samples x 4096 tokens Layers: 50 x 4 projections (qkv, o, up, down) = 200 files |