phi4-Instruct-AutoRound-GPTQ-4bit / quantize_config.json
vizsatiz's picture
Upload quantized model (4-bit AutoRound) Phi-4
92f9295 verified
{
"bits": 4,
"group_size": 128,
"sym": true,
"data_type": "int",
"iters": 256,
"autoround_version": "0.6.0",
"quant_method": "gptq",
"desc_act": false,
"true_sequential": false,
"damp_percent": 0.01
}