GLM-4-9b-AutoRound-4bits / quantization_config.json
Emanresu's picture
Upload folder using huggingface_hub
d5c4f32 verified
raw
history blame contribute delete
285 Bytes
{
"bits": 4,
"group_size": 128,
"sym": true,
"data_type": "int",
"seqlen": 128,
"batch_size": 1,
"lr": 0.005,
"minmax_lr": 0.005,
"iters": 50,
"amp": false,
"autoround_version": "0.7.1",
"quant_method": "auto-round",
"packing_format": "auto_round:auto_gptq"
}