File size: 215 Bytes
574a626 | 1 2 3 4 5 6 7 8 9 10 11 | {
"architectures": ["GPTQForCausalLM"],
"model_type": "gptq",
"quantization_config": {
"bits": 4,
"group_size": 128,
"desc_act": false
},
"torch_dtype": "float16",
"trust_remote_code": true
} |