{ "method": "gptq_v2", "version": "v2", "bits": 4, "group_size": 128, "quantization_time": 319.0039792060852, "model_size_gb": 0.5041657388210297, "peak_memory_gb": 2.2490234375, "calibration_size": 1024, "config": { "method": "gptq_v2", "version": "v2", "bits": 4, "group_size": 128, "desc_act": true, "sym": true, "true_sequential": true, "batch_size": 1, "auto_gc": false } }