File size: 804 Bytes
1534341 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 |
{
"quantization_method": "awq_4bit",
"linear_layers": "0/115 (4-bit AWQ)",
"embedding_layers": "0/0 (not quantized by AWQ)",
"total_quantized": "0/115",
"original_model": "luca-deandrea/MNLP_M3_mcqa_model",
"calibration_data_path": "smoothquant/smoothquant/calibration_prompts.json",
"calibration_samples": 100,
"quantization_timestamp": "2025-06-10 20:48:02",
"pytorch_version": "2.6.0+cu118",
"estimated_compression_ratio": "4.0x",
"estimated_size_mb": 284.08837890625,
"original_size_mb": 1136.353515625,
"formats_included": [
"awq_safetensors"
],
"lighteval_compatible": true,
"awq_config": {
"zero_point": true,
"q_group_size": 128,
"w_bit": 4,
"version": "GEMM"
},
"notes": "4-bit AWQ quantization with activation-aware weight optimization"
} |