awq / quantization_info.json
Thomaschtl's picture
Upload quantization_info.json with huggingface_hub
4e9354e verified
{
"quantization_method": "awq_4bit",
"linear_layers": "0/115 (4-bit AWQ)",
"embedding_layers": "0/0 (not quantized by AWQ)",
"total_quantized": "0/115",
"original_model": "Qwen/Qwen3-0.6B-Base",
"calibration_data_path": "smoothquant/smoothquant/calibration_prompts.json",
"calibration_samples": 100,
"quantization_timestamp": "2025-06-10 20:40:48",
"pytorch_version": "2.6.0+cu118",
"estimated_compression_ratio": "4.0x",
"estimated_size_mb": 284.21875,
"original_size_mb": 1136.875,
"formats_included": [
"awq_safetensors"
],
"lighteval_compatible": true,
"awq_config": {
"zero_point": true,
"q_group_size": 128,
"w_bit": 4,
"version": "GEMM"
},
"notes": "4-bit AWQ quantization with activation-aware weight optimization"
}