File size: 667 Bytes
0af2a02 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 |
{
"model_name": "luca-deandrea/MNLP_M3_mcqa_model",
"quantization_method": "bitsandbytes",
"config_name": "bnb_8bit_aggressive",
"description": "8-bit with aggressive quantization",
"expected_compression": "~4x",
"quantization_config": {
"load_in_4bit": false,
"load_in_8bit": true,
"bnb_4bit_quant_type": "fp4",
"bnb_4bit_use_double_quant": false,
"bnb_4bit_compute_dtype": "torch.float32",
"llm_int8_threshold": 0.0
},
"timestamp": "2025-06-10 21:32:18",
"loading_time_seconds": 10.248082876205444,
"memory_usage_gb": 0.7341856956481934,
"test_generation": "Hello, I am a student who has been working on a project to"
} |