Upload quantization_info.json with huggingface_hub
Browse files- quantization_info.json +5 -5
quantization_info.json
CHANGED
|
@@ -1,5 +1,5 @@
|
|
| 1 |
{
|
| 2 |
-
"model_name": "
|
| 3 |
"quantization_method": "bitsandbytes",
|
| 4 |
"config_name": "bnb_4bit_fp4",
|
| 5 |
"description": "4-bit FP4 with double quantization",
|
|
@@ -12,8 +12,8 @@
|
|
| 12 |
"bnb_4bit_compute_dtype": "torch.float16",
|
| 13 |
"llm_int8_threshold": 6.0
|
| 14 |
},
|
| 15 |
-
"timestamp": "2025-06-10 21:
|
| 16 |
-
"loading_time_seconds": 4.
|
| 17 |
-
"memory_usage_gb": 0.
|
| 18 |
-
"test_generation": "Hello, I am a
|
| 19 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"model_name": "luca-deandrea/MNLP_M3_mcqa_model",
|
| 3 |
"quantization_method": "bitsandbytes",
|
| 4 |
"config_name": "bnb_4bit_fp4",
|
| 5 |
"description": "4-bit FP4 with double quantization",
|
|
|
|
| 12 |
"bnb_4bit_compute_dtype": "torch.float16",
|
| 13 |
"llm_int8_threshold": 6.0
|
| 14 |
},
|
| 15 |
+
"timestamp": "2025-06-10 21:31:28",
|
| 16 |
+
"loading_time_seconds": 4.605607509613037,
|
| 17 |
+
"memory_usage_gb": 0.5091643333435059,
|
| 18 |
+
"test_generation": "Hello, I am a bit confused about the difference between the terms of"
|
| 19 |
}
|