Upload model_info.json with huggingface_hub
Browse files- model_info.json +40 -0
model_info.json
ADDED
|
@@ -0,0 +1,40 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"model_name": "bnb_8bit_conservative_stem_optimized",
|
| 3 |
+
"source_model": "luca-deandrea/MNLP_M3_mcqa_model",
|
| 4 |
+
"quantization_method": "BitsAndBytes_8bit",
|
| 5 |
+
"timestamp": "2025-06-10 23:38:18",
|
| 6 |
+
"processing_time_seconds": 62.69236969947815,
|
| 7 |
+
"quantization_config": {
|
| 8 |
+
"load_in_8bit": true,
|
| 9 |
+
"llm_int8_threshold": 8.0,
|
| 10 |
+
"llm_int8_skip_modules": [
|
| 11 |
+
"lm_head"
|
| 12 |
+
],
|
| 13 |
+
"llm_int8_enable_fp32_cpu_offload": false,
|
| 14 |
+
"llm_int8_has_fp16_weight": false
|
| 15 |
+
},
|
| 16 |
+
"model_stats": {
|
| 17 |
+
"size_mb": 732.8867292404175,
|
| 18 |
+
"validation_score": 0.4,
|
| 19 |
+
"compression_ratio": "~2x compared to FP16"
|
| 20 |
+
},
|
| 21 |
+
"stem_optimizations": [
|
| 22 |
+
"conservative_8bit_threshold_8.0",
|
| 23 |
+
"lm_head_layer_kept_in_fp16",
|
| 24 |
+
"deterministic_generation_config",
|
| 25 |
+
"optimized_for_stem_reasoning",
|
| 26 |
+
"cache_enabled_for_performance"
|
| 27 |
+
],
|
| 28 |
+
"usage_instructions": {
|
| 29 |
+
"library": "transformers",
|
| 30 |
+
"loading": "AutoModelForCausalLM.from_pretrained('thresh10')",
|
| 31 |
+
"recommended_for": "MNLP_STEM_mcqa_evals",
|
| 32 |
+
"notes": "Model ready for lighteval evaluation"
|
| 33 |
+
},
|
| 34 |
+
"performance_characteristics": {
|
| 35 |
+
"accuracy": "High (conservative quantization)",
|
| 36 |
+
"speed": "Good (8-bit quantization)",
|
| 37 |
+
"memory": "Reduced (~50% of original)",
|
| 38 |
+
"best_for": "STEM reasoning tasks requiring high accuracy"
|
| 39 |
+
}
|
| 40 |
+
}
|