Thomaschtl commited on
Commit
f2acd09
·
verified ·
1 Parent(s): f034a31

Upload model_info.json with huggingface_hub

Browse files
Files changed (1) hide show
  1. model_info.json +40 -0
model_info.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_name": "bnb_8bit_conservative_stem_optimized",
3
+ "source_model": "luca-deandrea/MNLP_M3_mcqa_model",
4
+ "quantization_method": "BitsAndBytes_8bit",
5
+ "timestamp": "2025-06-10 23:38:18",
6
+ "processing_time_seconds": 62.69236969947815,
7
+ "quantization_config": {
8
+ "load_in_8bit": true,
9
+ "llm_int8_threshold": 8.0,
10
+ "llm_int8_skip_modules": [
11
+ "lm_head"
12
+ ],
13
+ "llm_int8_enable_fp32_cpu_offload": false,
14
+ "llm_int8_has_fp16_weight": false
15
+ },
16
+ "model_stats": {
17
+ "size_mb": 732.8867292404175,
18
+ "validation_score": 0.4,
19
+ "compression_ratio": "~2x compared to FP16"
20
+ },
21
+ "stem_optimizations": [
22
+ "conservative_8bit_threshold_8.0",
23
+ "lm_head_layer_kept_in_fp16",
24
+ "deterministic_generation_config",
25
+ "optimized_for_stem_reasoning",
26
+ "cache_enabled_for_performance"
27
+ ],
28
+ "usage_instructions": {
29
+ "library": "transformers",
30
+ "loading": "AutoModelForCausalLM.from_pretrained('thresh10')",
31
+ "recommended_for": "MNLP_STEM_mcqa_evals",
32
+ "notes": "Model ready for lighteval evaluation"
33
+ },
34
+ "performance_characteristics": {
35
+ "accuracy": "High (conservative quantization)",
36
+ "speed": "Good (8-bit quantization)",
37
+ "memory": "Reduced (~50% of original)",
38
+ "best_for": "STEM reasoning tasks requiring high accuracy"
39
+ }
40
+ }