File size: 1,292 Bytes
ba6cfb8
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
{
  "model_name": "bnb_8bit_conservative_stem_optimized",
  "source_model": "luca-deandrea/MNLP_M3_mcqa_model",
  "quantization_method": "BitsAndBytes_8bit",
  "timestamp": "2025-06-10 23:50:21",
  "processing_time_seconds": 25.678223609924316,
  "quantization_config": {
    "load_in_8bit": true,
    "llm_int8_threshold": 8.0,
    "llm_int8_skip_modules": [
      "lm_head"
    ],
    "llm_int8_enable_fp32_cpu_offload": false,
    "llm_int8_has_fp16_weight": false
  },
  "model_stats": {
    "size_mb": 732.8879594802856,
    "validation_score": 0.6,
    "compression_ratio": "~2x compared to FP16"
  },
  "stem_optimizations": [
    "conservative_8bit_threshold_8.0",
    "lm_head_layer_kept_in_fp16",
    "deterministic_generation_config",
    "optimized_for_stem_reasoning",
    "cache_enabled_for_performance"
  ],
  "usage_instructions": {
    "library": "transformers",
    "loading": "AutoModelForCausalLM.from_pretrained('thresh10')",
    "recommended_for": "MNLP_STEM_mcqa_evals",
    "notes": "Model ready for lighteval evaluation"
  },
  "performance_characteristics": {
    "accuracy": "High (conservative quantization)",
    "speed": "Good (8-bit quantization)",
    "memory": "Reduced (~50% of original)",
    "best_for": "STEM reasoning tasks requiring high accuracy"
  }
}