Upload Biomni-R0-32B-PTQ-INT8 - PTQ INT8 quantized variant of Biomni-R0-32B-Preview
c839f1f verified | { | |
| "method": "PTQ-INT8", | |
| "original_model": "/projects/extern/kisski/kisski-narges-llm-interactive/dir.project/Biomni-R0-32B-Preview", | |
| "quantized_model": "/projects/extern/kisski/kisski-narges-llm-interactive/dir.project/Biomni-R0-32B-PTQ-INT8", | |
| "calibration_samples": 120, | |
| "quantization_scheme": "INT8 weights and activations", | |
| "backend": "optimum-quanto", | |
| "model_size_estimate": "~8-10 GB (vs ~60 GB original)", | |
| "note": "Quantization may be applied at serving time by vLLM" | |
| } |