metadata
license: mit
base_model: microsoft/MediPhi-Instruct
library_name: mlx
pipeline_tag: text-generation
language:
- en
tags:
- mlx
- mlx-lm
- phi
- phi-3
- medical
- clinical
- healthcare
- quantized
- 4bit
- on-device
- ios
- apple-silicon
model_type: phi
quantization:
- 4bit
datasets:
- microsoft/mediflow
- ncbi/pubmed
- epfl-llm/guidelines
- starmpcc/Asclepius-Synthetic-Clinical-Notes
- akemiH/NoteChat
- zhengyun21/PMC-Patients
- jpcorb20/medical_wikipedia
MediPhi-Instruct (MLX · 4-bit)
This repository contains an MLX-format 4-bit quantized version ofmicrosoft/MediPhi-Instruct,
converted using mlx-lm for efficient on-device inference on Apple silicon.
This model is intended for iOS / iPadOS / macOS usage where memory and power constraints require aggressive quantization while preserving clinical reasoning quality.
Model details
- Base model: MediPhi-Instruct (Phi-3 family)
- Parameters: ~3.8B
- Quantization: 4-bit (MLX)
- Format: MLX (not GGUF)
- Intended use: On-device medical and clinical QA, decision support, and explanations
- Language: English
⚠️ This is a conversion only. No additional fine-tuning was performed.
Why MLX 4-bit?
Compared to larger 4–7B medical models, MediPhi-Instruct shows:
- Strong clinical reasoning per parameter
- Better robustness under 4-bit quantization
- Lower memory footprint suitable for mobile devices
This makes it a strong candidate for on-device medical assistants on iPhone and iPad.
Usage (MLX-LM)
Install
pip install mlx-lm