File size: 1,662 Bytes
6b84460 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 |
---
license: mit
base_model: microsoft/MediPhi-Instruct
library_name: mlx
pipeline_tag: text-generation
language:
- en
tags:
- mlx
- mlx-lm
- phi
- phi-3
- medical
- clinical
- healthcare
- quantized
- 4bit
- on-device
- ios
- apple-silicon
model_type: phi
quantization:
- 4bit
datasets:
- microsoft/mediflow
- ncbi/pubmed
- epfl-llm/guidelines
- starmpcc/Asclepius-Synthetic-Clinical-Notes
- akemiH/NoteChat
- zhengyun21/PMC-Patients
- jpcorb20/medical_wikipedia
---
# MediPhi-Instruct (MLX · 4-bit)
This repository contains an **MLX-format 4-bit quantized** version of
[`microsoft/MediPhi-Instruct`](https://huggingface.co/microsoft/MediPhi-Instruct),
converted using **`mlx-lm`** for efficient **on-device inference** on Apple silicon.
This model is intended for **iOS / iPadOS / macOS** usage where memory and power
constraints require aggressive quantization while preserving clinical reasoning quality.
---
## Model details
- **Base model:** MediPhi-Instruct (Phi-3 family)
- **Parameters:** ~3.8B
- **Quantization:** 4-bit (MLX)
- **Format:** MLX (not GGUF)
- **Intended use:** On-device medical and clinical QA, decision support, and explanations
- **Language:** English
> ⚠️ This is a **conversion only**. No additional fine-tuning was performed.
---
## Why MLX 4-bit?
Compared to larger 4–7B medical models, MediPhi-Instruct shows:
- Strong clinical reasoning per parameter
- Better robustness under 4-bit quantization
- Lower memory footprint suitable for mobile devices
This makes it a strong candidate for **on-device medical assistants** on iPhone and iPad.
---
## Usage (MLX-LM)
### Install
```bash
pip install mlx-lm
|