File size: 1,662 Bytes
6b84460
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
---
license: mit
base_model: microsoft/MediPhi-Instruct
library_name: mlx
pipeline_tag: text-generation
language:
- en

tags:
- mlx
- mlx-lm
- phi
- phi-3
- medical
- clinical
- healthcare
- quantized
- 4bit
- on-device
- ios
- apple-silicon

model_type: phi
quantization:
- 4bit

datasets:
- microsoft/mediflow
- ncbi/pubmed
- epfl-llm/guidelines
- starmpcc/Asclepius-Synthetic-Clinical-Notes
- akemiH/NoteChat
- zhengyun21/PMC-Patients
- jpcorb20/medical_wikipedia
---

# MediPhi-Instruct (MLX · 4-bit)

This repository contains an **MLX-format 4-bit quantized** version of  
[`microsoft/MediPhi-Instruct`](https://huggingface.co/microsoft/MediPhi-Instruct),  
converted using **`mlx-lm`** for efficient **on-device inference** on Apple silicon.

This model is intended for **iOS / iPadOS / macOS** usage where memory and power
constraints require aggressive quantization while preserving clinical reasoning quality.

---

## Model details

- **Base model:** MediPhi-Instruct (Phi-3 family)
- **Parameters:** ~3.8B
- **Quantization:** 4-bit (MLX)
- **Format:** MLX (not GGUF)
- **Intended use:** On-device medical and clinical QA, decision support, and explanations
- **Language:** English

> ⚠️ This is a **conversion only**. No additional fine-tuning was performed.

---

## Why MLX 4-bit?

Compared to larger 4–7B medical models, MediPhi-Instruct shows:
- Strong clinical reasoning per parameter
- Better robustness under 4-bit quantization
- Lower memory footprint suitable for mobile devices

This makes it a strong candidate for **on-device medical assistants** on iPhone and iPad.

---

## Usage (MLX-LM)

### Install
```bash
pip install mlx-lm