File size: 1,653 Bytes
6664e77
79d9b0a
 
 
 
 
 
 
 
 
 
 
6664e77
 
79d9b0a
6664e77
79d9b0a
 
6664e77
79d9b0a
 
 
6664e77
79d9b0a
6664e77
 
 
79d9b0a
 
 
 
 
 
 
6664e77
79d9b0a
6664e77
79d9b0a
6664e77
79d9b0a
6664e77
79d9b0a
 
 
6664e77
79d9b0a
6664e77
79d9b0a
6664e77
79d9b0a
 
 
6664e77
79d9b0a
6664e77
79d9b0a
6664e77
79d9b0a
 
 
6664e77
79d9b0a
 
 
6664e77
79d9b0a
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
---
license: apache-2.0
base_model: deepseek-ai/deepseek-11m-7b-base
tags:
- phi-deidentification
- healthcare-nlp
- medical-text
- lora
- peft
- privacy
- ner
- safety
---

# DeepSeek PHI De-identification Adapter

This repository hosts a LoRA adapter fine-tuned for safe detection and redaction of
Protected Health Information (PHI) in clinical text.

The model is trained on a large synthetic and de-identified corpus derived from
MIMIC-III-style clinical notes and is designed to operate as part of a configurable,
explainable medical text de-identification pipeline.

---

## Model Details

- **Developed by:** Iftakhar Khandokar (Marquette University)
- **Funded by:** Academic research (EECE Department, Marquette University)
- **Shared by:** Iftakhar Khandokar
- **Model type:** LoRA adapter (PEFT)
- **Base model:** `deepseek-ai/deepseek-11m-7b-base`
- **Language:** English (clinical / biomedical NLP)
- **License:** Apache 2.0

---

## Intended Use

This adapter is intended for:

✅ Research on medical data de-identification  
✅ Benchmarking privacy-preserving NLP pipelines  
✅ Safety and explainability evaluation for clinical LLM workflows  

---

## Not Intended For

❌ Automated medical diagnosis  
❌ Direct patient care deployment without regulatory review  
❌ Generating synthetic patient records for real-world use

---

## Loading the Model

```python
from transformers import AutoModelForCausalLM
from peft import PeftModel

base = AutoModelForCausalLM.from_pretrained(
    "deepseek-ai/deepseek-11m-7b-base", trust_remote_code=True
)

model = PeftModel.from_pretrained(
    base,
    "Iftakhar/deepseek-phi-adapter"
)