File size: 3,078 Bytes
09aaa29 d37dfcd 09aaa29 d37dfcd 09aaa29 d37dfcd 09aaa29 d37dfcd 09aaa29 d37dfcd 09aaa29 d37dfcd 09aaa29 d37dfcd 09aaa29 d37dfcd 09aaa29 d37dfcd 09aaa29 d37dfcd 09aaa29 d37dfcd 09aaa29 d37dfcd 09aaa29 d37dfcd 09aaa29 d37dfcd 09aaa29 d37dfcd 09aaa29 d37dfcd 09aaa29 d37dfcd 09aaa29 d37dfcd 09aaa29 d37dfcd 09aaa29 d37dfcd 09aaa29 d37dfcd 09aaa29 d37dfcd 09aaa29 d37dfcd 09aaa29 d37dfcd | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 | ---
language: en
license: mit
tags:
- medical
- pharmaceutical
- autocomplete
- distillation
- gpt2
datasets:
- medmcqa
metrics:
- perplexity
model-index:
- name: codehance/distilgpt2-medical-pharma
results:
- task:
type: text-generation
dataset:
name: Medical Q&A
type: medmcqa
metrics:
- name: Perplexity
type: perplexity
value: 44.07
---
# DistilGPT-2 Medical Pharmaceutical Autocomplete
## Model Description
This is a distilled GPT-2 model fine-tuned for pharmaceutical autocomplete. It suggests drug names and medical terminology based on clinical context.
**Key Features:**
- 34% smaller than base fine-tuned model (81,912,576 parameters)
- 45% faster inference (347.9ms per generation)
- Specialized in pharmaceutical vocabulary
## Training Process
### Stage 1: Fine-Tuning
- Base model: GPT-2 (124M parameters)
- Dataset: Medical Q&A (medmcqa) - 4,500 training examples
- Training: 3 epochs
- Final perplexity: 23.61
### Stage 2: Knowledge Distillation
- Teacher: Fine-tuned GPT-2
- Student: DistilGPT-2
- Training: 2 epochs
- Compression: 34.2% size reduction
## Performance
| Metric | Value |
|--------|-------|
| Parameters | 81,912,576 |
| Perplexity | 44.07 |
| Inference Speed | 347.9ms |
| Quality Retained | 53.6% |
## Usage
```python
from transformers import GPT2LMHeadModel, GPT2Tokenizer
# Load model and tokenizer
model = GPT2LMHeadModel.from_pretrained("codehance/distilgpt2-medical-pharma")
tokenizer = GPT2Tokenizer.from_pretrained("codehance/distilgpt2-medical-pharma")
# Generate pharmaceutical suggestions
prompt = "The patient should take"
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(**inputs, max_length=30, num_return_sequences=3)
for output in outputs:
print(tokenizer.decode(output, skip_special_tokens=True))
```
## Intended Use
**Primary Use Cases:**
- Pharmaceutical autocomplete systems
- Medical documentation assistance
- Clinical note-taking tools
- Drug name suggestion
**Limitations:**
- Not a substitute for medical advice
- May suggest incorrect drugs - always verify with qualified professionals
- Trained on medical exam questions, not real prescriptions
- English language only
## Training Data
- **Source:** MedMCQA dataset (Indian medical entrance exam questions)
- **Size:** 4,500 training examples
- **Content:** Medical questions with pharmaceutical terminology
## Ethical Considerations
⚠️ **Important:** This model is for autocomplete assistance only. It should NOT be used as the sole basis for medical decisions. Always verify suggestions with qualified healthcare professionals.
## Model Card Authors
Created as part of a pharmaceutical autocomplete system tutorial demonstrating transfer learning, fine-tuning, and knowledge distillation.
## Citation
```bibtex
@misc{distilgpt2-medical-pharma,
author = {codehance},
title = {DistilGPT-2 Medical Pharmaceutical Autocomplete},
year = {2025},
publisher = {Hugging Face},
howpublished = {\url{https://huggingface.co/codehance/distilgpt2-medical-pharma}}
}
```
|