| --- |
| language: en |
| license: mit |
| tags: |
| - medical |
| - pharmaceutical |
| - autocomplete |
| - distillation |
| - gpt2 |
| datasets: |
| - medmcqa |
| metrics: |
| - perplexity |
| model-index: |
| - name: codehance/distilgpt2-medical-pharma |
| results: |
| - task: |
| type: text-generation |
| dataset: |
| name: Medical Q&A |
| type: medmcqa |
| metrics: |
| - name: Perplexity |
| type: perplexity |
| value: 44.07 |
| --- |
| |
| # DistilGPT-2 Medical Pharmaceutical Autocomplete |
|
|
| ## Model Description |
|
|
| This is a distilled GPT-2 model fine-tuned for pharmaceutical autocomplete. It suggests drug names and medical terminology based on clinical context. |
|
|
| **Key Features:** |
| - 34% smaller than base fine-tuned model (81,912,576 parameters) |
| - 45% faster inference (347.9ms per generation) |
| - Specialized in pharmaceutical vocabulary |
|
|
| ## Training Process |
|
|
| ### Stage 1: Fine-Tuning |
| - Base model: GPT-2 (124M parameters) |
| - Dataset: Medical Q&A (medmcqa) - 4,500 training examples |
| - Training: 3 epochs |
| - Final perplexity: 23.61 |
|
|
| ### Stage 2: Knowledge Distillation |
| - Teacher: Fine-tuned GPT-2 |
| - Student: DistilGPT-2 |
| - Training: 2 epochs |
| - Compression: 34.2% size reduction |
|
|
| ## Performance |
|
|
| | Metric | Value | |
| |--------|-------| |
| | Parameters | 81,912,576 | |
| | Perplexity | 44.07 | |
| | Inference Speed | 347.9ms | |
| | Quality Retained | 53.6% | |
|
|
| ## Usage |
| ```python |
| from transformers import GPT2LMHeadModel, GPT2Tokenizer |
| |
| # Load model and tokenizer |
| model = GPT2LMHeadModel.from_pretrained("codehance/distilgpt2-medical-pharma") |
| tokenizer = GPT2Tokenizer.from_pretrained("codehance/distilgpt2-medical-pharma") |
| |
| # Generate pharmaceutical suggestions |
| prompt = "The patient should take" |
| inputs = tokenizer(prompt, return_tensors="pt") |
| outputs = model.generate(**inputs, max_length=30, num_return_sequences=3) |
| |
| for output in outputs: |
| print(tokenizer.decode(output, skip_special_tokens=True)) |
| ``` |
|
|
| ## Intended Use |
|
|
| **Primary Use Cases:** |
| - Pharmaceutical autocomplete systems |
| - Medical documentation assistance |
| - Clinical note-taking tools |
| - Drug name suggestion |
|
|
| **Limitations:** |
| - Not a substitute for medical advice |
| - May suggest incorrect drugs - always verify with qualified professionals |
| - Trained on medical exam questions, not real prescriptions |
| - English language only |
|
|
| ## Training Data |
|
|
| - **Source:** MedMCQA dataset (Indian medical entrance exam questions) |
| - **Size:** 4,500 training examples |
| - **Content:** Medical questions with pharmaceutical terminology |
|
|
| ## Ethical Considerations |
|
|
| ⚠️ **Important:** This model is for autocomplete assistance only. It should NOT be used as the sole basis for medical decisions. Always verify suggestions with qualified healthcare professionals. |
|
|
| ## Model Card Authors |
|
|
| Created as part of a pharmaceutical autocomplete system tutorial demonstrating transfer learning, fine-tuning, and knowledge distillation. |
|
|
| ## Citation |
| ```bibtex |
| @misc{distilgpt2-medical-pharma, |
| author = {codehance}, |
| title = {DistilGPT-2 Medical Pharmaceutical Autocomplete}, |
| year = {2025}, |
| publisher = {Hugging Face}, |
| howpublished = {\url{https://huggingface.co/codehance/distilgpt2-medical-pharma}} |
| } |
| ``` |
|
|