jb10231's picture
Upload README.md with huggingface_hub
9c7cba6 verified
---
language: en
license: llama3.2
base_model: meta-llama/Llama-3.2-3B-Instruct
tags:
- medical
- healthcare
- lab-reports
- llama
- qlora
- peft
datasets:
- lavita/ChatDoctor-HealthCareMagic-100k
- lavita/ChatDoctor-iCliniq
---
# MedLLaMA-3.2-3B: AI Lab Report Analyzer
## Model Description
This is a fine-tuned version of [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct)
trained on medical Q&A data to answer patient queries about lab reports and health conditions.
## Training Details
- **Base Model:** LLaMA-3.2-3B-Instruct
- **Method:** QLoRA (4-bit quantization + LoRA rank 16)
- **Dataset:** MedQuAD + iCliniq (~10k examples)
- **Epochs:** 2
- **Hardware:** NVIDIA T4 (Google Colab)
## Intended Use
- Answering patient questions about lab report values
- Explaining medical terminology in plain language
- Providing general health information
## ⚠️ Limitations & Disclaimer
This model is for **educational and informational purposes only**.
It is **NOT a substitute for professional medical advice, diagnosis, or treatment.**
Always consult a qualified healthcare provider for medical decisions.
## Quick Start
```python
from transformers import AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig
from peft import PeftModel
import torch
base_model = AutoModelForCausalLM.from_pretrained(
'meta-llama/Llama-3.2-3B-Instruct',
quantization_config=BitsAndBytesConfig(load_in_4bit=True),
device_map='auto'
)
model = PeftModel.from_pretrained(base_model, 'jb10231/MedLLaMA-3.2-3B-LabReport')
tokenizer = AutoTokenizer.from_pretrained('jb10231/MedLLaMA-3.2-3B-LabReport')
```