--- language: en license: llama3.2 base_model: meta-llama/Llama-3.2-3B-Instruct tags: - medical - healthcare - lab-reports - llama - qlora - peft datasets: - lavita/ChatDoctor-HealthCareMagic-100k - lavita/ChatDoctor-iCliniq --- # MedLLaMA-3.2-3B: AI Lab Report Analyzer ## Model Description This is a fine-tuned version of [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct) trained on medical Q&A data to answer patient queries about lab reports and health conditions. ## Training Details - **Base Model:** LLaMA-3.2-3B-Instruct - **Method:** QLoRA (4-bit quantization + LoRA rank 16) - **Dataset:** MedQuAD + iCliniq (~10k examples) - **Epochs:** 2 - **Hardware:** NVIDIA T4 (Google Colab) ## Intended Use - Answering patient questions about lab report values - Explaining medical terminology in plain language - Providing general health information ## ⚠️ Limitations & Disclaimer This model is for **educational and informational purposes only**. It is **NOT a substitute for professional medical advice, diagnosis, or treatment.** Always consult a qualified healthcare provider for medical decisions. ## Quick Start ```python from transformers import AutoModelForCausalLM, AutoTokenizer, BitsAndBytesConfig from peft import PeftModel import torch base_model = AutoModelForCausalLM.from_pretrained( 'meta-llama/Llama-3.2-3B-Instruct', quantization_config=BitsAndBytesConfig(load_in_4bit=True), device_map='auto' ) model = PeftModel.from_pretrained(base_model, 'jb10231/MedLLaMA-3.2-3B-LabReport') tokenizer = AutoTokenizer.from_pretrained('jb10231/MedLLaMA-3.2-3B-LabReport') ```