--- language: - en license: apache-2.0 tags: - medical - clinical - mistral - gguf - unsloth - lora base_model: mistralai/Mistral-Nemo-Base-2407 --- # 🩺 Rapha — Clinical AI Physician Assistant (GGUF) **Rapha** is a clinical AI assistant fine-tuned on **Mistral-Nemo-12B** using Unsloth. It performs forward-chaining medical reasoning — gathering symptoms conversationally, reasoning step by step, and escalating structured findings to a physician. > ⚠️ Rapha is a research prototype. It does not diagnose. All outputs must be reviewed by a qualified medical professional. --- ## 🚀 Quickstart ### Ollama ```bash ollama run hf.co/Phora68/rapha ``` ### llama.cpp ```bash ./llama-cli -m rapha-q4_k_m.gguf \ --chat-template mistral \ -p "I've been having chest pain and shortness of breath for two days." \ -n 512 ``` ### Python (llama-cpp-python) ```python from llama_cpp import Llama llm = Llama( model_path = "rapha-q4_k_m.gguf", n_ctx = 2048, n_gpu_layers = -1, # use all GPU layers ) response = llm.create_chat_completion(messages=[ {"role": "user", "content": "I've had a persistent headache for three days and I'm really worried."} ]) print(response["choices"][0]["message"]["content"]) ``` --- ## 📊 Model Details | Property | Value | |---|---| | **Base model** | `mistralai/Mistral-Nemo-Base-2407` | | **Fine-tuning** | QLoRA (r=64, α=16) via Unsloth | | **Quantisation** | Q4_K_M | | **Context length** | 2048 tokens | | **Training format** | ShareGPT | | **Chat template** | Mistral `[INST]` | | **Domain** | Clinical / Medical triage | | **Dataset** | 200,000 samples (170k train / 20k val / 10k test) | --- ## ⚠️ Limitations - Not a medical device - Does not provide diagnoses - Must be reviewed by a qualified clinician - Not validated for clinical deployment