rapha / README.md
Phora68's picture
Add model card
b2a2f2e verified
metadata
language:
  - en
license: apache-2.0
tags:
  - medical
  - clinical
  - mistral
  - gguf
  - unsloth
  - lora
base_model: mistralai/Mistral-Nemo-Base-2407

๐Ÿฉบ Rapha โ€” Clinical AI Physician Assistant (GGUF)

Rapha is a clinical AI assistant fine-tuned on Mistral-Nemo-12B using Unsloth.
It performs forward-chaining medical reasoning โ€” gathering symptoms conversationally,
reasoning step by step, and escalating structured findings to a physician.

โš ๏ธ Rapha is a research prototype. It does not diagnose. All outputs must be reviewed by a qualified medical professional.


๐Ÿš€ Quickstart

Ollama

ollama run hf.co/Phora68/rapha

llama.cpp

./llama-cli -m rapha-q4_k_m.gguf \
  --chat-template mistral \
  -p "I've been having chest pain and shortness of breath for two days." \
  -n 512

Python (llama-cpp-python)

from llama_cpp import Llama

llm = Llama(
    model_path = "rapha-q4_k_m.gguf",
    n_ctx      = 2048,
    n_gpu_layers = -1,   # use all GPU layers
)

response = llm.create_chat_completion(messages=[
    {"role": "user", "content": "I've had a persistent headache for three days and I'm really worried."}
])
print(response["choices"][0]["message"]["content"])

๐Ÿ“Š Model Details

Property Value
Base model mistralai/Mistral-Nemo-Base-2407
Fine-tuning QLoRA (r=64, ฮฑ=16) via Unsloth
Quantisation Q4_K_M
Context length 2048 tokens
Training format ShareGPT
Chat template Mistral [INST]
Domain Clinical / Medical triage
Dataset 200,000 samples (170k train / 20k val / 10k test)

โš ๏ธ Limitations

  • Not a medical device
  • Does not provide diagnoses
  • Must be reviewed by a qualified clinician
  • Not validated for clinical deployment