SwarmMed-7B-v5 — Platinum-Verified Clinical Medical Model
The model that proved quality beats size. SwarmMed-7B-v5 outperforms a fine-tuned 14B model on clinical translation, syndrome naming, and urgency calibration — trained exclusively on CoVe-verified platinum data.
Built by Swarm & Bee on sovereign GPU infrastructure. Your data never leaves your rack.
Key Results
| Metric | Base Qwen2.5-7B | V3 (Gold) | V4 (Platinum) | V5 (Full Platinum) |
|---|---|---|---|---|
| Concept Coverage | 65.3% | 40.7% | 39.9% | 52.5% |
| Guidelines Cited | 47% | 35% | 47% | 82% |
| Specific Numbers | 82% | 82% | 65% | 94% |
| Disclaimer Rate | 53% | 94% | 88% | 94% |
Head-to-head: V5 beats V4 (7-4), V5 beats V3 (9-5), V5 beats fine-tuned 14B on clinical safety (20-9-6).
What Makes This Different
Platinum Verification Pipeline
Every training pair passed through a 4-stage quality gate + independent fact-checking:
- Clinical Verification — Is the scenario medically plausible?
- Accuracy Scoring — 5-dimension rubric (dosing, contraindications, guidelines, risk scores, treatment)
- Safety Check — Dangerous advice detection, disclaimer enforcement
- CoVe Fact-Check — Chain-of-Verification with Qwen3-235B. Every factual claim independently verified by a 235B parameter model.
Pairs that passed all 4 stages = PASS (platinum by verification). Pairs that flagged on CoVe = rewritten by 235B using only verified facts = platinum by construction.
Result: 1,191 platinum pairs. Zero hallucinations. Zero unverified claims.
235B Rewrite Pipeline
For pairs where CoVe found factual issues:
- Extract independently verified facts from 235B responses
- Have 235B write a new clinical answer using ONLY verified facts
- 100% conversion rate — every rewrite is platinum by construction
This is not prompt engineering. This is verified knowledge distillation from a 235B teacher into a 7B student, with every fact independently confirmed.
Model Details
| Parameter | Value |
|---|---|
| Base Model | Qwen/Qwen2.5-7B-Instruct |
| Method | LoRA r=64, alpha=128 |
| Training Data | 1,191 platinum pairs (876 pharmacology + 315 cardiology) |
| Data Sources | 284 CoVe PASS + 935 235B rewrites (from 618 FLAG + 425 silver) |
| Epochs | 3 |
| Loss Curve | 1.892 → 0.463 → 0.236 → 0.198 |
| Training Time | ~28 minutes |
| Hardware | 2x NVIDIA RTX 3090 Ti |
| Quantization | 4-bit (BitsAndBytes NF4) during training |
| License | Apache 2.0 |
Specialties Covered
- Pharmacology (876 pairs): Drug interactions, dosing, adverse effects, contraindications, monitoring, pharmacokinetics
- Cardiology (315 pairs): ACS management, heart failure (HFrEF/HFpEF), arrhythmias, GDMT, risk stratification
Usage
from transformers import AutoModelForCausalLM, AutoTokenizer
model_name = "SwarmOS/SwarmMed-7B-v5-merged"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(
model_name,
torch_dtype="auto",
device_map="auto"
)
messages = [
{"role": "system", "content": "You are a clinical medical assistant. Provide evidence-based responses citing current guidelines. Always include appropriate disclaimers. Name clinical syndromes explicitly. Provide time-critical action windows when relevant."},
{"role": "user", "content": "A 58-year-old male presents with acute onset crushing chest pain radiating to the left arm, diaphoresis, and ST elevation in leads II, III, and aVF. Troponin is elevated. What is the diagnosis and immediate management?"}
]
text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = tokenizer(text, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=1024, temperature=0.7)
response = tokenizer.decode(outputs[0][inputs["input_ids"].shape[1]:], skip_special_tokens=True)
print(response)
With vLLM (Recommended for Production)
python -m vllm.entrypoints.openai.api_server \
--model SwarmOS/SwarmMed-7B-v5-merged \
--port 8000 \
--max-model-len 4096
from openai import OpenAI
client = OpenAI(base_url="http://localhost:8000/v1", api_key="na")
response = client.chat.completions.create(
model="SwarmOS/SwarmMed-7B-v5-merged",
messages=[
{"role": "system", "content": "You are a clinical medical assistant..."},
{"role": "user", "content": "Your clinical question here"}
],
temperature=0.7,
max_tokens=1024
)
Training Data
This model was trained on the SwarmMed-Platinum-1K dataset — 1,191 CoVe-verified clinical QA pairs.
Every pair includes:
- Detailed clinical scenario with patient demographics
- Evidence-based response citing current guidelines (ACC/AHA, ESC, ACCP, etc.)
- Chain-of-thought clinical reasoning
- Appropriate disclaimers and safety language
- Specific dosing, monitoring parameters, and time-critical action windows
Limitations
- Not a medical device. This model is for research and educational purposes. It should not be used for clinical decision-making without physician oversight.
- Specialty coverage: Currently strongest in pharmacology and cardiology. Other specialties (neurology, oncology, emergency medicine) have limited representation.
- MCQ tradeoff: Fine-tuning on long-form clinical QA optimizes for detailed clinical explanations at the cost of standardized test (MCQ) performance. MedQA scores may be slightly lower than base model.
- English only: Training data is exclusively in English.
- Knowledge cutoff: Training data reflects guidelines and evidence available as of February 2026.
The Platinum Standard
We believe the future of medical AI is not bigger models — it's better data. A 7B model trained on 1,191 verified pairs outperforms a 14B trained on 5,009 unverified pairs on clinical safety metrics. Quality of signal beats quantity of parameters.
Every pair in our platinum vault has had every factual claim independently verified by a 235B parameter model. We don't ship slop.
About Swarm & Bee
Swarm & Bee builds sovereign compute infrastructure for regulated industries. Last mile intelligence — specialized models running on your hardware, with your data, under your control.
- Website: swarmandbee.com
- HuggingFace: SwarmOS
- ERC-8004 Agent: ID #17493 on Base Mainnet
Citation
@misc{swarmmed7bv5,
title={SwarmMed-7B-v5: Platinum-Verified Clinical Medical Model},
author={Swarm and Bee},
year={2026},
url={https://huggingface.co/SwarmOS/SwarmMed-7B-v5-merged},
note={CoVe-verified with Qwen3-235B, trained on 1,191 platinum pairs}
}
- Downloads last month
- 6
Model tree for SwarmOS/SwarmMed-7B-v5-merged
Dataset used to train SwarmOS/SwarmMed-7B-v5-merged
Evaluation results
- Concept Coverage (17-Q Clinical Eval)self-reported52.500
- Guidelines Cited Rateself-reported82.000
- Disclaimer Rateself-reported94.000