File size: 5,048 Bytes
983c1f2 2f97eee 983c1f2 2f97eee 983c1f2 2f97eee 983c1f2 2f97eee 983c1f2 2f97eee 983c1f2 2f97eee 983c1f2 2f97eee 983c1f2 2f97eee 983c1f2 2f97eee 983c1f2 2f97eee 983c1f2 2f97eee 983c1f2 2f97eee 983c1f2 2f97eee 983c1f2 2f97eee 983c1f2 2f97eee 983c1f2 2f97eee 983c1f2 2f97eee | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 | ---
language:
- en
license: cc-by-nc-4.0
tags:
- pharmacovigilance
- medical
- mistral
- qlora
- faers
- drug-safety
- adverse-events
base_model: mistralai/Mistral-7B-Instruct-v0.3
---
# pv-biomistral-7b
A pharmacovigilance-specialised language model fine-tuned from
[Mistral-7B-Instruct-v0.3](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.3)
on 100,000 FAERS-derived training examples across five structured PV tasks.
This is the community testing release. It contains only the Q4_K_M quantized
GGUF for local inference via Ollama or llama-cpp-python.
---
## ⚠️ Important Disclaimer
This model is a **research prototype** intended for pharmacovigilance
professionals to evaluate and provide feedback on. It is **not a validated
system** and must not be used for:
- Autonomous pharmacovigilance decision-making
- Generating or contributing to regulatory submissions
- Replacing qualified pharmacovigilance assessor judgment
- Clinical or safety-critical decisions of any kind
All model outputs require review by a qualified pharmacovigilance professional.
This tool is for exploratory and research purposes only.
---
## Model Details
| Property | Value |
|---|---|
| Base model | mistralai/Mistral-7B-Instruct-v0.3 |
| Fine-tuning method | QLoRA (4-bit NF4, LoRA r=16) |
| Training records | 100,000 |
| Training epochs | 3 |
| Data source | FAERS public database (FDA) |
| Quantization | Q4_K_M (GGUF) |
| Model size | 4.37 GB |
| Context window | 8192 tokens |
| Framework | TRL 1.0.0, Transformers, PEFT |
## Setup — Ollama (Recommended)
### Requirements
- [Ollama](https://ollama.com/download) installed
- ~5 GB free disk space
- 8 GB RAM minimum, 16 GB recommended
- GPU optional but recommended for faster inference
### Installation
**Step 1 — Download both files from this repository:**
- `pv-biomistral-7b-Q4_K_M.gguf` (4.37 GB)
- `Modelfile`
Place both in the same folder.
**Step 2 — Create the Ollama model**
```bash
cd /path/to/downloaded/files
ollama create pv-mistral-v2 -f Modelfile
```
**Step 3 — Run**
```bash
ollama run pv-mistral-v2
```
**Windows users:** Use the full path e.g. `cd C:\Users\YourName\Downloads\pv-model\`
---
## Setup — llama-cpp-python (Alternative)
```bash
pip install llama-cpp-python[server]
python -m llama_cpp.server \
--model pv-biomistral-7b-Q4_K_M.gguf \
--chat_format mistral-instruct \
--n_gpu_layers -1 \
--n_ctx 8192
```
Then open `http://localhost:8000/docs` for the Swagger UI.
---
## Setup — Jan App (Windows/Mac)
1. Download [Jan](https://jan.ai)
2. Import Model → select the GGUF file
3. Set temperature to 0.1 in chat settings
4. Add system prompt from the Modelfile SYSTEM field
---
## Expected Performance by Hardware
| Hardware | Speed | Response Time |
|---|---|---|
| Mac Mini M4 / Apple Silicon | 25-35 tokens/sec | 2-5 sec/case |
| Windows + NVIDIA GPU (8GB+ VRAM) | 25-40 tokens/sec | 2-4 sec/case |
| Snapdragon X Elite (16GB) | 8-15 tokens/sec | 5-12 sec/case |
| Windows CPU only (16-24GB RAM) | 3-6 tokens/sec | 15-30 sec/case |
---
## Known Limitations
- **Probable causality underrepresented:** Training data contained only 70 Probable
causality examples out of 100,000 records, reflecting real-world FAERS spontaneous
reporting patterns. The model may default to Possible even for cases with confirmed
positive dechallenge and no confounders.
- **Spontaneous reports only:** Trained exclusively on FAERS spontaneous adverse
event reports. Performance on clinical trial safety data, EHR-derived cases,
or non-English source material is untested.
- **Not formally validated:** The model has not been validated against any regulatory
standard including ICH E2D, ICH E2A, or WHO-UMC guidelines.
- **Short context optimised:** Designed for single-case inputs under 512 tokens.
---
## CIOMS WG XIV Alignment
This model is designed to operate within a Human-in-the-Loop (HITL) framework
consistent with CIOMS Working Group XIV recommendations for AI in drug safety.
All outputs are decision-support signals requiring human adjudication by a
qualified pharmacovigilance professional.
---
## Feedback
This is a community testing release. Please evaluate the model on real cases
from your practice area and share findings. Particular interest in:
- Causality outputs where you would classify Probable
- Cases with unusual drug combinations or rare reactions
- Narrative quality from a safety database entry perspective
- Therapeutic areas where performance appears weaker
---
## Training Data
Trained on 10,000 cases from the FDA Adverse Event Reporting System (FAERS),
accessed via public database export. No proprietary, confidential, or
patient-identifiable data beyond what is publicly available in FAERS was used.
---
## License
Base model (Mistral-7B-Instruct-v0.3): Apache 2.0
Fine-tuned weights: CC BY-NC 4.0 (non-commercial research use only)
By downloading this model you agree to use it for research purposes only
and not for any commercial application or regulatory submission.
|