Nosa Daniel
Update README.md
6cd8fe8
---
base_model: unsloth/mistral-7b-instruct-v0.3-bnb-4bit
library_name: peft
pipeline_tag: text-generation
tags:
- base_model:adapter:unsloth/mistral-7b-instruct-v0.3-bnb-4bit
- lora
- sft
- transformers
- trl
- unsloth
---
# Model Card for mistral-7b-instruct-sft
## Abstract
This repository hosts a **fine‑tuned Mistral 7B‑Instruct** model that leverages **parameter‑efficient LoRA** adaptation via the **Unsloth** framework. The model is adapted for email‑security tasks using a curated phishing‑email training dataset and demonstrates state‑of‑the‑art performance (94.9 % accuracy, 93.9 % precision, 96.1 % recall).
## Model Details
### Model Description
- **Developed by:** Montimage
- **Model type:** Large Language Model (LLM)
- **Language(s):** English
- **License:** Apache‑2.0
- **Finetuned from model:** unsloth/mistral-7b-instruct-v0.3-bnb-4bit
### Model Sources
- **Repository:** https://huggingface.co/nosadaniel/mistral-7b-instruct-tuned
## How to Get Started with the Model
You can download the model using the [Hugging Face Space](https://huggingface.co/spaces/nosadaniel/mistral-7b-instruct-tuned-demo).
### Locally using Ollama
1. **Install Ollama** – https://ollama.com
2. **Pull the base model**:
```bash
ollama pull unsloth/mistral-7b-instruct-v0.3-bnb-4bit
```
3. **Create a merged‑model manifest** (no file extension) with the following content and place it in a folder of your choice:
```text
FROM mistral-7b-instruct-v0.3-bnb-4bit
ADAPTER /path/to/your/downloaded/adapter
```
Then run:
```bash
ollama create merged-model --from ./merged-model
```
4. **Run the merged model**:
```bash
ollama run merged-model
```
## Training Details
### Training Data
- **Dataset:** Phishing Email Training Dataset
- **Link:** https://huggingface.co/datasets/nosadelian/phishing-email-training-dataset
### Training Procedure
- **Fine‑tuning method:** LoRA (Low‑Rank Adaptation) via Unsloth
- **Training regime:** fp16 mixed precision
- **Epochs:** 3 (full dataset)
- **Learning rate:** 2e‑4
- **Batch size:** 32
### Model Comparison Table (selected row for this model)
| Model | Samples | Accuracy | Precision | Recall | F1‑Score | Specificity | FPR | FNR | MCC | Validity | Avg Response Time (s) | Total Input Tokens | Total Output Tokens | Avg Input Tokens | Avg Output Tokens | Quality Mean | Quality Std | Excellent (%) | Good (%) | Fair (%) | Poor (%) |
|-------|---------|----------|-----------|--------|----------|-------------|-----|-----|-----|----------|----------------------|-------------------|---------------------|------------------|-------------------|-------------|-------------|---------------|----------|----------|----------|
| **mistral-7b-instruct-sft** | 256 | 0.949 | 0.939 | 0.961 | 0.950 | 0.938 | 0.062 | 0.039 | 0.899 | 100.0 % | 18.70 | 144,247 | 72,299 | 563.5 | 282.4 | 0.951 | 0.098 | 94.9 | 0.0 | 5.1 | 0.0 |
| **mistral:7b** | 256 | 0.840 | 0.939 | 0.727 | 0.819 | 0.953 | 0.047 | 0.273 | 0.698 | 100.0 % | 12.57 | 144,247 | 76,797 | 563.5 | 300.0 | 0.850 | 0.160 | 84.0 | 0.0 | 6.6 | 9.4 |
**Comparison with Base Model (mistral:7b):** The fine‑tuned model achieves substantially higher accuracy (94.9 % vs 84.0 %), recall (96.1 % vs 72.7 %), and overall quality metrics, while maintaining comparable precision, demonstrating the effectiveness of LoRA fine‑tuning for email‑phishing detection.
### Model Performance Analysis – mistral‑7b‑instruct‑sft
- **Total Responses:** 256
- **Accuracy:** 94.9 % (243/256)
- **Valid Responses:** 100 % (256/256)
- **Average Confidence:** 0.921
#### Classification Metrics
| Metric | Value |
|--------|-------|
| **Accuracy** | 94.9 % |
| **Precision** | 93.9 % |
| **Recall** | 96.1 % |
| **F1‑Score** | 95.0 % |
| **Specificity** | 93.8 % |
#### Confusion Matrix
| | Predicted Positive | Predicted Negative |
|--|-------------------|-------------------|
| **Actual Positive** | 123 (TP) | 5 (FN) |
| **Actual Negative** | 8 (FP) | 120 (TN) |
#### Additional Metrics
- **False Positive Rate:** 6.2 %
- **False Negative Rate:** 3.9 %
- **Negative Predictive Value:** 96.0 %
- **Matthews Correlation Coefficient:** 0.899
#### Performance Insights
-**High Precision** – Low false‑positive rate, fostering user trust.
-**High Recall** – Catches the vast majority of phishing attempts, enhancing security.
-**Excellent F1‑Score** – Well‑balanced precision and recall.
-**Strong MCC** – Strong overall correlation between predictions and ground truth.
## Citation
**BibTeX:**
```bibtex
@dataset{mistral-7b-instruct-sft,
title={Fine‑tuned Mistral 7B‑Instruct model for Email Phishing Detection},
author={Montimage, Nosadaniel, Luong89},
year={2025},
publisher={Montimage}
}
```
**APA:**
Montimage, Nosakhare Daniel Ahanor, & Luong89. (2025). *Fine‑tuned Mistral 7B‑Instruct model for Email Phishing Detection*. Montimage.
## Model Card Authors
Montimage
Email Security Research Division
AI/ML Engineering Team
Cybersecurity Domain Experts
### Framework versions
- PEFT 0.17.1