Model Card for mistral-7b-instruct-sft

Abstract

This repository hosts a fine‑tuned Mistral 7B‑Instruct model that leverages parameter‑efficient LoRA adaptation via the Unsloth framework. The model is adapted for email‑security tasks using a curated phishing‑email training dataset and demonstrates state‑of‑the‑art performance (94.9 % accuracy, 93.9 % precision, 96.1 % recall).

Model Details

Model Description

  • Developed by: Montimage
  • Model type: Large Language Model (LLM)
  • Language(s): English
  • License: Apache‑2.0
  • Finetuned from model: unsloth/mistral-7b-instruct-v0.3-bnb-4bit

Model Sources

How to Get Started with the Model

You can download the model using the Hugging Face Space.

Locally using Ollama

  1. Install Ollamahttps://ollama.com
  2. Pull the base model:
    ollama pull unsloth/mistral-7b-instruct-v0.3-bnb-4bit
    
  3. Create a merged‑model manifest (no file extension) with the following content and place it in a folder of your choice:
    FROM mistral-7b-instruct-v0.3-bnb-4bit
    ADAPTER /path/to/your/downloaded/adapter
    
    Then run:
    ollama create merged-model --from ./merged-model
    
  4. Run the merged model:
    ollama run merged-model
    

Training Details

Training Data

Training Procedure

  • Fine‑tuning method: LoRA (Low‑Rank Adaptation) via Unsloth
  • Training regime: fp16 mixed precision
  • Epochs: 3 (full dataset)
  • Learning rate: 2e‑4
  • Batch size: 32

Model Comparison Table (selected row for this model)

Model Samples Accuracy Precision Recall F1‑Score Specificity FPR FNR MCC Validity Avg Response Time (s) Total Input Tokens Total Output Tokens Avg Input Tokens Avg Output Tokens Quality Mean Quality Std Excellent (%) Good (%) Fair (%) Poor (%)
mistral-7b-instruct-sft 256 0.949 0.939 0.961 0.950 0.938 0.062 0.039 0.899 100.0 % 18.70 144,247 72,299 563.5 282.4 0.951 0.098 94.9 0.0 5.1 0.0
mistral:7b 256 0.840 0.939 0.727 0.819 0.953 0.047 0.273 0.698 100.0 % 12.57 144,247 76,797 563.5 300.0 0.850 0.160 84.0 0.0 6.6 9.4

Comparison with Base Model (mistral:7b): The fine‑tuned model achieves substantially higher accuracy (94.9 % vs 84.0 %), recall (96.1 % vs 72.7 %), and overall quality metrics, while maintaining comparable precision, demonstrating the effectiveness of LoRA fine‑tuning for email‑phishing detection.

Model Performance Analysis – mistral‑7b‑instruct‑sft

  • Total Responses: 256
  • Accuracy: 94.9 % (243/256)
  • Valid Responses: 100 % (256/256)
  • Average Confidence: 0.921

Classification Metrics

Metric Value
Accuracy 94.9 %
Precision 93.9 %
Recall 96.1 %
F1‑Score 95.0 %
Specificity 93.8 %

Confusion Matrix

Predicted Positive Predicted Negative
Actual Positive 123 (TP) 5 (FN)
Actual Negative 8 (FP) 120 (TN)

Additional Metrics

  • False Positive Rate: 6.2 %
  • False Negative Rate: 3.9 %
  • Negative Predictive Value: 96.0 %
  • Matthews Correlation Coefficient: 0.899

Performance Insights

  • High Precision – Low false‑positive rate, fostering user trust.
  • High Recall – Catches the vast majority of phishing attempts, enhancing security.
  • Excellent F1‑Score – Well‑balanced precision and recall.
  • Strong MCC – Strong overall correlation between predictions and ground truth.

Citation

BibTeX:

@dataset{mistral-7b-instruct-sft,
  title={Fine‑tuned Mistral 7B‑Instruct model for Email Phishing Detection},
  author={Montimage, Nosadaniel, Luong89},
  year={2025},
  publisher={Montimage}
}

APA:

Montimage, Nosakhare Daniel Ahanor, & Luong89. (2025). Fine‑tuned Mistral 7B‑Instruct model for Email Phishing Detection. Montimage.

Model Card Authors

Montimage Email Security Research Division AI/ML Engineering Team Cybersecurity Domain Experts

Framework versions

  • PEFT 0.17.1
Downloads last month
-
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for nosadaniel/mistral-7b-instruct-tuned