| | ---
|
| | language:
|
| | - en
|
| | pipeline_tag: text-classification
|
| | tags:
|
| | - sms-spam
|
| | - phishing-detection
|
| | - scam-detection
|
| | - security
|
| | metrics:
|
| | - f1
|
| | - accuracy
|
| | - precision
|
| | - recall
|
| | widget:
|
| | - text: "Your account is blocked! Verify immediately with OTP. Send money to scam@ybl using https://scam.xyz/"
|
| | example_title: "Bank KYC Scam"
|
| | - text: "Congratulations! You won Rs 50,000 lottery prize. Contact urgently to claim via link: http://bit.ly/claim"
|
| | example_title: "Lottery Scam"
|
| | - text: "Hey, are we still meeting for lunch tomorrow at 12?"
|
| | example_title: "Safe Message"
|
| | ---
|
| |
|
| | # SCAMBERT: DistilBERT for SMS Fraud & Scam Detection
|
| |
|
| | SCAMBERT is a fine-tuned `distilbert-base-uncased` model specifically designed to detect social engineering, financial fraud, phishing, and scam payloads in SMS and short-form conversational text. It is built as Layer 3 of the AI Honeypot (CIPHER) Threat Intelligence Pipeline.
|
| |
|
| | ## Model Summary
|
| |
|
| | - **Model Type:** Text Classification (Binary)
|
| | - **Base Model:** `distilbert-base-uncased`
|
| | - **Language:** English (en)
|
| | - **Task:** Spam/Scam Detection
|
| | - **License:** MIT (or your project's license)
|
| | - **Size:** ~255 MB
|
| |
|
| | ### Labels
|
| |
|
| | - `0`: Safe / Legitimate
|
| | - `1`: Scam / Fraud / Phishing
|
| |
|
| | ## Performance & Metrics
|
| |
|
| | The model was fine-tuned on a dataset of **8,438** samples (27.5% Scam / 72.5% Safe). Due to class imbalance, class weights were applied during training.
|
| |
|
| | ### Calibration & Validation Results
|
| |
|
| | - **Best Accuracy:** 99.41%
|
| | - **Best F1-Score:** 98.92%
|
| | - **Calibrated Precision:** 95.08%
|
| | - **Calibrated Recall:** 100.0%
|
| | - **Optimal Threshold:** `0.0028` (For high-recall environments)
|
| |
|
| | ### Robustness Evaluation
|
| |
|
| | The model was tested against common bad-actor obfuscation tactics:
|
| |
|
| | | Tactic | Example Input | Prediction Probability | Passed |
|
| | | :--- | :--- | :--- | :--- |
|
| | | **URL Obfuscation** | `Win $1000 fast! Click hxxp://scammy...` | 99.9% Scam | ✅ |
|
| | | **Numeric Substitution** | `W1NNER! Y0u have b33n select3d...` | 99.3% Scam | ✅ |
|
| | | **Mixed Case** | `cOnGrAtUlAtIoNs, yOu WoN a FrEe...` | 89.8% Scam | ✅ |
|
| |
|
| | *Note: The model occasionally struggles with extremely short, contextless messages (e.g., "Call me now") as intended, relying on earlier heuristic layers for context.*
|
| |
|
| | ## Usage
|
| |
|
| | You can use this model directly with Hugging Face's `pipeline`:
|
| |
|
| | ```python
|
| | from transformers import pipeline
|
| |
|
| | # Load the pipeline
|
| | classifier = pipeline("text-classification", model="Digvijay05/SCAMBERT")
|
| |
|
| | # Inference
|
| | text = "Earn Rs 5000 daily income from home part time. Click this link: http://bit.ly/job"
|
| | result = classifier(text)
|
| |
|
| | print(result)
|
| | # [{'label': 'LABEL_1', 'score': 0.99...}]
|
| | ```
|
| |
|
| | Or run via the **Inference API**:
|
| |
|
| | ```python
|
| | import httpx
|
| |
|
| | API_URL = "https://api-inference.huggingface.co/models/Digvijay05/SCAMBERT"
|
| | headers = {"Authorization": "Bearer YOUR_HF_TOKEN"}
|
| |
|
| | response = httpx.post(API_URL, headers=headers, json={"inputs": "Your account is locked. Verify at bit.ly/secure"})
|
| | print(response.json())
|
| | ```
|
| |
|
| | ## Deployment Considerations
|
| |
|
| | - **CPU Latency Estimate:** ~10-30ms / sequence
|
| | - **GPU Latency Estimate:** ~2-5ms / sequence
|
| | - **Recommendation:** Can be efficiently hosted on serverless CPU environments (like Render Free Tier) using Hugging Face's Inference API, or deployed natively if 512MB+ RAM is available. ONNX quantization is recommended for edge deployments.
|
| |
|
| | ## Intended Use
|
| |
|
| | This model is designed as a *semantic booster* tie-breaker layer within a multi-layered classification engine. It excels at detecting complex sentence structures, urgency, and manipulative context that standard Regex/Heuristic rules might miss.
|
| |
|