| --- |
| base_model: |
| - cservan/malbert-base-cased-128k |
| license: apache-2.0 |
| language: |
| - en |
| - fr |
| pipeline_tag: text-classification |
| inference: false |
| tags: |
| - classification |
| - emails |
| - multilingual |
| - albert |
| - onnx |
| - mobile |
| - int8 |
| widget: |
| - text: "Subject: Your order has shipped\n\nBody: Your order #12345 is on its way and will arrive by Monday." |
| example_title: Transaction (EN) |
| - text: "Subject: Réunion demain\n\nBody: Salut, peut-on reporter notre réunion de 14h à 15h ? Dis-moi." |
| example_title: Personal (FR) |
| - text: "Subject: Weekly Newsletter\n\nBody: Check out our latest deals! 50% off everything this weekend." |
| example_title: Newsletter (EN) |
| - text: "Subject: Alerte de sécurité\n\nBody: Une nouvelle connexion à votre compte depuis Paris, France. Vérifiez que c'est bien vous." |
| example_title: Alert (FR) |
| --- |
| |
| # Email Classifier (mALBERT ONNX) |
|
|
| A dual-head **mALBERT** classifier for email category + action prediction, optimized for on-device inference using ONNX Runtime. Bilingual (English + French), 24M parameters, 50.7 MB after INT8 quantization. |
|
|
| ## Model Description |
|
|
| Classifies emails into 5 categories and predicts whether the recipient should take action: |
|
|
| | Category | Description | |
| |----------|-------------| |
| | **PERSONAL** | Direct 1:1 human communication, calendar invites from real people, direct messages. Excludes platform notifications. | |
| | **NEWSLETTER** | Marketing, promotions, subscribed content. Includes weekly digests, year-in-review recaps, marketing-flavored surveys with rewards. | |
| | **TRANSACTION** | Money or order events: receipts, charges, refunds, shipping confirmations with order/booking IDs, payslips, money-transfer notifications. | |
| | **ALERT** | Account, security, or infrastructure messages: password resets, login alerts, CI failures, booking-bound expiry, satisfaction surveys without rewards, named-product update notifications. | |
| | **SOCIAL** | Platform activity *between people*: post mentions, comment notifications, PR review requests from real users. Excludes automated platform mail (those are ALERT). | |
|
|
| The action flag is `true` only when the email requires a concrete response tied to something the user owns or initiated — pay to keep an existing booking, verify a code you requested, accept/decline a calendar invite, reply to a 1:1 message, security event needing verification, or a support ticket follow-up. |
|
|
| ### Output Format |
|
|
| Single forward pass producing two tensors: |
| - `category_probs`: Float32[5] — softmax probabilities per category (argmax = predicted category) |
| - `action_prob`: Float32[1] — sigmoid probability of action required (threshold 0.5) |
|
|
| No text generation, no decoder, no beam search. |
|
|
| **Example:** |
|
|
| ``` |
| Input: "Subject: Your order has shipped\n\nBody: Your order #12345 is on its way..." |
| Output: category_probs → TRANSACTION (0.94), action_prob → 0.08 (NO_ACTION) |
| ``` |
|
|
| ## Intended Use |
|
|
| - **Primary:** On-device email triage in mobile apps (iOS/Android) |
| - **Runtime:** ONNX Runtime React Native |
| - **Use case:** Prioritizing inbox, filtering noise, surfacing actionable emails |
|
|
| ## Model Details |
|
|
| | Attribute | Value | |
| |-----------|-------| |
| | Base Model | `cservan/malbert-base-cased-128k` | |
| | Parameters | ~24M | |
| | Architecture | ALBERT encoder (parameter-shared, 1 physical block × 12 virtual layers) + dual classification heads | |
| | Pooling | `pooler_output` (SOP-pretrained linear + tanh) | |
| | ONNX Size | 50.7 MB (INT8 quantized, 1.8× compression from FP32) | |
| | Max Sequence | 384 tokens | |
| | Tokenizer | SentencePiece Unigram (128K vocab, French-aware) | |
| | Hidden Size | 768 | |
| | Special Tokens | `[CLS]=2`, `[SEP]=3`, `<pad>=0`, `<unk>=1` | |
|
|
| ## Performance |
|
|
| Test set metrics (250 emails, balanced across categories, EN+FR): |
|
|
| | Metric | Score | |
| |--------|-------| |
| | **Category Accuracy** | **86.0%** (single seed) / **88.4%** (2-seed soft-vote ensemble) | |
| | **Action Accuracy** | **84.8%** | |
| | Quantization | INT8 dynamic, 20/20 PyTorch↔ONNX argmax parity | |
|
|
| ### Per-language breakdown (single seed) |
|
|
| | | English | French | |
| |---|---|---| |
| | Category accuracy | 85.4% | **87.0%** | |
| | Action accuracy | 89.2% | 77.2% | |
|
|
| Notable: French slightly outperforms English on category — the multilingual signal is symmetric. Action accuracy retains an EN advantage (~12 pts) reflecting heavier representation of EN action patterns in training data. |
|
|
| ### Per-class F1 (single seed) |
|
|
| | Class | Precision | Recall | F1 | |
| |---|---|---|---| |
| | ALERT | 0.885 | 0.900 | 0.893 | |
| | NEWSLETTER | 0.771 | 0.900 | 0.831 | |
| | PERSONAL | 0.917 | 0.892 | 0.904 | |
| | SOCIAL | 0.862 | 0.758 | 0.807 | |
| | TRANSACTION | 0.907 | 0.817 | 0.860 | |
|
|
| ## Training Data |
|
|
| - **Source:** Personal Gmail inboxes (anonymized) |
| - **Languages:** English, French |
| - **Size:** 2,005 train / 251 val / 250 test (balanced) |
| - **Labeling:** Human-annotated with category + action flag, prompt-assisted with v7 labeling rules (precise tie-breakers for booking-bound deadlines, marketing recaps with reward language, CI/security automation, curated personalized outreach, satisfaction surveys with/without incentives) |
| - **Input format:** `Subject: ...\n\nBody: ...` (no instruction prefix) |
|
|
| ## How to Use |
|
|
| ### ONNX Runtime (React Native) |
|
|
| ```typescript |
| import { InferenceSession, Tensor } from 'onnxruntime-react-native'; |
| |
| const session = await InferenceSession.create('model.onnx'); |
| |
| const outputs = await session.run({ |
| input_ids: inputIdsTensor, // int64[1, seq_len] |
| attention_mask: attentionMaskTensor, // int64[1, seq_len] |
| token_type_ids: tokenTypeIdsTensor, // int64[1, seq_len], all zeros |
| }); |
| |
| const categoryProbs = outputs.category_probs.data; // Float32[5] |
| const actionProb = outputs.action_prob.data[0]; // Float32 |
| ``` |
|
|
| ### Python (PyTorch reference) |
|
|
| ```python |
| from transformers import AutoTokenizer |
| import torch |
| |
| tokenizer = AutoTokenizer.from_pretrained("Ippoboi/malbert-email-classifier") |
| # Load DualHeadClassifier from checkpoint (see ml/scripts/train_classifier.py) |
| |
| text = "Subject: Réunion demain\n\nBody: Peut-on reporter à 15h ?" |
| inputs = tokenizer(text, return_tensors="pt", max_length=384, truncation=True) |
| |
| with torch.no_grad(): |
| cat_logits, act_logits = model(inputs["input_ids"], inputs["attention_mask"]) |
| category = ["ALERT", "NEWSLETTER", "PERSONAL", "SOCIAL", "TRANSACTION"][cat_logits.argmax()] |
| action = torch.sigmoid(act_logits).item() > 0.5 |
| ``` |
|
|
| ### ONNX Runtime (Python) |
|
|
| ```python |
| import onnxruntime as ort |
| from transformers import AutoTokenizer |
| import numpy as np |
| |
| session = ort.InferenceSession("model.onnx") |
| tokenizer = AutoTokenizer.from_pretrained("Ippoboi/malbert-email-classifier") |
| |
| inputs = tokenizer( |
| "Subject: Your order has shipped\n\nBody: ...", |
| return_tensors="np", |
| max_length=384, |
| truncation=True, |
| padding="max_length", |
| ) |
| cat_probs, act_prob = session.run( |
| ["category_probs", "action_prob"], |
| { |
| "input_ids": inputs["input_ids"].astype(np.int64), |
| "attention_mask": inputs["attention_mask"].astype(np.int64), |
| "token_type_ids": np.zeros_like(inputs["input_ids"], dtype=np.int64), |
| }, |
| ) |
| categories = ["ALERT", "NEWSLETTER", "PERSONAL", "SOCIAL", "TRANSACTION"] |
| print(categories[cat_probs[0].argmax()], "action:", act_prob[0] > 0.5) |
| ``` |
|
|
| ## Files |
|
|
| | File | Size | Description | |
| |------|------|-------------| |
| | `model.onnx` | 50.7 MB | INT8 quantized ONNX model | |
| | `tokenizer.json` | 8.2 MB | Fast tokenizer (SentencePiece Unigram, 128K vocab) | |
| | `spiece.model` | 2.3 MB | Raw SentencePiece vocab (optional, for Python reload) | |
| | `tokenizer_config.json` | 1.4 KB | Tokenizer config | |
| | `special_tokens_map.json` | 970 B | Special token names → IDs | |
|
|
| ## Architecture |
|
|
| ``` |
| Input → ALBERT Encoder (12 virtual layers × 1 shared block, hidden=768) |
| ↓ |
| pooler_output (Linear+tanh on [CLS]) |
| ↓ |
| ┌─────┴─────┐ |
| ↓ ↓ |
| Category Head Action Head |
| Linear(768→5) Linear(768→1) |
| ↓ ↓ |
| softmax sigmoid |
| ↓ ↓ |
| category_probs action_prob |
| ``` |
|
|
| ALBERT shares one physical transformer block across all 12 virtual layers. This gives ~24M total parameters (vs ~110M for an equivalent BERT-base) at the cost of representational capacity per virtual depth. |
|
|
| ## Compared to Previous Model (MiniLM v1) |
|
|
| | | MiniLM v1 | mALBERT v3 (this) | |
| |---|---|---| |
| | Base architecture | XLM-R encoder, independent layers | ALBERT, parameter-shared | |
| | Parameters | ~117M | ~24M | |
| | ONNX size | 113 MB | **50.7 MB** | |
| | Max sequence | 256 | **384** | |
| | Vocab size | 250K | 128K | |
| | Category accuracy | 92.0% | 86.0% / 88.4% (ensemble) | |
| | Action accuracy | 82.8% | **84.8%** | |
| | FR cat parity | EN-favored | **EN/FR symmetric** | |
|
|
| mALBERT v3 trades raw category accuracy for **less than half the on-device footprint**, **wider context** (384 vs 256 tokens), and **balanced multilingual performance**. Action accuracy is higher; category accuracy is lower in absolute terms but the language gap closes. |
|
|
| ## Limitations |
|
|
| - Trained on personal email patterns; may not generalize to enterprise/corporate email styles |
| - Classification accuracy depends on text quality (plain text preferred over heavy HTML) |
| - French action accuracy lags English by ~12 points; the v7 labeling prompt is EN-leaning in its action examples |
| - SOCIAL is the weakest category (F1 0.81) — smallest training class (268 examples) and shares features with NEWSLETTER for platform-mass-emails |
| - 384-token cap may truncate long emails; ~17% of training emails exceeded this limit |
| - ALBERT parameter sharing limits representational depth; for harder boundaries, a non-shared encoder (mDeBERTa-v3-base, MiniLM-L12) would have more capacity at higher inference cost |
|
|
| ## License |
|
|
| Apache 2.0 |
|
|