HuaminChen's picture
Upload mmBERT LoRA adapter: mmBERT Intent Classifier with LoRA - 14-class MMLU-Pro classification (77.9% accuracy)
b6e5d70 verified
metadata
license: apache-2.0
language:
  - en
  - zh
  - multilingual
library_name: peft
base_model: jhu-clsp/mmBERT-base
tags:
  - text-classification
  - intent-classification
  - mmbert
  - lora
  - multilingual
  - vllm-semantic-router
datasets:
  - TIGER-Lab/MMLU-Pro
metrics:
  - accuracy
  - f1
pipeline_tag: text-classification

mmBERT Intent Classifier (LoRA Adapter)

A multilingual intent classification model based on mmBERT (Multilingual ModernBERT) with LoRA adapters for efficient inference.

Model Description

This model classifies text into 14 MMLU-Pro academic categories using a LoRA-enhanced mmBERT backbone. It supports 1800+ languages through mmBERT's multilingual pretraining.

Categories

  • biology, business, chemistry, computer science, economics
  • engineering, health, history, law, math
  • other, philosophy, physics, psychology

Performance

Metric Score
Accuracy 77.9%
F1 (weighted) 78.0%
Training Time 139 seconds (MI300X GPU)

Training Details

  • Base Model: jhu-clsp/mmBERT-base
  • LoRA Rank: 32
  • LoRA Alpha: 64
  • Trainable Parameters: 6.8M / 314M (2.2%)
  • Epochs: 10
  • Batch Size: 64
  • Learning Rate: 2e-5
  • Dataset: TIGER-Lab/MMLU-Pro (9,144 samples)

Usage

from peft import PeftModel
from transformers import AutoModelForSequenceClassification, AutoTokenizer

# Load base model and LoRA adapter
base_model = AutoModelForSequenceClassification.from_pretrained(
    "jhu-clsp/mmBERT-base",
    num_labels=14
)
model = PeftModel.from_pretrained(base_model, "llm-semantic-router/mmbert-intent-classifier-lora")
tokenizer = AutoTokenizer.from_pretrained("jhu-clsp/mmBERT-base")

# Classify
text = "What are the legal requirements for forming a corporation?"
inputs = tokenizer(text, return_tensors="pt", truncation=True, max_length=512)
outputs = model(**inputs)
predicted_class = outputs.logits.argmax(-1).item()

Multilingual Support

This model supports cross-lingual transfer:

  • Fine-tuned on English MMLU-Pro data
  • Can classify queries in 1800+ languages
  • Best performance on English, good transfer to Chinese, Spanish, French, German, etc.

Part of vLLM Semantic Router

This model is part of the vLLM Semantic Router project - a Mixture-of-Models (MoM) router that understands request intent.

Citation

@misc{mmbert-intent-classifier,
  author = {vLLM Semantic Router Team},
  title = {mmBERT Intent Classifier with LoRA},
  year = {2026},
  publisher = {Hugging Face},
  url = {https://huggingface.co/llm-semantic-router/mmbert-intent-classifier-lora}
}

License

Apache 2.0