multi-lingual
Collection
multi-lingual mixture-of-models
•
10 items
•
Updated
A multilingual intent classification model based on mmBERT (Multilingual ModernBERT) with LoRA adapters for efficient inference.
This model classifies text into 14 MMLU-Pro academic categories using a LoRA-enhanced mmBERT backbone. It supports 1800+ languages through mmBERT's multilingual pretraining.
| Metric | Score |
|---|---|
| Accuracy | 77.9% |
| F1 (weighted) | 78.0% |
| Training Time | 139 seconds (MI300X GPU) |
from peft import PeftModel
from transformers import AutoModelForSequenceClassification, AutoTokenizer
# Load base model and LoRA adapter
base_model = AutoModelForSequenceClassification.from_pretrained(
"jhu-clsp/mmBERT-base",
num_labels=14
)
model = PeftModel.from_pretrained(base_model, "llm-semantic-router/mmbert-intent-classifier-lora")
tokenizer = AutoTokenizer.from_pretrained("jhu-clsp/mmBERT-base")
# Classify
text = "What are the legal requirements for forming a corporation?"
inputs = tokenizer(text, return_tensors="pt", truncation=True, max_length=512)
outputs = model(**inputs)
predicted_class = outputs.logits.argmax(-1).item()
This model supports cross-lingual transfer:
This model is part of the vLLM Semantic Router project - a Mixture-of-Models (MoM) router that understands request intent.
@misc{mmbert-intent-classifier,
author = {vLLM Semantic Router Team},
title = {mmBERT Intent Classifier with LoRA},
year = {2026},
publisher = {Hugging Face},
url = {https://huggingface.co/llm-semantic-router/mmbert-intent-classifier-lora}
}
Apache 2.0
Base model
jhu-clsp/mmBERT-base