MALM: Modular Adapter-based Language Model
๐ Read the full paper (MALM.pdf)
๐ Author: Hilal Limo (Independent Researcher, 15)
๐ License: Apache-2.0
Overview
This repository contains the research paper MALM: Modular Adapter-based Language Model, which introduces a lightweight and scalable framework for multilingual AI.
Instead of relying on massive monolithic models, MALM separates reasoning and translation into two modular parts:
- Core Language Model (CLM): A compact, English-focused reasoning engine.
- Specialized Translation Adapters (STAs): Lightweight, swappable neural machine translation models.
- Orchestration Layer: Connects the pieces, parsing delegation tokens (e.g.
<to:de> ... </to>) and routing requests to the right adapter.
This design drastically reduces compute cost, makes it easier to add new languages, and is especially useful for small models, edge devices, and research settings.
Why MALM?
- ๐ Efficiency: Keep one reasoning core small and sharp.
- ๐ Scalability: Add or update languages by swapping STAs.
- ๐ ๏ธ Maintainability: Upgrade individual adapters without retraining the whole system.
- ๐ฑ Small Models: Perfect for low-resource environments, edge devices, and startups.
Example Conversation Flows
User: Translate "my name is Adam" into German.
CLM โ <to:de> my name is Adam </to>
STA โ "Mein Name ist Adam"
User (in Spanish): "ยฟCuรกnto es 12 + 7?"
Input STA (esโen) โ "How much is 12 + 7?"
CLM โ "The answer is <to:es> 19 </to>"
Output STA โ "La respuesta es 19"