File size: 1,679 Bytes
8856078 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 |
# MALM: Modular Adapter-based Language Model
๐ [Read the full paper (MALM.pdf)](./MALM.pdf)
๐ Author: **Hilal Limo (Independent Researcher, 15)**
๐ License: [Apache-2.0](./LICENSE)
---
## Overview
This repository contains the research paper **MALM: Modular Adapter-based Language Model**, which introduces a lightweight and scalable framework for multilingual AI.
Instead of relying on massive monolithic models, MALM separates **reasoning** and **translation** into two modular parts:
- **Core Language Model (CLM):** A compact, English-focused reasoning engine.
- **Specialized Translation Adapters (STAs):** Lightweight, swappable neural machine translation models.
- **Orchestration Layer:** Connects the pieces, parsing delegation tokens (e.g. `<to:de> ... </to>`) and routing requests to the right adapter.
This design drastically reduces compute cost, makes it easier to add new languages, and is especially useful for **small models**, edge devices, and research settings.
---
## Why MALM?
- ๐ **Efficiency:** Keep one reasoning core small and sharp.
- ๐ **Scalability:** Add or update languages by swapping STAs.
- ๐ ๏ธ **Maintainability:** Upgrade individual adapters without retraining the whole system.
- ๐ฑ **Small Models:** Perfect for low-resource environments, edge devices, and startups.
---
## Example Conversation Flows
```text
User: Translate "my name is Adam" into German.
CLM โ <to:de> my name is Adam </to>
STA โ "Mein Name ist Adam"
User (in Spanish): "ยฟCuรกnto es 12 + 7?"
Input STA (esโen) โ "How much is 12 + 7?"
CLM โ "The answer is <to:es> 19 </to>"
Output STA โ "La respuesta es 19"
|