Update README.md
Browse files
README.md
CHANGED
|
@@ -15,11 +15,9 @@ library_name: transformers
|
|
| 15 |
|
| 16 |
## Overview
|
| 17 |
|
| 18 |
-
With the rapid proliferation of large language models (LLMs) -- each optimized for different strengths, style, or latency/cost profile -- routing has become an essential technique to operationalize the use of different models.
|
| 19 |
|
| 20 |
-
|
| 21 |
-
We introduce a preference-aligned routing framework that guides model selection by matching queries to user-defined domains (e.g., travel) or action types (e.g., image editing) -- offering a practical mechanism to encode preferences in routing decisions.
|
| 22 |
-
Specifically, we introduce Arch-Router, a compact 1.5B model that learns to map queries to domain-action preferences for model routing decisions.
|
| 23 |
|
| 24 |
Experiments on conversational datasets demonstrate that our approach achieves state-of-the-art (SOTA) results in matching queries with human preferences, outperforming top proprietary models.
|
| 25 |
|
|
|
|
| 15 |
|
| 16 |
## Overview
|
| 17 |
|
| 18 |
+
With the rapid proliferation of large language models (LLMs) -- each optimized for different strengths, style, or latency/cost profile -- routing has become an essential technique to operationalize the use of different models. However, existing LLM routing approaches are limited in two key ways: they evaluate performance using benchmarks that often fail to capture human preferences driven by subjective evaluation criteria, and they typically select from a limited pool of models.
|
| 19 |
|
| 20 |
+
We introduce a preference-aligned routing framework that guides model selection by matching queries to user-defined domains (e.g., travel) or action types (e.g., image editing) -- offering a practical mechanism to encode preferences in routing decisions. Specifically, we introduce Arch-Router, a compact 1.5B model that learns to map queries to domain-action preferences for model routing decisions.
|
|
|
|
|
|
|
| 21 |
|
| 22 |
Experiments on conversational datasets demonstrate that our approach achieves state-of-the-art (SOTA) results in matching queries with human preferences, outperforming top proprietary models.
|
| 23 |
|