Improve model card: Add pipeline tag, library name, and correct GitHub link
#1
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -1,12 +1,14 @@
|
|
| 1 |
---
|
| 2 |
license: apache-2.0
|
|
|
|
|
|
|
| 3 |
---
|
| 4 |
|
| 5 |
<h1 align="center">Metis-HOME: Hybrid Optimized Mixture-of-Experts for Multimodal Reasoning</h1>
|
| 6 |
|
| 7 |
<h5 align="center">
|
| 8 |
|
| 9 |
-
[](https://arxiv.org/pdf/2510.20519) <a href='https://huggingface.co/mmthinking/Metis-HOME'><img src='https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face%20-models-blue'></a> [](https://arxiv.org/pdf/2510.20519) <a href='https://huggingface.co/mmthinking/Metis-HOME'><img src='https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face%20-models-blue'></a> [](https://github.com/MM-Thinking/Metis-HOME)
|
| 12 |
|
| 13 |
</h5>
|
| 14 |
|
|
|
|
| 16 |
## π‘ Overview
|
| 17 |
Current multimodal reasoning models face a critical dilemma: they often "overthink" on simple tasks (inefficiency) and suffer from general capability degradation when optimized for reasoning.
|
| 18 |
|
| 19 |
+
We introduce **Metis-HOME** (**H**ybrid **O**ptimized **M**ixture-of-**E**xperts), a novel framework that enables a "Hybrid Thinking" paradigm. By structuring the original dense model (Qwen2.5-VL-7B) into two distinct expert branches: a Thinking Expert for complex reasoning, and a Non-Thinking Expert for rapid inference, controlled by a lightweight router, Metis-HOME effectively resolves the reasoning-vs-generalization trade-off.
|
| 20 |
|
| 21 |
<div style="display: flex; justify-content: center; gap: 20px; flex-wrap: wrap;">
|
| 22 |
<img src="https://raw.githubusercontent.com/MM-Thinking/Metis-HOME/main/assets/framework.png" alt="Metis-RISE Framework Overview" style="width:400px; max-width:100%;">
|
|
|
|
| 38 |
|
| 39 |
### Thinking Ratio
|
| 40 |
As shown in the following figure, the **thinking ratio** analysis of Metis-HOME reveals adaptive routing behavior:
|
| 41 |
+
- **High ratios (78%β98%)** on reasoning-heavy benchmarks (*WeMath*, *MathVision*, etc.), indicating effective use of the *thinking expert* for multi-step inference.
|
| 42 |
+
- **Low ratios (2%β5%)** on general benchmarks (*MMBench*, *OCRBench*), showing preference for the *non-thinking expert*.
|
| 43 |
|
| 44 |
This aligns with our design: **deliberate reasoning for complex tasks**, **fast inference for simple ones**, optimizing computational efficiency.
|
| 45 |
|