Improve model card: Add pipeline tag, library name, and correct GitHub link

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +6 -4
README.md CHANGED
@@ -1,12 +1,14 @@
1
  ---
2
  license: apache-2.0
 
 
3
  ---
4
 
5
  <h1 align="center">Metis-HOME: Hybrid Optimized Mixture-of-Experts for Multimodal Reasoning</h1>
6
 
7
  <h5 align="center">
8
 
9
- [![arXiv](https://img.shields.io/badge/Arxiv-2510.20519-b31b1b.svg?logo=arXiv)](https://arxiv.org/pdf/2510.20519)&ensp;<a href='https://huggingface.co/mmthinking/Metis-HOME'><img src='https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face%20-models-blue'></a>&ensp;[![Code License](https://img.shields.io/badge/License-Apache_2.0-green.svg)](https://github.com/tatsu-lab/stanford_alpaca/blob/main/LICENSE)
10
 
11
  </h5>
12
 
@@ -14,7 +16,7 @@ license: apache-2.0
14
  ## πŸ’‘ Overview
15
  Current multimodal reasoning models face a critical dilemma: they often "overthink" on simple tasks (inefficiency) and suffer from general capability degradation when optimized for reasoning.
16
 
17
- We introduce **Metis-HOME** (**H**ybrid **O**ptimized **M**ixture-of-**E**xperts), a novel framework that enables a "Hybrid Thinking" paradigm. By structuring the original dense model (Qwen2.5-VL-7B) into two distinct expert branchesβ€”a Thinking Expert for complex reasoning and a Non-Thinking Expert for rapid inferenceβ€”controlled by a lightweight router, Metis-HOME effectively resolves the reasoning-vs-generalization trade-off.
18
 
19
  <div style="display: flex; justify-content: center; gap: 20px; flex-wrap: wrap;">
20
  <img src="https://raw.githubusercontent.com/MM-Thinking/Metis-HOME/main/assets/framework.png" alt="Metis-RISE Framework Overview" style="width:400px; max-width:100%;">
@@ -36,8 +38,8 @@ We introduce **Metis-HOME** (**H**ybrid **O**ptimized **M**ixture-of-**E**xperts
36
 
37
  ### Thinking Ratio
38
  As shown in the following figure, the **thinking ratio** analysis of Metis-HOME reveals adaptive routing behavior:
39
- - **High ratios (78\%–98\%)** on reasoning-heavy benchmarks (*WeMath*, *MathVision*, etc.), indicating effective use of the *thinking expert* for multi-step inference.
40
- - **Low ratios (2\%–5\%)** on general benchmarks (*MMBench*, *OCRBench*), showing preference for the *non-thinking expert*.
41
 
42
  This aligns with our design: **deliberate reasoning for complex tasks**, **fast inference for simple ones**, optimizing computational efficiency.
43
 
 
1
  ---
2
  license: apache-2.0
3
+ pipeline_tag: image-text-to-text
4
+ library_name: transformers
5
  ---
6
 
7
  <h1 align="center">Metis-HOME: Hybrid Optimized Mixture-of-Experts for Multimodal Reasoning</h1>
8
 
9
  <h5 align="center">
10
 
11
+ [![arXiv](https://img.shields.io/badge/Arxiv-2510.20519-b31b1b.svg?logo=arXiv)](https://arxiv.org/pdf/2510.20519)&ensp;<a href='https://huggingface.co/mmthinking/Metis-HOME'><img src='https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face%20-models-blue'></a>&ensp;[![Code](https://img.shields.io/badge/GitHub-Code-blue.svg?logo=github)](https://github.com/MM-Thinking/Metis-HOME)
12
 
13
  </h5>
14
 
 
16
  ## πŸ’‘ Overview
17
  Current multimodal reasoning models face a critical dilemma: they often "overthink" on simple tasks (inefficiency) and suffer from general capability degradation when optimized for reasoning.
18
 
19
+ We introduce **Metis-HOME** (**H**ybrid **O**ptimized **M**ixture-of-**E**xperts), a novel framework that enables a "Hybrid Thinking" paradigm. By structuring the original dense model (Qwen2.5-VL-7B) into two distinct expert branches: a Thinking Expert for complex reasoning, and a Non-Thinking Expert for rapid inference, controlled by a lightweight router, Metis-HOME effectively resolves the reasoning-vs-generalization trade-off.
20
 
21
  <div style="display: flex; justify-content: center; gap: 20px; flex-wrap: wrap;">
22
  <img src="https://raw.githubusercontent.com/MM-Thinking/Metis-HOME/main/assets/framework.png" alt="Metis-RISE Framework Overview" style="width:400px; max-width:100%;">
 
38
 
39
  ### Thinking Ratio
40
  As shown in the following figure, the **thinking ratio** analysis of Metis-HOME reveals adaptive routing behavior:
41
+ - **High ratios (78%–98%)** on reasoning-heavy benchmarks (*WeMath*, *MathVision*, etc.), indicating effective use of the *thinking expert* for multi-step inference.
42
+ - **Low ratios (2%–5%)** on general benchmarks (*MMBench*, *OCRBench*), showing preference for the *non-thinking expert*.
43
 
44
  This aligns with our design: **deliberate reasoning for complex tasks**, **fast inference for simple ones**, optimizing computational efficiency.
45