Add metadata and link to paper
#2
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -1,3 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
<div align="center">
|
| 2 |
<h1>
|
| 3 |
Yuan 3.0 Multimodal Foundation Model
|
|
@@ -33,7 +39,7 @@
|
|
| 33 |
|
| 34 |
## 1. Introduction
|
| 35 |
|
| 36 |
-
Yuan 3.0 Flash, developed by the **YuanLab.ai team**, is a **40B parameter multimodal foundation model** that employs a Mixture of Experts (MoE) architecture, activating only approximately **3.7B parameters** per inference. Through innovative reinforcement learning training methods (RAPO), it significantly reduces inference token consumption while improving reasoning accuracy, exploring the innovative path of "less computation, higher intelligence" for large language models. We have also released the
|
| 37 |
|
| 38 |
<div align="center">
|
| 39 |
<img src="https://huggingface.co/YuanLabAI/Yuan3.0-Flash-4bit/resolve/main/docs/Yuan3.0-architecture.png" width="80%" />
|
|
@@ -55,7 +61,7 @@ Yuan 3.0 Flash outperforms GPT-5.1 in enterprise-grade RAG, multimodal retrieval
|
|
| 55 |
|
| 56 |
<div align="center">
|
| 57 |
<img src="https://huggingface.co/YuanLabAI/Yuan3.0-Flash-4bit/resolve/main/docs/Yuan3.0-benchmarks.png" width="80%" />
|
| 58 |
-
Fig.
|
| 59 |
</div>
|
| 60 |
|
| 61 |
|
|
@@ -177,4 +183,5 @@ Summarization generation is a core requirement for historical information compre
|
|
| 177 |
| **OpenAI GPT-5.1** | 49.44 | 27.48 | 10.16 | 84.63 | 40.50 |
|
| 178 |
| **Yuan3.0 Flash** | **59.31** | 51.32 | 28.32 | 89.99 | 45.34 |
|
| 179 |
|
| 180 |
-
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: other
|
| 3 |
+
library_name: transformers
|
| 4 |
+
pipeline_tag: image-text-to-text
|
| 5 |
+
---
|
| 6 |
+
|
| 7 |
<div align="center">
|
| 8 |
<h1>
|
| 9 |
Yuan 3.0 Multimodal Foundation Model
|
|
|
|
| 39 |
|
| 40 |
## 1. Introduction
|
| 41 |
|
| 42 |
+
Yuan 3.0 Flash, developed by the **YuanLab.ai team**, is a **40B parameter multimodal foundation model** that employs a Mixture of Experts (MoE) architecture, activating only approximately **3.7B parameters** per inference. Through innovative reinforcement learning training methods (RAPO), it significantly reduces inference token consumption while improving reasoning accuracy, exploring the innovative path of "less computation, higher intelligence" for large language models. We have also released the [**technical report**](https://huggingface.co/papers/2601.01718) for the Yuan3.0 model, where you can find more detailed technical information and evaluation results.
|
| 43 |
|
| 44 |
<div align="center">
|
| 45 |
<img src="https://huggingface.co/YuanLabAI/Yuan3.0-Flash-4bit/resolve/main/docs/Yuan3.0-architecture.png" width="80%" />
|
|
|
|
| 61 |
|
| 62 |
<div align="center">
|
| 63 |
<img src="https://huggingface.co/YuanLabAI/Yuan3.0-Flash-4bit/resolve/main/docs/Yuan3.0-benchmarks.png" width="80%" />
|
| 64 |
+
Fig.2: Yuan3.0 Flash Evaluation Results
|
| 65 |
</div>
|
| 66 |
|
| 67 |
|
|
|
|
| 183 |
| **OpenAI GPT-5.1** | 49.44 | 27.48 | 10.16 | 84.63 | 40.50 |
|
| 184 |
| **Yuan3.0 Flash** | **59.31** | 51.32 | 28.32 | 89.99 | 45.34 |
|
| 185 |
|
| 186 |
+
## 6. License Agreement
|
| 187 |
+
The use of Yuan 3.0 code and models must comply with the [《Yuan 3.0 Model License Agreement》](https://github.com/Yuan-lab-LLM/Yuan3.0?tab=License-1-ov-file). The Yuan 3.0 model supports commercial use without requiring authorization application. Please understand and comply with the agreement, and do not use the open-source model and code, as well as derivatives generated based on the open-source project, for any purpose that may bring harm to the country and society, or for any service that has not undergone security assessment and filing.
|