Add metadata and link to paper

#2
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +10 -3
README.md CHANGED
@@ -1,3 +1,9 @@
 
 
 
 
 
 
1
  <div align="center">
2
  <h1>
3
  Yuan 3.0 Multimodal Foundation Model
@@ -33,7 +39,7 @@
33
 
34
  ## 1. Introduction
35
 
36
- Yuan 3.0 Flash, developed by the **YuanLab.ai team**, is a **40B parameter multimodal foundation model** that employs a Mixture of Experts (MoE) architecture, activating only approximately **3.7B parameters** per inference. Through innovative reinforcement learning training methods (RAPO), it significantly reduces inference token consumption while improving reasoning accuracy, exploring the innovative path of "less computation, higher intelligence" for large language models. We have also released the <a href="https://arxiv.org/abs/2601.01718" target="_blank">**technical report**</a> for the Yuan3.0 model, where you can find more detailed technical information and evaluation results.
37
 
38
  <div align="center">
39
  <img src="https://huggingface.co/YuanLabAI/Yuan3.0-Flash-4bit/resolve/main/docs/Yuan3.0-architecture.png" width="80%" />
@@ -55,7 +61,7 @@ Yuan 3.0 Flash outperforms GPT-5.1 in enterprise-grade RAG, multimodal retrieval
55
 
56
  <div align="center">
57
  <img src="https://huggingface.co/YuanLabAI/Yuan3.0-Flash-4bit/resolve/main/docs/Yuan3.0-benchmarks.png" width="80%" />
58
- Fig.1: Yuan3.0 Multimodal Large Language Model Architecture
59
  </div>
60
 
61
 
@@ -177,4 +183,5 @@ Summarization generation is a core requirement for historical information compre
177
  | **OpenAI GPT-5.1** | 49.44 | 27.48 | 10.16 | 84.63 | 40.50 |
178
  | **Yuan3.0 Flash** | **59.31** | 51.32 | 28.32 | 89.99 | 45.34 |
179
 
180
-
 
 
1
+ ---
2
+ license: other
3
+ library_name: transformers
4
+ pipeline_tag: image-text-to-text
5
+ ---
6
+
7
  <div align="center">
8
  <h1>
9
  Yuan 3.0 Multimodal Foundation Model
 
39
 
40
  ## 1. Introduction
41
 
42
+ Yuan 3.0 Flash, developed by the **YuanLab.ai team**, is a **40B parameter multimodal foundation model** that employs a Mixture of Experts (MoE) architecture, activating only approximately **3.7B parameters** per inference. Through innovative reinforcement learning training methods (RAPO), it significantly reduces inference token consumption while improving reasoning accuracy, exploring the innovative path of "less computation, higher intelligence" for large language models. We have also released the [**technical report**](https://huggingface.co/papers/2601.01718) for the Yuan3.0 model, where you can find more detailed technical information and evaluation results.
43
 
44
  <div align="center">
45
  <img src="https://huggingface.co/YuanLabAI/Yuan3.0-Flash-4bit/resolve/main/docs/Yuan3.0-architecture.png" width="80%" />
 
61
 
62
  <div align="center">
63
  <img src="https://huggingface.co/YuanLabAI/Yuan3.0-Flash-4bit/resolve/main/docs/Yuan3.0-benchmarks.png" width="80%" />
64
+ Fig.2: Yuan3.0 Flash Evaluation Results
65
  </div>
66
 
67
 
 
183
  | **OpenAI GPT-5.1** | 49.44 | 27.48 | 10.16 | 84.63 | 40.50 |
184
  | **Yuan3.0 Flash** | **59.31** | 51.32 | 28.32 | 89.99 | 45.34 |
185
 
186
+ ## 6. License Agreement
187
+ The use of Yuan 3.0 code and models must comply with the [《Yuan 3.0 Model License Agreement》](https://github.com/Yuan-lab-LLM/Yuan3.0?tab=License-1-ov-file). The Yuan 3.0 model supports commercial use without requiring authorization application. Please understand and comply with the agreement, and do not use the open-source model and code, as well as derivatives generated based on the open-source project, for any purpose that may bring harm to the country and society, or for any service that has not undergone security assessment and filing.