Yuan-lab commited on
Commit
b721cac
·
verified ·
1 Parent(s): b2d4dae

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -40
README.md CHANGED
@@ -6,8 +6,8 @@
6
 
7
  <hr>
8
  <div align="center" style="line-height: 1;">
9
- <a href="https://huggingface.co/YuanLabAI"><img alt="Hugging Face"
10
- src="https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-Yuan%20LLM-ffc107?color=ffc107&logoColor=white"/></a>
11
  <a href="https://www.modelscope.cn/profile/Yuanlab"><img alt="ModelScope"
12
  src="https://img.shields.io/badge/💾%20ModelScope-Yuan%20LLM-6b4fbb?color=6b4fbb&logoColor=white"/></a>
13
  <a href="https://x.com/Yuanlabai"><img alt="Twitter Follow"
@@ -21,15 +21,6 @@
21
  </div>
22
 
23
 
24
-
25
- <h4 align="center">
26
- <p>
27
- <a href="./README_ZH.md">简体中文</a> |
28
- <b>English</b>
29
- <p>
30
- </h4>
31
-
32
-
33
  -----
34
 
35
 
@@ -44,10 +35,9 @@
44
 
45
  Yuan 3.0 Flash, developed by the **YuanLab.ai team**, is a **40B parameter multimodal foundation model** that employs a Mixture of Experts (MoE) architecture, activating only approximately **3.7B parameters** per inference. Through innovative reinforcement learning training methods (RAPO), it significantly reduces inference token consumption while improving reasoning accuracy, exploring the innovative path of "less computation, higher intelligence" for large language models. We have also released the <a href="https://github.com/Yuan-lab-LLM/Yuan3.0/blob/main/docs/YUAN3.0_FLASH-paper.pdf" target="_blank">**technical report**</a> for the Yuan3.0 model, where you can find more detailed technical information and evaluation results.
46
 
47
- <div align=center> <img src=https://github.com/Yuan-lab-LLM/Yuan3.0/blob/main/docs/Yuan3.0-architecture.png width=80% />
48
-
49
  Fig.1: Yuan3.0 Multimodal Large Language Model Architecture
50
-
51
  </div>
52
 
53
  ### Core Features
@@ -62,10 +52,10 @@ Fig.1: Yuan3.0 Multimodal Large Language Model Architecture
62
 
63
  Yuan 3.0 Flash outperforms GPT-5.1 in enterprise-grade RAG, multimodal retrieval, table understanding, summary generation and other tasks. With 40B parameters, it achieves the reasoning accuracy of 235B/671B models while reducing token consumption by 50%-75%, providing enterprises with high-performance, low-cost large language model solutions.
64
 
65
- <div align=center> <img src=https://github.com/Yuan-lab-LLM/Yuan3.0/blob/main/docs/Yuan3.0-benchmarks.png width=80% />
66
-
67
- Fig.2: Yuan3.0 Flash Evaluation Results
68
 
 
 
 
69
  </div>
70
 
71
 
@@ -188,26 +178,3 @@ Summarization generation is a core requirement for historical information compre
188
  | **Yuan3.0 Flash** | **59.31** | 51.32 | 28.32 | 89.99 | 45.34 |
189
 
190
 
191
- ---
192
-
193
- ## 6. Quick Start
194
-
195
- **6.1 Yuan3.0 Flash Inference**
196
-
197
- Yuan3.0 Flash supports bfloat16 and int4 quantized models. For specific usage methods, please refer to [QuickStart](vllm/README_Yuan.md)
198
-
199
-
200
- **6.2 Data Preprocessing**
201
-
202
- We provide data preprocessing scripts. Please refer to the data preprocessing [documentation](rlhf/docs/data_process.md).
203
-
204
- **6.3 Model Fine-tuning Training**
205
-
206
- We provide supervised fine-tuning scripts and reinforcement learning workflows for the Yuan3.0 Flash model. Please refer to the fine-tuning training [documentation](rlhf/docs/instruct_tuning.md) and reinforcement learning [documentation](rlhf/docs/RL_training.md).
207
-
208
-
209
-
210
- ## 7. License Agreement
211
- The use of Yuan 3.0 code and models must comply with the [《Yuan 3.0 Model License Agreement》](https://github.com/Yuan-lab-LLM/Yuan3.0?tab=License-1-ov-file). The Yuan 3.0 model supports commercial use without requiring authorization application. Please understand and comply with the agreement, and do not use the open-source model and code, as well as derivatives generated based on the open-source project, for any purpose that may bring harm to the country and society, or for any service that has not undergone security assessment and filing.
212
-
213
- Although we have taken measures to ensure the compliance and accuracy of the data during model training, due to the large number of model parameters and the influence of probabilistic randomness, we cannot guarantee the accuracy of the output content, and the model is easily misled by input instructions. This project does not assume responsibility for data security and public opinion risks caused by open-source models and code, or any risks and responsibilities arising from the model being misled, abused, disseminated, or improperly used. You will independently bear all risks and consequences arising from the use, copying, distribution, and modification of the model through using this open-source project.
 
6
 
7
  <hr>
8
  <div align="center" style="line-height: 1;">
9
+ <a href="https://github.com/Yuan-lab-LLM/Yuan3.0"><img alt="GitHub"
10
+ src="https://img.shields.io/badge/GitHub-Yuan%203.0%20Repo-181717?logo=github&logoColor=white"/></a>
11
  <a href="https://www.modelscope.cn/profile/Yuanlab"><img alt="ModelScope"
12
  src="https://img.shields.io/badge/💾%20ModelScope-Yuan%20LLM-6b4fbb?color=6b4fbb&logoColor=white"/></a>
13
  <a href="https://x.com/Yuanlabai"><img alt="Twitter Follow"
 
21
  </div>
22
 
23
 
 
 
 
 
 
 
 
 
 
24
  -----
25
 
26
 
 
35
 
36
  Yuan 3.0 Flash, developed by the **YuanLab.ai team**, is a **40B parameter multimodal foundation model** that employs a Mixture of Experts (MoE) architecture, activating only approximately **3.7B parameters** per inference. Through innovative reinforcement learning training methods (RAPO), it significantly reduces inference token consumption while improving reasoning accuracy, exploring the innovative path of "less computation, higher intelligence" for large language models. We have also released the <a href="https://github.com/Yuan-lab-LLM/Yuan3.0/blob/main/docs/YUAN3.0_FLASH-paper.pdf" target="_blank">**technical report**</a> for the Yuan3.0 model, where you can find more detailed technical information and evaluation results.
37
 
38
+ <div align="center">
39
+ <img src="https://huggingface.co/YuanLabAI/Yuan3.0-Flash-4bit/resolve/main/docs/Yuan3.0-architecture.png" width="80%" />
40
  Fig.1: Yuan3.0 Multimodal Large Language Model Architecture
 
41
  </div>
42
 
43
  ### Core Features
 
52
 
53
  Yuan 3.0 Flash outperforms GPT-5.1 in enterprise-grade RAG, multimodal retrieval, table understanding, summary generation and other tasks. With 40B parameters, it achieves the reasoning accuracy of 235B/671B models while reducing token consumption by 50%-75%, providing enterprises with high-performance, low-cost large language model solutions.
54
 
 
 
 
55
 
56
+ <div align="center">
57
+ <img src="https://huggingface.co/YuanLabAI/Yuan3.0-Flash-4bit/resolve/main/docs/Yuan3.0-benchmarks.png" width="80%" />
58
+ Fig.1: Yuan3.0 Multimodal Large Language Model Architecture
59
  </div>
60
 
61
 
 
178
  | **Yuan3.0 Flash** | **59.31** | 51.32 | 28.32 | 89.99 | 45.34 |
179
 
180