Yuan-lab commited on
Commit
0c059f4
·
verified ·
1 Parent(s): 4e18cb4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -10,9 +10,9 @@
10
  src="https://img.shields.io/badge/GitHub-Yuan%203.0%20Repo-181717?logo=github&logoColor=white"/></a>
11
  <a href="https://www.modelscope.cn/profile/Yuanlab"><img alt="ModelScope"
12
  src="https://img.shields.io/badge/💾%20ModelScope-Yuan3.0-6b4fbb?color=6b4fbb&logoColor=white"/></a>
13
- <a href="https://x.com/Yuanlabai"><img alt="Twitter Follow"
14
  src="https://img.shields.io/badge/Twitter-Yuanlabai-white?logo=x&logoColor=white"/></a>
15
- <a href="https://github.com/Yuan-lab-LLM/Yuan3.0/blob/main/docs/YUAN3.0_FLASH-paper.pdf"><img alt="arXiv"
16
  src="https://img.shields.io/badge/arXiv-Yuan3.0%20Paper-b31b1b?logo=arxiv&logoColor=white"/></a>
17
  </a>
18
 
@@ -33,7 +33,7 @@
33
 
34
  ## 1. Introduction
35
 
36
- Yuan 3.0 Flash, developed by the **YuanLab.ai team**, is a **40B parameter multimodal foundation model** that employs a Mixture of Experts (MoE) architecture, activating only approximately **3.7B parameters** per inference. Through innovative reinforcement learning training methods (RAPO), it significantly reduces inference token consumption while improving reasoning accuracy, exploring the innovative path of "less computation, higher intelligence" for large language models. We have also released the <a href="https://github.com/Yuan-lab-LLM/Yuan3.0/blob/main/docs/YUAN3.0_FLASH-paper.pdf" target="_blank">**technical report**</a> for the Yuan3.0 model, where you can find more detailed technical information and evaluation results.
37
 
38
  <div align="center">
39
  <img src="https://huggingface.co/YuanLabAI/Yuan3.0-Flash-4bit/resolve/main/docs/Yuan3.0-architecture.png" width="80%" />
 
10
  src="https://img.shields.io/badge/GitHub-Yuan%203.0%20Repo-181717?logo=github&logoColor=white"/></a>
11
  <a href="https://www.modelscope.cn/profile/Yuanlab"><img alt="ModelScope"
12
  src="https://img.shields.io/badge/💾%20ModelScope-Yuan3.0-6b4fbb?color=6b4fbb&logoColor=white"/></a>
13
+ <a href="https://x.com/YuanAI_Lab"><img alt="Twitter Follow"
14
  src="https://img.shields.io/badge/Twitter-Yuanlabai-white?logo=x&logoColor=white"/></a>
15
+ <a href="https://arxiv.org/abs/2601.01718"><img alt="arXiv"
16
  src="https://img.shields.io/badge/arXiv-Yuan3.0%20Paper-b31b1b?logo=arxiv&logoColor=white"/></a>
17
  </a>
18
 
 
33
 
34
  ## 1. Introduction
35
 
36
+ Yuan 3.0 Flash, developed by the **YuanLab.ai team**, is a **40B parameter multimodal foundation model** that employs a Mixture of Experts (MoE) architecture, activating only approximately **3.7B parameters** per inference. Through innovative reinforcement learning training methods (RAPO), it significantly reduces inference token consumption while improving reasoning accuracy, exploring the innovative path of "less computation, higher intelligence" for large language models. We have also released the <a href="https://arxiv.org/abs/2601.01718" target="_blank">**technical report**</a> for the Yuan3.0 model, where you can find more detailed technical information and evaluation results.
37
 
38
  <div align="center">
39
  <img src="https://huggingface.co/YuanLabAI/Yuan3.0-Flash-4bit/resolve/main/docs/Yuan3.0-architecture.png" width="80%" />