Update README.md
Browse files
README.md
CHANGED
|
@@ -14,10 +14,14 @@ base_model:
|
|
| 14 |
|
| 15 |
<p align="center">🤗 <a href="https://huggingface.co/inclusionAI">Hugging Face</a>   |   🤖 <a href="https://modelscope.cn/organization/inclusionAI">ModelScope</a></p>
|
| 16 |
|
| 17 |
-
|
| 18 |
-
|
| 19 |
-
|
|
|
|
| 20 |
|
|
|
|
|
|
|
|
|
|
| 21 |
|
| 22 |
## Model Downloads
|
| 23 |
|
|
@@ -26,18 +30,8 @@ We present a compact yet powerful reasoning model **Ring-mini-2.0**. It has 16B
|
|
| 26 |
| **Model** | **#Total Params** | **#Activated Params** | **Context Length** | **Download** |
|
| 27 |
| :----------------: | :---------------: | :-------------------: | :----------------: | :----------: |
|
| 28 |
| Ring-mini-2.0 | 16.8B | 1.4B | 128K | [🤗 HuggingFace](https://huggingface.co/inclusionAI/Ring-mini-2.0) |
|
| 29 |
-
| Ring-lite-2507 | 16.8B | 2.75B | 128K | [🤗 HuggingFace](https://huggingface.co/inclusionAI/Ring-lite-2507) |
|
| 30 |
-
|
| 31 |
</div>
|
| 32 |
|
| 33 |
-
## Evaluation
|
| 34 |
-
For a comprehensive evaluation of the quality of our reasoning models, we implemented automatic benchmarks to assess their performance including math, code and science. The results indicate **Ring-mini-2.0** achieves comparable performace with **Ring-lite-2507** while activating only half parameters.
|
| 35 |
-
|
| 36 |
-
<p align="center">
|
| 37 |
-
<img src="https://mdn.alipayobjects.com/huamei_qa8qxu/afts/img/A*mJKfTaP18DwAAAAAUIAAAAgAemJ7AQ/fmt.avif" width="1000"/>
|
| 38 |
-
<p>
|
| 39 |
-
|
| 40 |
-
|
| 41 |
## Quickstart
|
| 42 |
|
| 43 |
### 🤗 Hugging Face Transformers
|
|
|
|
| 14 |
|
| 15 |
<p align="center">🤗 <a href="https://huggingface.co/inclusionAI">Hugging Face</a>   |   🤖 <a href="https://modelscope.cn/organization/inclusionAI">ModelScope</a></p>
|
| 16 |
|
| 17 |
+
Today, we officially release Ring-mini-2.0 — a high-performance inference-oriented MoE model deeply optimized based on the Ling 2.0 architecture. With only 16B total parameters and 1.4B activated parameters, it achieves comprehensive reasoning capabilities comparable to dense models below the 10B scale. It excels particularly in logical reasoning, code generation, and mathematical tasks, while supporting 128K long-context processing and 300+ tokens/s high-speed generation.
|
| 18 |
+
Enhanced Reasoning: Joint Training with SFT + RLVR + RLHF
|
| 19 |
+
Built upon Ling-mini-2.0-base, Ring-mini-2.0 undergoes further training with Long-CoT SFT, more stable and continuous RLVR, and RLHF joint optimization, significantly improving the stability and generalization of complex reasoning. On multiple challenging benchmarks (LiveCodeBench, AIME 2025, GPQA, ARC-AGI-v1, etc.), it outperforms dense models below 10B and even rivals larger MoE models (e.g., gpt-oss-20B-medium) with comparable output lengths, particularly excelling in logical reasoning.
|
| 20 |
+
For a comprehensive evaluation of the quality of our reasoning models, we implemented automatic benchmarks to assess their performance including math, code and science. The results indicate **Ring-mini-2.0** achieves comparable performace with **Ring-lite-2507** while activating only half parameters.
|
| 21 |
|
| 22 |
+
<p align="center">
|
| 23 |
+
<img src="https://mdn.alipayobjects.com/huamei_d2byvp/afts/img/OQWDT7e6BVwAAAAATGAAAAgADod9AQFr/original" width="1000"/>
|
| 24 |
+
<p>
|
| 25 |
|
| 26 |
## Model Downloads
|
| 27 |
|
|
|
|
| 30 |
| **Model** | **#Total Params** | **#Activated Params** | **Context Length** | **Download** |
|
| 31 |
| :----------------: | :---------------: | :-------------------: | :----------------: | :----------: |
|
| 32 |
| Ring-mini-2.0 | 16.8B | 1.4B | 128K | [🤗 HuggingFace](https://huggingface.co/inclusionAI/Ring-mini-2.0) |
|
|
|
|
|
|
|
| 33 |
</div>
|
| 34 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 35 |
## Quickstart
|
| 36 |
|
| 37 |
### 🤗 Hugging Face Transformers
|