LiangJiang commited on
Commit
687c625
·
verified ·
1 Parent(s): f12538b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -1
README.md CHANGED
@@ -18,7 +18,6 @@ Today, we officially release Ring-mini-2.0 — a high-performance inference-orie
18
 
19
  ## Enhanced Reasoning: Joint Training with SFT + RLVR + RLHF
20
  Built upon Ling-mini-2.0-base, Ring-mini-2.0 undergoes further training with Long-CoT SFT, more stable and continuous RLVR, and RLHF joint optimization, significantly improving the stability and generalization of complex reasoning. On multiple challenging benchmarks (LiveCodeBench, AIME 2025, GPQA, ARC-AGI-v1, etc.), it outperforms dense models below 10B and even rivals larger MoE models (e.g., gpt-oss-20B-medium) with comparable output lengths, particularly excelling in logical reasoning.
21
- For a comprehensive evaluation of the quality of our reasoning models, we implemented automatic benchmarks to assess their performance including math, code and science. The results indicate **Ring-mini-2.0** achieves comparable performace with **Ring-lite-2507** while activating only half parameters.
22
 
23
  <p align="center">
24
  <img src="https://mdn.alipayobjects.com/huamei_d2byvp/afts/img/OQWDT7e6BVwAAAAATGAAAAgADod9AQFr/original" width="1000"/>
 
18
 
19
  ## Enhanced Reasoning: Joint Training with SFT + RLVR + RLHF
20
  Built upon Ling-mini-2.0-base, Ring-mini-2.0 undergoes further training with Long-CoT SFT, more stable and continuous RLVR, and RLHF joint optimization, significantly improving the stability and generalization of complex reasoning. On multiple challenging benchmarks (LiveCodeBench, AIME 2025, GPQA, ARC-AGI-v1, etc.), it outperforms dense models below 10B and even rivals larger MoE models (e.g., gpt-oss-20B-medium) with comparable output lengths, particularly excelling in logical reasoning.
 
21
 
22
  <p align="center">
23
  <img src="https://mdn.alipayobjects.com/huamei_d2byvp/afts/img/OQWDT7e6BVwAAAAATGAAAAgADod9AQFr/original" width="1000"/>