LiangJiang commited on
Commit
19461fd
·
verified ·
1 Parent(s): db274d2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +153 -3
README.md CHANGED
@@ -1,3 +1,153 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ language:
4
+ - zh
5
+ - en
6
+ base_model:
7
+ - inclusionAI/Ling-lite-base-1.5
8
+ ---
9
+ # Ring-lite-2507
10
+
11
+ <p align="center">
12
+ <img src="https://mdn.alipayobjects.com/huamei_qa8qxu/afts/img/A*4QxcQrBlTiAAAAAAQXAAAAgAemJ7AQ/original" width="100"/>
13
+ <p>
14
+
15
+ <p align="center">
16
+ 🤗 <a href="https://huggingface.co/inclusionAI">Hugging Face</a>
17
+ <p>
18
+
19
+ ## Introduction
20
+
21
+ We present a compact yet powerful reasoning model **Ring-mini-2.0**. It has 16B total parameters, with 1.4B parameters are activated per input token (non-embedding 789M). Trained on more than 20T tokens of high-quality data and enhanced through long-cot supervised fine-tuning and multi-stage reinforcement learning, **Ring-mini-2.0** still reaches the top-tier level of sub-10B dense LLMs and even matches or surpasses much larger MoE models.
22
+
23
+
24
+ ## Model Downloads
25
+
26
+ <div align="center">
27
+
28
+ | **Model** | **#Total Params** | **#Activated Params** | **Context Length** | **Download** |
29
+ | :----------------: | :---------------: | :-------------------: | :----------------: | :----------: |
30
+ | Ring-mini-2.0 | 16.8B | 1.4B | 128K | [🤗 HuggingFace](https://huggingface.co/inclusionAI/Ring-mini-2.0) |
31
+
32
+ </div>
33
+
34
+ ## Evaluation
35
+ For a comprehensive evaluation of the quality of our reasoning models, we implemented automatic benchmarks to assess their performance including math, code and science.
36
+
37
+ <p align="center">
38
+ <img src="https://mdn.alipayobjects.com/huamei_qa8qxu/afts/img/A*5F9KR7Tm4MAAAAAARzAAAAgAemJ7AQ/original" width="1000"/>
39
+ <p>
40
+
41
+ To compare the performance of Ring-lite-2507 and Ring-lite, we evaluate the two models on a broader range of reasoning and general-purpose benchmarks, including knowledge understanding, math, coding, reasoning & agentic and alignment.
42
+
43
+ ### Knowledge Understanding
44
+
45
+ | **Benchmark** | **Ring-mini-2.0** | **Ring-lite-2507** | **Qwen3-8B-Thinking**
46
+ | :-------------: | :---------------: | :-----------: | :-------------------: |
47
+ | MMLU-Pro (EM) | 71.52 | 72.50 | 72.56 |
48
+ | GPQA-Diamond (Pass@1) | 68.24 | 69.35 | 62.00 |
49
+ | SuperGPQA (EM) | 36.21 | 39.57 | 42.42 |
50
+ | Phybench (Pass@1) | 25.80 | 28.51 | 22.14 |
51
+
52
+
53
+ ### Math
54
+
55
+ | **Benchmark** | **Ring-lite-2507** | **Ring-lite-2506** | **Qwen3-8B-Thinking**
56
+ | :-------------: | :---------------: | :-----------: | :-------------------: |
57
+ | MATH-500 (Pass@1) | 97.60 | 76.95 | 97.30 |
58
+ | CNMO 2024 (Pass@1) | 76.91 | 77.78 | 75.09 |
59
+ | AIME 2024 (Pass@1) | 79.69 | 84.06 | 79.27 |
60
+ | AIME 2025 (Pass@1) | 74.06 | 79.74 | 71.25 |
61
+ | LiveMathBench (Pass@1) | 83.98 | 84.94 | 82.92 |
62
+ | TheoremQA (Pass@1) | 70.09 | 70.00 | 68.81 |
63
+ | OlympiadBench (math) (Pass@1) | 82.91 | 84.94 | 82.27 |
64
+
65
+ ### Coding
66
+
67
+ | **Benchmark** | **Ring-lite-2507** | **Ring-lite-2506** | **Qwen3-8B-Thinking**
68
+ | :-------------: | :---------------: | :-----------: | :-------------------: |
69
+ | LiveCodeBench(2408-2505) (Pass@1) |62.56 | 63.27 | 56.94 |
70
+ | Codeforces | 84.80 | 89.09 | 73.31 |
71
+
72
+ ### Reasoning \& Agentic
73
+
74
+ | **Benchmark** | **Ring-lite-2507** | **Ring-lite-2506** | **Qwen3-8B-Thinking**
75
+ | :-------------: | :---------------: | :-----------: | :-------------------: |
76
+ | DROP (zero-shot F1) | 88.55 | 89.27 | 87.13 |
77
+ | BBH (EM) | 87.59 | 88.65 | 87.30 |
78
+ | ARCPrize (Pass@1) | 20.12 | 21.25 | 4.38 |
79
+ | MuSR (EM) | 75.99 | 77.19 | 76.92 |
80
+ | BFCL_Live (Pass@1) | 74.26 | 74.81 | 75.99 |
81
+
82
+ ### Alignment
83
+
84
+ | **Benchmark** | **Ring-lite-2507** | **Ring-lite-2506** | **Qwen3-8B-Thinking**
85
+ | :-------------: | :---------------: | :-----------: | :-------------------: |
86
+ | IFEval (Prompt Strict) | 78.93 | 82.99 | 85.0 |
87
+ | AlignBench v1.1(gpt-4.1) | 80.69 | 80.90 | 74.70 |
88
+ | FoFo (gpt-4-turbo) | 84.11 | 85.02 | 81.93 |
89
+ | ArenaHard (gpt-4.1) | 85.19 | 88.85 | 86.14 |
90
+
91
+
92
+ ## Quickstart
93
+
94
+ ### 🤗 Hugging Face Transformers
95
+
96
+ Here is a code snippet to show you how to use the chat model with `transformers`:
97
+
98
+ ```python
99
+ from transformers import AutoModelForCausalLM, AutoTokenizer
100
+
101
+ model_name = "inclusionAI/Ring-lite-2507"
102
+
103
+ model = AutoModelForCausalLM.from_pretrained(
104
+ model_name,
105
+ torch_dtype="auto",
106
+ device_map="auto",
107
+ trust_remote_code=True
108
+ )
109
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
110
+
111
+ prompt = "Give me a short introduction to large language models."
112
+ messages = [
113
+ {"role": "system", "content": "You are Ring, an assistant created by inclusionAI"},
114
+ {"role": "user", "content": prompt}
115
+ ]
116
+ text = tokenizer.apply_chat_template(
117
+ messages,
118
+ tokenize=False,
119
+ add_generation_prompt=True,
120
+ enable_thinking=True
121
+ )
122
+ model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
123
+
124
+ generated_ids = model.generate(
125
+ **model_inputs,
126
+ max_new_tokens=8192
127
+ )
128
+ generated_ids = [
129
+ output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
130
+ ]
131
+
132
+ response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
133
+ ```
134
+
135
+
136
+ ## Deployment
137
+ Please refer to [GitHub](https://github.com/inclusionAI/Ring/blob/main/README.md)
138
+
139
+ ## License
140
+ This code repository is licensed under [the MIT License](https://huggingface.co/inclusionAI/Ring-lite-2507/blob/main/LICENSE).
141
+
142
+ ## Citation
143
+ ```
144
+ @misc{ringteam2025ringlitescalablereasoningc3postabilized,
145
+ title={Ring-lite: Scalable Reasoning via C3PO-Stabilized Reinforcement Learning for LLMs},
146
+ author={Ling Team},
147
+ year={2025},
148
+ eprint={2506.14731},
149
+ archivePrefix={arXiv},
150
+ primaryClass={cs.CL},
151
+ url={https://arxiv.org/abs/2506.14731},
152
+ }
153
+ ```