File size: 6,621 Bytes
70cdbe7 a6a59bc 70cdbe7 a67eef7 70cdbe7 51aea39 70cdbe7 51aea39 70cdbe7 46a81f3 70cdbe7 9ce7d76 8cf0ec2 9ce7d76 8cf0ec2 9ce7d76 8cf0ec2 9ce7d76 8cf0ec2 9ce7d76 8cf0ec2 9ce7d76 8cf0ec2 9ce7d76 8cf0ec2 9ce7d76 70cdbe7 05dec7c decc0f1 70cdbe7 51aea39 70cdbe7 51aea39 70cdbe7 51aea39 70cdbe7 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 | ---
license: mit
language:
- zh
- en
base_model:
- inclusionAI/Ling-lite-base-1.5
---
# Ring-lite-2507
<p align="center">
<img src="https://mdn.alipayobjects.com/huamei_qa8qxu/afts/img/A*4QxcQrBlTiAAAAAAQXAAAAgAemJ7AQ/original" width="100"/>
<p>
<p align="center">
🤗 <a href="https://huggingface.co/inclusionAI">Hugging Face</a>
<p>
## Introduction
We present **Ring-lite-2507**, an upgraded version of our previously released lightweight reasoning model, **Ring-lite**. Building upon 16.8B Mixture-of-Experts (MoE)-based large language model with 2.75B activated parameters, Ring-lite-2507 further pushes its reasoning ability to an advanced level, meanwhile, it demonstrates superior performance on a comprehensive range of LLM benchmarks, including general text understanding, alignment, coding, logical and agentic tasks. Thanks to our innovative and robust reinforcement learning training pipeline, Ring-lite-2507 distinguished itself from latest public dense models under 10B parameters by showing competitive performance across various tasks while activating only 1/3 of their parameter size.
## Model Downloads
<div align="center">
| **Model** | **#Total Params** | **#Activated Params** | **Context Length** | **Download** |
| :----------------: | :---------------: | :-------------------: | :----------------: | :----------: |
| Ring-lite-2507 | 16.8B | 2.75B | 128K | [🤗 HuggingFace](https://huggingface.co/inclusionAI/Ring-lite-2507) |
| Ring-lite | 16.8B | 2.75B | 128K | [🤗 HuggingFace](https://huggingface.co/inclusionAI/Ring-lite) |
</div>
## Evaluation
For a comprehensive evaluation of the quality of our reasoning models, we implemented automatic benchmarks to assess their performance including math, code and science.
<p align="center">
<img src="https://mdn.alipayobjects.com/huamei_qa8qxu/afts/img/A*4FXmTLeqUWMAAAAARqAAAAgAemJ7AQ/original" width="1000"/>
<p>
To compare the performance of Ring-lite-2507 and Ring-lite, we evaluate the two models on a broader range of reasoning and general-purpose benchmarks, including knowledge understanding, math, coding, reasoning & agentic and alignment.
### Knowledge Understanding
| **Benchmark** | **Ring-lite-2507** | **Ring-lite-2506** | **Qwen3-8B-Thinking**
| :-------------: | :---------------: | :-----------: | :-------------------: |
| MMLU-Pro (EM) | 72.50 | 63.44 | **72.56** |
| GPQA-Diamond (Pass@1) | **69.35** | 63.51 | 62.00 |
| SuperGPQA (EM) | 40.05 | 13.97 | **40.36** |
| Phybench (Pass@1) | 28.51 | **29.19** | 22.14 |
### Math
| **Benchmark** | **Ring-lite-2507** | **Ring-lite-2506** | **Qwen3-8B-Thinking**
| :-------------: | :---------------: | :-----------: | :-------------------: |
| MATH-500 (Pass@1) | **97.95** | 96.80 | 97.30 |
| CNMO 2024 (Pass@1) | 75.09 | **77.26** | 74.57 |
| AIME 2024 (Pass@1) | **79.79** | 79.00 | 74.90 |
| AIME 2025 (Pass@1) | **72.92** | 69.50 | 67.19 |
| LiveMathBench (Pass@1) | 83.37 | **85.08** | 81.90 |
| TheoremQA (Pass@1) | 70.00 | **70.19** | 68.81 |
| OlympiadBench (math) (Pass@1) | 80.64 | **82.86** | 80.20 |
### Coding
| **Benchmark** | **Ring-lite-2507** | **Ring-lite-2506** | **Qwen3-8B-Thinking**
| :-------------: | :---------------: | :-----------: | :-------------------: |
| LiveCodeBench(2408-2505) (Pass@1) |**60.35** | 59.53 | 55.12 |
| Codeforces(Percentile) (Pass@1) |**1830** | 1673 | 1580 |
| Codeforces(Rating) |**92.16** | 88.00 | 79.44 |
### Reasoning \& Agentic
| **Benchmark** | **Ring-lite-2507** | **Ring-lite-2506** | **Qwen3-8B-Thinking**
| :-------------: | :---------------: | :-----------: | :-------------------: |
| DROP (zero-shot F1) | **89.27** | 60.21 | 87.13 |
| BBH (EM) | **88.65** | 50.84 | 87.30 |
| ARCPrize (Pass@1) | **19.00** | 3.12 | 3.88 |
| MuSR (EM) | **77.19** | 66.77 | 76.92 |
| BFCL_Live (Pass@1) | 74.81 | 66.76 | **75.99** |
### Alignment
| **Benchmark** | **Ring-lite-2507** | **Ring-lite-2506** | **Qwen3-8B-Thinking**
| :-------------: | :---------------: | :-----------: | :-------------------: |
| IFEval (Prompt Strict) | 84.66 | 54.34 | **85.40** |
| AlignBench v1.1(gpt-4.1) | **80.90** | 69.60 | 74.70 |
| FoFo (gpt-4-turbo) | **85.02** | 67.81 | 81.93 |
| ArenaHard (gpt-4.1) | **88.85** | 56.12 | 86.14 |
### Blog
More details are reported in our [blog](https://inclusionai.github.io/blog/ring-lite-2507/).
## Quickstart
### 🤗 Hugging Face Transformers
Here is a code snippet to show you how to use the chat model with `transformers`:
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model_name = "inclusionAI/Ring-lite-2507"
model = AutoModelForCausalLM.from_pretrained(
model_name,
torch_dtype="auto",
device_map="auto",
trust_remote_code=True
)
tokenizer = AutoTokenizer.from_pretrained(model_name)
prompt = "Give me a short introduction to large language models."
messages = [
{"role": "system", "content": "You are Ring, an assistant created by inclusionAI"},
{"role": "user", "content": prompt}
]
text = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True,
enable_thinking=True
)
model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
generated_ids = model.generate(
**model_inputs,
max_new_tokens=8192
)
generated_ids = [
output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
]
response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
```
## Deployment
Please refer to [GitHub](https://github.com/inclusionAI/Ring/blob/main/README.md)
## License
This code repository is licensed under [the MIT License](https://huggingface.co/inclusionAI/Ring-lite-2507/blob/main/LICENSE).
## Citation
```
@misc{ringteam2025ringlitescalablereasoningc3postabilized,
title={Ring-lite: Scalable Reasoning via C3PO-Stabilized Reinforcement Learning for LLMs},
author={Ling Team},
year={2025},
eprint={2506.14731},
archivePrefix={arXiv},
primaryClass={cs.CL},
url={https://arxiv.org/abs/2506.14731},
}
``` |