Add library_name and paper link
#1
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -1,11 +1,12 @@
|
|
| 1 |
---
|
| 2 |
-
|
|
|
|
| 3 |
language:
|
| 4 |
- zh
|
| 5 |
- en
|
| 6 |
-
|
| 7 |
-
- inclusionAI/Ling-lite
|
| 8 |
pipeline_tag: text-generation
|
|
|
|
| 9 |
---
|
| 10 |
|
| 11 |
# Ring-lite-distill-preview
|
|
@@ -22,6 +23,8 @@ pipeline_tag: text-generation
|
|
| 22 |
|
| 23 |
Ring-lite-distill-preview is an MoE LLM provided and open-sourced by InclusionAI, which has 16.8B parameters with 2.75B activated parameters. It was fine-tuned from [Ling-lite](https://modelscope.cn/models/inclusionAI/Ling-lite) using extensive reasoning-focused instruction data. This model delivers performance comparable to DeepSeek-R1-Distill-Qwen-7B on reasoning benchmarks while achieving better results on general benchmarks, especially superior performance on function-calling evaluation benchmarks (e.g., TEval, BFCl_v2) and instruction-following benchmarks (e.g., IFEval). This demonstrates that Ring-lite-distill is a more balanced and versatile model. Additionaly, it maintains competitive latency and throughput compared to other reasoning LLMs of similar size.
|
| 24 |
|
|
|
|
|
|
|
| 25 |
## Model Downloads
|
| 26 |
|
| 27 |
<div align="center">
|
|
@@ -108,4 +111,4 @@ Please refer to [Github](https://github.com/inclusionAI/Ring/blob/main/README.md
|
|
| 108 |
This code repository is licensed under [the MIT License](https://huggingface.co/inclusionAI/Ring-lite-distill/blob/main/LICENSE).
|
| 109 |
|
| 110 |
## Citation
|
| 111 |
-
[TBD]
|
|
|
|
| 1 |
---
|
| 2 |
+
base_model:
|
| 3 |
+
- inclusionAI/Ling-lite
|
| 4 |
language:
|
| 5 |
- zh
|
| 6 |
- en
|
| 7 |
+
license: mit
|
|
|
|
| 8 |
pipeline_tag: text-generation
|
| 9 |
+
library_name: transformers
|
| 10 |
---
|
| 11 |
|
| 12 |
# Ring-lite-distill-preview
|
|
|
|
| 23 |
|
| 24 |
Ring-lite-distill-preview is an MoE LLM provided and open-sourced by InclusionAI, which has 16.8B parameters with 2.75B activated parameters. It was fine-tuned from [Ling-lite](https://modelscope.cn/models/inclusionAI/Ling-lite) using extensive reasoning-focused instruction data. This model delivers performance comparable to DeepSeek-R1-Distill-Qwen-7B on reasoning benchmarks while achieving better results on general benchmarks, especially superior performance on function-calling evaluation benchmarks (e.g., TEval, BFCl_v2) and instruction-following benchmarks (e.g., IFEval). This demonstrates that Ring-lite-distill is a more balanced and versatile model. Additionaly, it maintains competitive latency and throughput compared to other reasoning LLMs of similar size.
|
| 25 |
|
| 26 |
+
The model was presented in the paper [](https://huggingface.co/papers/2504.07158).
|
| 27 |
+
|
| 28 |
## Model Downloads
|
| 29 |
|
| 30 |
<div align="center">
|
|
|
|
| 111 |
This code repository is licensed under [the MIT License](https://huggingface.co/inclusionAI/Ring-lite-distill/blob/main/LICENSE).
|
| 112 |
|
| 113 |
## Citation
|
| 114 |
+
[TBD]
|