Add library_name and paper link

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +7 -4
README.md CHANGED
@@ -1,11 +1,12 @@
1
  ---
2
- license: mit
 
3
  language:
4
  - zh
5
  - en
6
- base_model:
7
- - inclusionAI/Ling-lite
8
  pipeline_tag: text-generation
 
9
  ---
10
 
11
  # Ring-lite-distill-preview
@@ -22,6 +23,8 @@ pipeline_tag: text-generation
22
 
23
  Ring-lite-distill-preview is an MoE LLM provided and open-sourced by InclusionAI, which has 16.8B parameters with 2.75B activated parameters. It was fine-tuned from [Ling-lite](https://modelscope.cn/models/inclusionAI/Ling-lite) using extensive reasoning-focused instruction data. This model delivers performance comparable to DeepSeek-R1-Distill-Qwen-7B on reasoning benchmarks while achieving better results on general benchmarks, especially superior performance on function-calling evaluation benchmarks (e.g., TEval, BFCl_v2) and instruction-following benchmarks (e.g., IFEval). This demonstrates that Ring-lite-distill is a more balanced and versatile model. Additionaly, it maintains competitive latency and throughput compared to other reasoning LLMs of similar size.
24
 
 
 
25
  ## Model Downloads
26
 
27
  <div align="center">
@@ -108,4 +111,4 @@ Please refer to [Github](https://github.com/inclusionAI/Ring/blob/main/README.md
108
  This code repository is licensed under [the MIT License](https://huggingface.co/inclusionAI/Ring-lite-distill/blob/main/LICENSE).
109
 
110
  ## Citation
111
- [TBD]
 
1
  ---
2
+ base_model:
3
+ - inclusionAI/Ling-lite
4
  language:
5
  - zh
6
  - en
7
+ license: mit
 
8
  pipeline_tag: text-generation
9
+ library_name: transformers
10
  ---
11
 
12
  # Ring-lite-distill-preview
 
23
 
24
  Ring-lite-distill-preview is an MoE LLM provided and open-sourced by InclusionAI, which has 16.8B parameters with 2.75B activated parameters. It was fine-tuned from [Ling-lite](https://modelscope.cn/models/inclusionAI/Ling-lite) using extensive reasoning-focused instruction data. This model delivers performance comparable to DeepSeek-R1-Distill-Qwen-7B on reasoning benchmarks while achieving better results on general benchmarks, especially superior performance on function-calling evaluation benchmarks (e.g., TEval, BFCl_v2) and instruction-following benchmarks (e.g., IFEval). This demonstrates that Ring-lite-distill is a more balanced and versatile model. Additionaly, it maintains competitive latency and throughput compared to other reasoning LLMs of similar size.
25
 
26
+ The model was presented in the paper [](https://huggingface.co/papers/2504.07158).
27
+
28
  ## Model Downloads
29
 
30
  <div align="center">
 
111
  This code repository is licensed under [the MIT License](https://huggingface.co/inclusionAI/Ring-lite-distill/blob/main/LICENSE).
112
 
113
  ## Citation
114
+ [TBD]