Add library_name and links to paper, code, and project page

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +13 -8
README.md CHANGED
@@ -1,26 +1,31 @@
1
  ---
 
2
  language:
3
  - zh
4
  - en
 
5
  pipeline_tag: text-generation
 
6
  tags:
7
  - deepscaler
8
  - reasoning
9
  - grpo
10
  - qwen2
11
- base_model: deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
12
- license: other
13
  ---
14
 
15
  # DECS_1.5B
16
- This is the official model for ICLR 2026 Oral "Overthinking Reduction with Decoupled Rewards and Curriculum Data Scheduling".
17
- DECS_1.5B is a reasoning-focused causal language model built from `deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B` and further trained with DECS algorithm, focused on 50% fewer tokens when answering a reasoning-required problem.
 
 
 
 
18
 
19
  ## Model Summary
20
 
21
- - Base model: `deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B`
22
- - Upload date: `2026-02-24`
23
- - Recommended use: long-form reasoning and mathematical/problem-solving style generation
24
 
25
  ## Quick Start (Transformers)
26
 
@@ -86,4 +91,4 @@ booktitle={The Fourteenth International Conference on Learning Representations},
86
  year={2026},
87
  url={https://openreview.net/forum?id=kdeiRledV6}
88
  }
89
- ```
 
1
  ---
2
+ base_model: deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
3
  language:
4
  - zh
5
  - en
6
+ license: other
7
  pipeline_tag: text-generation
8
+ library_name: transformers
9
  tags:
10
  - deepscaler
11
  - reasoning
12
  - grpo
13
  - qwen2
 
 
14
  ---
15
 
16
  # DECS_1.5B
17
+
18
+ This is the official model for the ICLR 2026 Oral paper: "**Overthinking Reduction with Decoupled Rewards and Curriculum Data Scheduling**".
19
+
20
+ [**Paper**](https://huggingface.co/papers/2509.25827) | [**Code**](https://github.com/pixas/DECS) | [**Project Page**](https://pixas.github.io/decs-iclr26-site/)
21
+
22
+ DECS_1.5B is a reasoning-focused causal language model built from `deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B` and further trained with the DECS algorithm, focused on 50% fewer tokens when answering a reasoning-required problem.
23
 
24
  ## Model Summary
25
 
26
+ - **Base model:** `deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B`
27
+ - **Upload date:** `2026-02-24`
28
+ - **Recommended use:** long-form reasoning and mathematical/problem-solving style generation
29
 
30
  ## Quick Start (Transformers)
31
 
 
91
  year={2026},
92
  url={https://openreview.net/forum?id=kdeiRledV6}
93
  }
94
+ ```