Improve model card: Add license, library name, pipeline tag, and GitHub link

#2
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +18 -11
README.md CHANGED
@@ -1,12 +1,18 @@
1
  ---
2
- datasets:
3
- - zwhe99/DeepMath-103K
4
  base_model:
5
  - deepseek-ai/DeepSeek-R1-Distill-Qwen-7B
 
 
 
 
 
6
  ---
 
7
  # AutoDeco
8
  Official Implementation of "[The End of Manual Decoding: Towards Truly End-to-End Language Models](https://arxiv.org/abs/2510.26697)"
9
 
 
 
10
  **AutoDeco** is a framework that adds token-level adaptive decoding parameter prediction capabilities to Large Language Models (LLMs). By adding lightweight prediction heads on top of pre-trained models, AutoDeco can dynamically predict optimal temperature and top-p parameters for each token during decoding.
11
 
12
  ## 🎯 Key Features
@@ -146,8 +152,15 @@ Training data should be in JSONL format, with one sample per line. AutoDeco supp
146
 
147
  # example
148
  {
149
- "prompt": "<|im_start|>user\nEvaluate the limit:$$\\lim_{(x, y) \\to (1, 2)} \\frac{(x-1)(y-2)-x+3}{x^2-2x+y^2-4}$$\nMake sure you output the final answer within \\boxed{}<|im_end|>\n< im_start>assistant\n",
150
- "completion": "......### Final Answer:\n$$\n\\boxed{-1}\n$$""
 
 
 
 
 
 
 
151
  }
152
  ```
153
 
@@ -277,10 +290,4 @@ If you use AutoDeco in your research, please cite:
277
  primaryClass={cs.CL},
278
  url={https://arxiv.org/abs/2510.26697},
279
  }
280
- ```
281
-
282
- <!-- ## Acknowledgments
283
-
284
- - Built on [Transformers](https://github.com/huggingface/transformers) and [TRL](https://github.com/huggingface/trl)
285
- - Training framework uses [DeepSpeed](https://github.com/microsoft/DeepSpeed)
286
- - Inference optimization uses [vLLM](https://github.com/vllm-project/vllm) -->
 
1
  ---
 
 
2
  base_model:
3
  - deepseek-ai/DeepSeek-R1-Distill-Qwen-7B
4
+ datasets:
5
+ - zwhe99/DeepMath-103K
6
+ license: apache-2.0
7
+ library_name: transformers
8
+ pipeline_tag: text-generation
9
  ---
10
+
11
  # AutoDeco
12
  Official Implementation of "[The End of Manual Decoding: Towards Truly End-to-End Language Models](https://arxiv.org/abs/2510.26697)"
13
 
14
+ Code: https://github.com/Zacks917/AutoDeco
15
+
16
  **AutoDeco** is a framework that adds token-level adaptive decoding parameter prediction capabilities to Large Language Models (LLMs). By adding lightweight prediction heads on top of pre-trained models, AutoDeco can dynamically predict optimal temperature and top-p parameters for each token during decoding.
17
 
18
  ## 🎯 Key Features
 
152
 
153
  # example
154
  {
155
+ "prompt": "<|im_start|>user
156
+ Evaluate the limit:$$\\lim_{(x, y) \\to (1, 2)} \\frac{(x-1)(y-2)-x+3}{x^2-2x+y^2-4}$$
157
+ Make sure you output the final answer within \\boxed{}<|im_end|>
158
+ < im_start>assistant
159
+ ",
160
+ "completion": "......### ✅ Final Answer:
161
+ $$
162
+ \\boxed{-1}
163
+ $$""
164
  }
165
  ```
166
 
 
290
  primaryClass={cs.CL},
291
  url={https://arxiv.org/abs/2510.26697},
292
  }
293
+ ```