Improve model card: Add license, library name, pipeline tag, and GitHub link
#2
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -1,12 +1,18 @@
|
|
| 1 |
---
|
| 2 |
-
datasets:
|
| 3 |
-
- zwhe99/DeepMath-103K
|
| 4 |
base_model:
|
| 5 |
- deepseek-ai/DeepSeek-R1-Distill-Qwen-7B
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 6 |
---
|
|
|
|
| 7 |
# AutoDeco
|
| 8 |
Official Implementation of "[The End of Manual Decoding: Towards Truly End-to-End Language Models](https://arxiv.org/abs/2510.26697)"
|
| 9 |
|
|
|
|
|
|
|
| 10 |
**AutoDeco** is a framework that adds token-level adaptive decoding parameter prediction capabilities to Large Language Models (LLMs). By adding lightweight prediction heads on top of pre-trained models, AutoDeco can dynamically predict optimal temperature and top-p parameters for each token during decoding.
|
| 11 |
|
| 12 |
## 🎯 Key Features
|
|
@@ -146,8 +152,15 @@ Training data should be in JSONL format, with one sample per line. AutoDeco supp
|
|
| 146 |
|
| 147 |
# example
|
| 148 |
{
|
| 149 |
-
"prompt": "<|im_start|>user
|
| 150 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 151 |
}
|
| 152 |
```
|
| 153 |
|
|
@@ -277,10 +290,4 @@ If you use AutoDeco in your research, please cite:
|
|
| 277 |
primaryClass={cs.CL},
|
| 278 |
url={https://arxiv.org/abs/2510.26697},
|
| 279 |
}
|
| 280 |
-
```
|
| 281 |
-
|
| 282 |
-
<!-- ## Acknowledgments
|
| 283 |
-
|
| 284 |
-
- Built on [Transformers](https://github.com/huggingface/transformers) and [TRL](https://github.com/huggingface/trl)
|
| 285 |
-
- Training framework uses [DeepSpeed](https://github.com/microsoft/DeepSpeed)
|
| 286 |
-
- Inference optimization uses [vLLM](https://github.com/vllm-project/vllm) -->
|
|
|
|
| 1 |
---
|
|
|
|
|
|
|
| 2 |
base_model:
|
| 3 |
- deepseek-ai/DeepSeek-R1-Distill-Qwen-7B
|
| 4 |
+
datasets:
|
| 5 |
+
- zwhe99/DeepMath-103K
|
| 6 |
+
license: apache-2.0
|
| 7 |
+
library_name: transformers
|
| 8 |
+
pipeline_tag: text-generation
|
| 9 |
---
|
| 10 |
+
|
| 11 |
# AutoDeco
|
| 12 |
Official Implementation of "[The End of Manual Decoding: Towards Truly End-to-End Language Models](https://arxiv.org/abs/2510.26697)"
|
| 13 |
|
| 14 |
+
Code: https://github.com/Zacks917/AutoDeco
|
| 15 |
+
|
| 16 |
**AutoDeco** is a framework that adds token-level adaptive decoding parameter prediction capabilities to Large Language Models (LLMs). By adding lightweight prediction heads on top of pre-trained models, AutoDeco can dynamically predict optimal temperature and top-p parameters for each token during decoding.
|
| 17 |
|
| 18 |
## 🎯 Key Features
|
|
|
|
| 152 |
|
| 153 |
# example
|
| 154 |
{
|
| 155 |
+
"prompt": "<|im_start|>user
|
| 156 |
+
Evaluate the limit:$$\\lim_{(x, y) \\to (1, 2)} \\frac{(x-1)(y-2)-x+3}{x^2-2x+y^2-4}$$
|
| 157 |
+
Make sure you output the final answer within \\boxed{}<|im_end|>
|
| 158 |
+
< im_start>assistant
|
| 159 |
+
",
|
| 160 |
+
"completion": "......### ✅ Final Answer:
|
| 161 |
+
$$
|
| 162 |
+
\\boxed{-1}
|
| 163 |
+
$$""
|
| 164 |
}
|
| 165 |
```
|
| 166 |
|
|
|
|
| 290 |
primaryClass={cs.CL},
|
| 291 |
url={https://arxiv.org/abs/2510.26697},
|
| 292 |
}
|
| 293 |
+
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|