Add model card with metadata and sample usage

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +32 -2
README.md CHANGED
@@ -1,6 +1,36 @@
1
- This is the PosS-3 model of the paper **PosS:Position Specialist Generates Better Draft for Speculative Decoding**
 
 
 
 
2
 
3
- If the code fails to auto-download the models, you may mannually download the following files.
 
 
 
 
 
 
 
 
 
 
4
 
5
  - `pytorch_model.bin`: Model weights
6
  - `config.json`: Model config
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ pipeline_tag: text-generation
3
+ library_name: transformers
4
+ license: apache-2.0
5
+ ---
6
 
7
+ # POSS: Position Specialist Generates Better Draft for Speculative Decoding
8
+
9
+ This repository contains the PosS-3 model described in the paper [POSS: Position Specialist Generates Better Draft for Speculative Decoding](https://arxiv.org/abs/2506.03566).
10
+
11
+ **Authors:** [Langlin Huang](https://shrango.github.io/), [Chengsong Huang](https://chengsong-huang.github.io/), [Jixuan Leng](https://jixuanleng.com/), Di Huang, [Jiaxin Huang](https://teapot123.github.io/)
12
+
13
+ The PosS model improves speculative decoding by using multiple position-specialized draft layers. This approach mitigates error accumulation in draft model-generated features, leading to improved token acceptance rates, especially at later positions.
14
+
15
+ For code and further details, please refer to the GitHub repository: [https://github.com/shrango/PosS](https://github.com/shrango/PosS)
16
+
17
+ If the code fails to auto-download the models, you may manually download the following files:
18
 
19
  - `pytorch_model.bin`: Model weights
20
  - `config.json`: Model config
21
+
22
+ **Sample Usage (Inference):**
23
+
24
+ The following command demonstrates how to use the model for inference (replace placeholders with actual paths):
25
+
26
+ ```bash
27
+ python spec_decode.py \
28
+ --device-num 0 \
29
+ --target-model llama3-8b \
30
+ --method poss-3 \
31
+ --temperature 0 \
32
+ --total-token 60 \
33
+ --depth 6 \
34
+ --repeat-time 3 \
35
+ --dataset mt_bench
36
+ ```