Feng Luo commited on
Commit
4f3bf85
·
1 Parent(s): e73d95b

update README

Browse files
Files changed (1) hide show
  1. README.md +15 -6
README.md CHANGED
@@ -1,5 +1,7 @@
1
  ---
2
  license: apache-2.0
 
 
3
  ---
4
 
5
  # AutoL2S-7B
@@ -8,16 +10,25 @@ This is the official model repository for **AutoL2S-7B**, a model fine-tuned for
8
 
9
  ## 💡 Overview
10
 
11
- AutoL2S enables automatically switching between short and long reasoning paths based on input complexity. This repository contains:
 
 
 
 
 
 
12
 
13
  - Model weights
14
  - Configuration files
15
  - necessary scripts in the `examples/` directory
16
 
17
- ---
 
 
 
18
 
 
19
  ## 🧩 Dependencies
20
-
21
  We recommend using the model with [vLLM](https://github.com/vllm-project/vllm).
22
  The code has been tested with:
23
 
@@ -26,7 +37,6 @@ vLLM == 0.6.2
26
  ```
27
 
28
  ---
29
-
30
  ## 🚀 How to Use
31
 
32
  Run the inference example:
@@ -72,5 +82,4 @@ If you use this model in your work, please consider citing:
72
  journal={arXiv preprint},
73
  year={2025}
74
  }
75
- ```
76
-
 
1
  ---
2
  license: apache-2.0
3
+ base_model:
4
+ - Qwen/Qwen2.5-7B-Instruct
5
  ---
6
 
7
  # AutoL2S-7B
 
10
 
11
  ## 💡 Overview
12
 
13
+ AutoL2S enables automatically switching between short and long reasoning paths based on input complexity.
14
+ Auto Long-Short Reasoning (AutoL2S), a dynamic and model-agnostic framework that enables LLMs to dynamically compress their generated reasoning
15
+ path based on the complexity of the reasoning question. AutoL2S enables a learned paradigm, in which LLMs themselves can decide when longer reasoning is necessary and when shorter reasoning
16
+ suffices, by training on data annotated with our proposed method, which includes both long and short CoT paths and a special \<EASY\> token (\<specialLong\> in the implementation). We then use <EASY> token to indicate when the model can
17
+ skip generating lengthy CoT reasoning. This proposed annotation strategy can enhance the LLMs’ ability to generate shorter CoT reasoning paths with improved quality after training.
18
+
19
+ This repository contains:
20
 
21
  - Model weights
22
  - Configuration files
23
  - necessary scripts in the `examples/` directory
24
 
25
+ <p align="left">
26
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/66f9bb2dd5575ad6914756ce/dVpIjeIaU8Hv1M5z5VWYS.png" width="35%" style="display:inline-block; margin-right: 10px;" />
27
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/66f9bb2dd5575ad6914756ce/qxHTE-ZGTpxVjmkIX6Fk-.png" width="35%" style="display:inline-block;" />
28
+ </p>
29
 
30
+ ---
31
  ## 🧩 Dependencies
 
32
  We recommend using the model with [vLLM](https://github.com/vllm-project/vllm).
33
  The code has been tested with:
34
 
 
37
  ```
38
 
39
  ---
 
40
  ## 🚀 How to Use
41
 
42
  Run the inference example:
 
82
  journal={arXiv preprint},
83
  year={2025}
84
  }
85
+ ```