Text Generation
Transformers
Safetensors
qwen2
conversational
text-generation-inference
Vinnnf commited on
Commit
c34c51c
·
verified ·
1 Parent(s): ae0d674

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +26 -1
README.md CHANGED
@@ -14,7 +14,32 @@ library_name: transformers
14
 
15
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/646a1939c37ca1e12308fe81/SRxJKkSuC0y-oMB7SFeR6.png)
16
 
17
- [[**ArXiv**]]() | [[**GitHub**](https://github.com/VainF/Thinkless)]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
18
 
19
  > [!IMPORTANT]
20
  > This is a warm-up model and should be used as an initialization for RL. It was trained on [OpenThoughts-1M-Hybrid-1.5B](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M) and can generate both long and short answers with comparable probabilities (~50%).
 
14
 
15
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/646a1939c37ca1e12308fe81/SRxJKkSuC0y-oMB7SFeR6.png)
16
 
17
+ <table>
18
+ <thead>
19
+ </thead>
20
+ <tbody>
21
+ <tr>
22
+ <td>📄 <strong>Paper Link</strong></td>
23
+ <td><a href="https://arxiv.org/abs/">ArXiv</a></td>
24
+ </tr>
25
+ <tr>
26
+ <td>🤖 <strong>RL Model</strong></td>
27
+ <td><a href="https://huggingface.co/Vinnnf/Thinkless-1.5B-RL-DeepScaleR">Thinkless-1.5B-RL-DeepScaleR</a></td>
28
+ </tr>
29
+ <tr>
30
+ <td>🐣 <strong>Warmup Model</strong></td>
31
+ <td><a href="https://huggingface.co/Vinnnf/Thinkless-1.5B-Warmup">Thinkless-1.5B-Warmup</a></td>
32
+ </tr>
33
+ <tr>
34
+ <td>📊 <strong>Data for Warmup</strong></td>
35
+ <td><a href="https://huggingface.co/datasets/Vinnnf/Hybrid-OpenThoughts-1M-1.5B">Hybrid-OpenThoughts-1M-1.5B</a></td>
36
+ </tr>
37
+ <tr>
38
+ <td>📊 <strong>Data for RL</strong></td>
39
+ <td><a href="https://huggingface.co/datasets/agentica-org/DeepScaleR-Preview-Dataset">agentica-org/DeepScaleR-Preview-Dataset</a></td>
40
+ </tr>
41
+ </tbody>
42
+ </table>
43
 
44
  > [!IMPORTANT]
45
  > This is a warm-up model and should be used as an initialization for RL. It was trained on [OpenThoughts-1M-Hybrid-1.5B](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M) and can generate both long and short answers with comparable probabilities (~50%).