nielsr HF Staff commited on
Commit
1910866
·
verified ·
1 Parent(s): 0dccf55

Improve model card: Add paper URL, pipeline tag and Github URL

Browse files

This PR improves the model card by adding the paper URL, the pipeline tag, and the Github URL. It also adds the blog post URLs in the description.

Files changed (1) hide show
  1. README.md +9 -5
README.md CHANGED
@@ -1,14 +1,15 @@
1
  ---
2
- library_name: transformers
 
 
3
  datasets:
4
  - BAAI/TACO
5
  - tasksource/PRM800K
6
  language:
7
  - en
8
- base_model:
9
- - Qwen/Qwen2.5-32B-Instruct
10
- - NovaSky-AI/Sky-T1-32B-Preview
11
  license: apache-2.0
 
12
  ---
13
 
14
  ## Model Details
@@ -18,9 +19,11 @@ license: apache-2.0
18
  <!-- Provide a longer summary of what this model is. -->
19
 
20
  This is a 32B reasoning model preference optimized on top of Sky-T1-32B-Preview to significantly reduce generation lengths while maintaining accuracy. The performance is on par with o1-preview model in both math and coding, while reducing generation lengths by up to 57% relative to Sky-T1-32B-Preview.
21
- Please see our [blog post](https://novasky-ai.github.io/posts/reduce-overthinking/) for more details.
22
 
23
  - **Developed by:** NovaSky Team from Sky Computing Lab at UC Berkeley.
 
 
24
 
25
  ## Training Details
26
 
@@ -71,3 +74,4 @@ Please considering citing our blog post if you found it useful for your research
71
  note = {Accessed: 2025-01-23},
72
  year = {2025}
73
  }
 
 
1
  ---
2
+ base_model:
3
+ - Qwen/Qwen2.5-32B-Instruct
4
+ - NovaSky-AI/Sky-T1-32B-Preview
5
  datasets:
6
  - BAAI/TACO
7
  - tasksource/PRM800K
8
  language:
9
  - en
10
+ library_name: transformers
 
 
11
  license: apache-2.0
12
+ pipeline_tag: text-generation
13
  ---
14
 
15
  ## Model Details
 
19
  <!-- Provide a longer summary of what this model is. -->
20
 
21
  This is a 32B reasoning model preference optimized on top of Sky-T1-32B-Preview to significantly reduce generation lengths while maintaining accuracy. The performance is on par with o1-preview model in both math and coding, while reducing generation lengths by up to 57% relative to Sky-T1-32B-Preview.
22
+ Please see our [blog post](https://novasky-ai.github.io/posts/reduce-overthinking/) and [Sky-T1 blog post](https://novasky-ai.github.io/posts/sky-t1/) for more details.
23
 
24
  - **Developed by:** NovaSky Team from Sky Computing Lab at UC Berkeley.
25
+ - **Paper:** [LLMs Can Easily Learn to Reason from Demonstrations Structure, not content, is what matters!](https://hf.co/papers/2502.07374)
26
+ - **Code:** [https://github.com/NovaSky-AI/SkyThought](https://github.com/NovaSky-AI/SkyThought)
27
 
28
  ## Training Details
29
 
 
74
  note = {Accessed: 2025-01-23},
75
  year = {2025}
76
  }
77
+ ```