Improve model card: Add paper URL, pipeline tag and Github URL
Browse filesThis PR improves the model card by adding the paper URL, the pipeline tag, and the Github URL. It also adds the blog post URLs in the description.
README.md
CHANGED
|
@@ -1,14 +1,15 @@
|
|
| 1 |
---
|
| 2 |
-
|
|
|
|
|
|
|
| 3 |
datasets:
|
| 4 |
- BAAI/TACO
|
| 5 |
- tasksource/PRM800K
|
| 6 |
language:
|
| 7 |
- en
|
| 8 |
-
|
| 9 |
-
- Qwen/Qwen2.5-32B-Instruct
|
| 10 |
-
- NovaSky-AI/Sky-T1-32B-Preview
|
| 11 |
license: apache-2.0
|
|
|
|
| 12 |
---
|
| 13 |
|
| 14 |
## Model Details
|
|
@@ -18,9 +19,11 @@ license: apache-2.0
|
|
| 18 |
<!-- Provide a longer summary of what this model is. -->
|
| 19 |
|
| 20 |
This is a 32B reasoning model preference optimized on top of Sky-T1-32B-Preview to significantly reduce generation lengths while maintaining accuracy. The performance is on par with o1-preview model in both math and coding, while reducing generation lengths by up to 57% relative to Sky-T1-32B-Preview.
|
| 21 |
-
Please see our [blog post](https://novasky-ai.github.io/posts/reduce-overthinking/) for more details.
|
| 22 |
|
| 23 |
- **Developed by:** NovaSky Team from Sky Computing Lab at UC Berkeley.
|
|
|
|
|
|
|
| 24 |
|
| 25 |
## Training Details
|
| 26 |
|
|
@@ -71,3 +74,4 @@ Please considering citing our blog post if you found it useful for your research
|
|
| 71 |
note = {Accessed: 2025-01-23},
|
| 72 |
year = {2025}
|
| 73 |
}
|
|
|
|
|
|
| 1 |
---
|
| 2 |
+
base_model:
|
| 3 |
+
- Qwen/Qwen2.5-32B-Instruct
|
| 4 |
+
- NovaSky-AI/Sky-T1-32B-Preview
|
| 5 |
datasets:
|
| 6 |
- BAAI/TACO
|
| 7 |
- tasksource/PRM800K
|
| 8 |
language:
|
| 9 |
- en
|
| 10 |
+
library_name: transformers
|
|
|
|
|
|
|
| 11 |
license: apache-2.0
|
| 12 |
+
pipeline_tag: text-generation
|
| 13 |
---
|
| 14 |
|
| 15 |
## Model Details
|
|
|
|
| 19 |
<!-- Provide a longer summary of what this model is. -->
|
| 20 |
|
| 21 |
This is a 32B reasoning model preference optimized on top of Sky-T1-32B-Preview to significantly reduce generation lengths while maintaining accuracy. The performance is on par with o1-preview model in both math and coding, while reducing generation lengths by up to 57% relative to Sky-T1-32B-Preview.
|
| 22 |
+
Please see our [blog post](https://novasky-ai.github.io/posts/reduce-overthinking/) and [Sky-T1 blog post](https://novasky-ai.github.io/posts/sky-t1/) for more details.
|
| 23 |
|
| 24 |
- **Developed by:** NovaSky Team from Sky Computing Lab at UC Berkeley.
|
| 25 |
+
- **Paper:** [LLMs Can Easily Learn to Reason from Demonstrations Structure, not content, is what matters!](https://hf.co/papers/2502.07374)
|
| 26 |
+
- **Code:** [https://github.com/NovaSky-AI/SkyThought](https://github.com/NovaSky-AI/SkyThought)
|
| 27 |
|
| 28 |
## Training Details
|
| 29 |
|
|
|
|
| 74 |
note = {Accessed: 2025-01-23},
|
| 75 |
year = {2025}
|
| 76 |
}
|
| 77 |
+
```
|