Update README.md
Browse files
README.md
CHANGED
|
@@ -1,7 +1,7 @@
|
|
| 1 |
# **PCL-Reasoner-V1**
|
| 2 |
|
| 3 |
## Model Overview
|
| 4 |
-
We release PCL-Reasoner-V1, a model trained based on Qwen2.5-32B-Base and undergoes high-performance supervised fine-tuning based on the MindSpore framework and Ascend hardware. After fine-tuning, the model demonstrates significant improvements in mathematical reasoning capabilities. PCL-Reasoner-V1 achieves 85.7% and 84.2% respectively on AIME 24 and AIME 25, which position PCL-Reasoner-V1 among the top-tier models in the 32B parameter class.
|
| 5 |
|
| 6 |
We have fully open-sourced the model weights, dataset and training code. Follow the tutorial below to deploy and explore post-training!
|
| 7 |
|
|
@@ -12,8 +12,8 @@ https://github.com/PCL-Reasoner/V1
|
|
| 12 |
|
| 13 |
https://openi.pcl.ac.cn/PCL-Reasoner/V1
|
| 14 |
|
| 15 |
-
#### Evaluation
|
| 16 |
-
|
| 17 |
|
| 18 |
|
| 19 |
<table>
|
|
|
|
| 1 |
# **PCL-Reasoner-V1**
|
| 2 |
|
| 3 |
## Model Overview
|
| 4 |
+
We release PCL-Reasoner-V1, a model trained based on Qwen2.5-32B-Base and undergoes high-performance supervised fine-tuning based on the MindSpore framework and Ascend hardware. After fine-tuning, the model demonstrates significant improvements in mathematical reasoning capabilities. PCL-Reasoner-V1 achieves 85.7% and 84.2% respectively on AIME 24 and AIME 25, which position PCL-Reasoner-V1 among the top-tier models in the 32B parameter class on AIME24/25.
|
| 5 |
|
| 6 |
We have fully open-sourced the model weights, dataset and training code. Follow the tutorial below to deploy and explore post-training!
|
| 7 |
|
|
|
|
| 12 |
|
| 13 |
https://openi.pcl.ac.cn/PCL-Reasoner/V1
|
| 14 |
|
| 15 |
+
#### Evaluation
|
| 16 |
+
We used the **Avg@32 metric** (averaging 32 sampling attempts per query) for evaluation
|
| 17 |
|
| 18 |
|
| 19 |
<table>
|