wandermay commited on
Commit
733ed8f
·
verified ·
1 Parent(s): 105607a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -6,7 +6,7 @@ license: mit
6
  ### Model description
7
  Most existing mthods focused on distilling DeepSeek-R1 to improve reasoning ability. However, as far as we know, there is no distilled model could surpass DeepSeek-R1 or QwQ-32B. We introduce NTele-R1-32B-DS , a state-of-the-art mathematical reasoning model that outperforms QwQ-32B across common reasoning benchmarks, including AIME2024/2025, MATH500 and GPQA-Diamond.
8
  Notebly, NTele-R1-32B-DS is the first that achieves **more than 80/70 in challenging AIME2024/2025**.
9
- | Model | Trained From | Release Date | AIME2024(ours/reported) | AIME2025(ours/reported) | MATH500(ours/reported) | GPQA-Diamond(ours/reported) |
10
  |-------|-------|-------|-------|-------|-------|-------|
11
  | QwQ-32B | - | 25.3.6 | 76.25 / 79.5 | 67.30 / - | 94.6 / - | 63.6 / - |
12
  | DeepSeek-32B-Distill | Qwen2.5-32B-Instruct | 25.1.20 | 64.17 / 72.6 | 55.21 / - | 89.8 / 94.3 | 62.1 / 62.1 |
 
6
  ### Model description
7
  Most existing mthods focused on distilling DeepSeek-R1 to improve reasoning ability. However, as far as we know, there is no distilled model could surpass DeepSeek-R1 or QwQ-32B. We introduce NTele-R1-32B-DS , a state-of-the-art mathematical reasoning model that outperforms QwQ-32B across common reasoning benchmarks, including AIME2024/2025, MATH500 and GPQA-Diamond.
8
  Notebly, NTele-R1-32B-DS is the first that achieves **more than 80/70 in challenging AIME2024/2025**.
9
+ | Model | Trained From | Release Date | AIME2024(ours/reported) | AIME2025(o/r) | MATH500(o/r) | GPQA-Diamond(o/r) |
10
  |-------|-------|-------|-------|-------|-------|-------|
11
  | QwQ-32B | - | 25.3.6 | 76.25 / 79.5 | 67.30 / - | 94.6 / - | 63.6 / - |
12
  | DeepSeek-32B-Distill | Qwen2.5-32B-Instruct | 25.1.20 | 64.17 / 72.6 | 55.21 / - | 89.8 / 94.3 | 62.1 / 62.1 |