RabotniKuma commited on
Commit
cc6b91b
·
verified ·
1 Parent(s): d495492

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -8
README.md CHANGED
@@ -12,25 +12,27 @@ base_model:
12
 
13
  # Summary
14
  By applying SFT and GRPO on difficult math problems, we enhanced the performance of `DeepSeek-R1-Distill-Qwen-14B` and developed `Fast-Math-R1-14B`,
15
- which achieves up to 60% faster inference while maintaining accuracy.
16
 
17
  Technical details can be found in [Kaggle Discussion](https://www.kaggle.com/competitions/ai-mathematical-olympiad-progress-prize-2/discussion/571252) and [Github](https://github.com/analokmaus/kaggle-aimo2-fast-math-r1).
18
 
19
- <img src="https://www.googleapis.com/download/storage/v1/b/kaggle-forum-message-attachments/o/inbox%2F1973217%2F2bebc2bf743e7fe92f9e1fa9527220fc%2Fpass1_aime_answers_only.png?generation=1744851657610346&alt=media" max-height="300px">
20
  <img src="https://www.googleapis.com/download/storage/v1/b/kaggle-forum-message-attachments/o/inbox%2F1973217%2F4f221ab914f3e950fa35bdab5723d462%2Fpass1_aime_all.png?generation=1744851665782759&alt=media" max-height="300px">
21
 
22
  | | | AIME 2024 | | AIME 2025 | |
23
  | ---------------------------- | ------------ | ---------------- | ------------- | ---------------- | ------------- |
24
  | Model | Token budget | Pass@1 (avg. 64) | Output tokens | Pass@1 (avg. 64) | Output tokens |
25
  | DeepSeek-R1-Distill-Qwen-14B | 16384 | 63.3 | 9590 | 46.7 | 10602 |
26
- | | 12800 | 58 | 6444 | 41.9 | 6684 |
27
- | | 8192 | 45.6 | 4920 | 30.6 | 4611 |
28
  | Light-R1-14B-DS | 16384 | **66.8** | 10146 | **51.3** | 11308 |
29
- | | 12800 | 59.2 | 6974 | 43.8 | 6869 |
30
- | | 8192 | 42.4 | 5500 | 30.4 | 4908 |
31
  | Fast-Math-R1-14B | 16384 | 66 | **7932** | 49.2 | **9066** |
32
- | | 12800 | **63** | **5996** | **46.1** | **6127** |
33
- | | 8192 | **51.4** | **4269** | **37.2** | **3905** |
 
 
 
34
 
35
 
36
  # Dataset
 
12
 
13
  # Summary
14
  By applying SFT and GRPO on difficult math problems, we enhanced the performance of `DeepSeek-R1-Distill-Qwen-14B` and developed `Fast-Math-R1-14B`,
15
+ which achieves up to 60% (on average approx. 30%) faster inference while maintaining accuracy.
16
 
17
  Technical details can be found in [Kaggle Discussion](https://www.kaggle.com/competitions/ai-mathematical-olympiad-progress-prize-2/discussion/571252) and [Github](https://github.com/analokmaus/kaggle-aimo2-fast-math-r1).
18
 
 
19
  <img src="https://www.googleapis.com/download/storage/v1/b/kaggle-forum-message-attachments/o/inbox%2F1973217%2F4f221ab914f3e950fa35bdab5723d462%2Fpass1_aime_all.png?generation=1744851665782759&alt=media" max-height="300px">
20
 
21
  | | | AIME 2024 | | AIME 2025 | |
22
  | ---------------------------- | ------------ | ---------------- | ------------- | ---------------- | ------------- |
23
  | Model | Token budget | Pass@1 (avg. 64) | Output tokens | Pass@1 (avg. 64) | Output tokens |
24
  | DeepSeek-R1-Distill-Qwen-14B | 16384 | 63.3 | 9590 | 46.7 | 10602 |
25
+ | | 12800 | 58 | 8632 | 41.9 | 9363 |
26
+ | | 8192 | 45.6 | 6638 | 30.6 | 6897 |
27
  | Light-R1-14B-DS | 16384 | **66.8** | 10146 | **51.3** | 11308 |
28
+ | | 12800 | 59.2 | 9110 | 43.8 | 9834 |
29
+ | | 8192 | 42.4 | 7020 | 30.4 | 7124 |
30
  | Fast-Math-R1-14B | 16384 | 66 | **7932** | 49.2 | **9066** |
31
+ | | 12800 | **63** | **7449** | **46.1** | **8282** |
32
+ | | 8192 | **51.4** | **5963** | **37.2** | **6256** |
33
+ | Fast-Math-R1-14B-SFT Only | 16384 | 65.2 | 10268 | 49.7 | 11264 |
34
+ | | 12800 | 57.2 | 9180 | 42.8 | 9805 |
35
+ | | 8192 | 41.3 | 7015 | 30.1 | 7074 |
36
 
37
 
38
  # Dataset