chenggong commited on
Commit
020a9a6
·
verified ·
1 Parent(s): bbf0040

Model save

Browse files
README.md CHANGED
@@ -27,7 +27,7 @@ print(output["generated_text"])
27
 
28
  ## Training procedure
29
 
30
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/gongc1995-city-university-of-hong-kong/huggingface/runs/tw88l9f9)
31
 
32
 
33
  This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models](https://huggingface.co/papers/2402.03300).
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/gongc1995-city-university-of-hong-kong/huggingface/runs/o4ou28au)
31
 
32
 
33
  This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models](https://huggingface.co/papers/2402.03300).
all_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "total_flos": 0.0,
3
- "train_loss": 0.04267983935002623,
4
- "train_runtime": 7907.107,
5
  "train_samples": 7500,
6
- "train_samples_per_second": 0.949,
7
  "train_steps_per_second": 0.007
8
  }
 
1
  {
2
  "total_flos": 0.0,
3
+ "train_loss": 0.04428564981910689,
4
+ "train_runtime": 7921.619,
5
  "train_samples": 7500,
6
+ "train_samples_per_second": 0.947,
7
  "train_steps_per_second": 0.007
8
  }
model-00001-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8da2c323738051ed0b639c6c6c3cabc3a7186aa2833f3340294351c8e4a4abeb
3
  size 4877660776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8cb74df1f51c7b5cf6f92df96b571f86a5d24c12666a82626c80881f07f102a7
3
  size 4877660776
model-00002-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9be10ba534848b1d7712651d761f47d9857ce2deafdf659b3e64780412dab1ae
3
  size 4932751008
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1562089ab4298c8e3ab1d9b5112aa468472695bc876edbaa0d3da760884391b4
3
  size 4932751008
model-00003-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ab0d56ac8ae6378b286f76fdbd6b9d2c2194938d9d386ce97316567d20706da4
3
  size 4330865200
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97ddbf5e2431a28323e4f5261064c74d112b605caf7bc43ae5eb36599756d68b
3
  size 4330865200
model-00004-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:95098f0abc1df599d0b888c5de8f72d3da1a83efe178a52514487eefbaea1fbb
3
  size 1089994880
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:012dcf0ed4f1423109622326588a9ee01e2288cffd2ee73564633813291ea143
3
  size 1089994880
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "total_flos": 0.0,
3
- "train_loss": 0.04267983935002623,
4
- "train_runtime": 7907.107,
5
  "train_samples": 7500,
6
- "train_samples_per_second": 0.949,
7
  "train_steps_per_second": 0.007
8
  }
 
1
  {
2
  "total_flos": 0.0,
3
+ "train_loss": 0.04428564981910689,
4
+ "train_runtime": 7921.619,
5
  "train_samples": 7500,
6
+ "train_samples_per_second": 0.947,
7
  "train_steps_per_second": 0.007
8
  }
trainer_state.json CHANGED
@@ -10,186 +10,186 @@
10
  "log_history": [
11
  {
12
  "clip_ratio": 0.0,
13
- "completion_length": 611.475471496582,
14
  "epoch": 0.017057569296375266,
15
- "grad_norm": 0.3608338534832001,
16
  "kl": 0.0,
17
  "learning_rate": 5e-07,
18
- "loss": 0.0485,
19
  "reward": 0.7064732536673546,
20
- "reward_std": 0.29637950100004673,
21
- "rewards/accuracy_reward": 0.6975446715950966,
22
- "rewards/format_reward": 0.008928571944124997,
23
  "step": 1
24
  },
25
  {
26
  "clip_ratio": 0.0,
27
- "completion_length": 603.289924621582,
28
  "epoch": 0.08528784648187633,
29
- "grad_norm": 0.24193009734153748,
30
- "kl": 0.00037926435470581055,
31
  "learning_rate": 2.5e-06,
32
- "loss": 0.0419,
33
- "reward": 0.6629464607685804,
34
- "reward_std": 0.3164387275464833,
35
- "rewards/accuracy_reward": 0.6540178805589676,
36
- "rewards/format_reward": 0.008928571856813505,
37
  "step": 5
38
  },
39
  {
40
  "clip_ratio": 0.0,
41
- "completion_length": 602.1417655944824,
42
  "epoch": 0.17057569296375266,
43
- "grad_norm": 0.21082398295402527,
44
- "kl": 0.0039100170135498045,
45
  "learning_rate": 2.956412726139078e-06,
46
- "loss": 0.0591,
47
- "reward": 0.6986607477068901,
48
- "reward_std": 0.29480395577847957,
49
- "rewards/accuracy_reward": 0.6926339596509934,
50
- "rewards/format_reward": 0.006026786076836288,
51
  "step": 10
52
  },
53
  {
54
  "clip_ratio": 0.0,
55
- "completion_length": 597.1254707336426,
56
  "epoch": 0.255863539445629,
57
- "grad_norm": 0.11069315671920776,
58
- "kl": 0.010307502746582032,
59
  "learning_rate": 2.7836719084521715e-06,
60
- "loss": 0.0681,
61
- "reward": 0.7524553880095481,
62
- "reward_std": 0.22803126517683267,
63
- "rewards/accuracy_reward": 0.7508928924798965,
64
- "rewards/format_reward": 0.0015625000698491931,
65
  "step": 15
66
  },
67
  {
68
  "clip_ratio": 0.0,
69
- "completion_length": 586.2643089294434,
70
  "epoch": 0.3411513859275053,
71
- "grad_norm": 0.5713587999343872,
72
- "kl": 0.00367889404296875,
73
  "learning_rate": 2.4946839873611927e-06,
74
- "loss": 0.058,
75
- "reward": 0.7647321790456771,
76
- "reward_std": 0.21236355928704143,
77
- "rewards/accuracy_reward": 0.7631696805357933,
78
- "rewards/format_reward": 0.0015625000698491931,
79
  "step": 20
80
  },
81
  {
82
  "clip_ratio": 0.0,
83
- "completion_length": 596.1701164245605,
84
  "epoch": 0.42643923240938164,
85
- "grad_norm": 0.2316303700208664,
86
- "kl": 0.007324790954589844,
87
  "learning_rate": 2.1156192081791355e-06,
88
- "loss": 0.0476,
89
- "reward": 0.7662946805357933,
90
- "reward_std": 0.19612560272216797,
91
- "rewards/accuracy_reward": 0.761383967101574,
92
- "rewards/format_reward": 0.004910714505240321,
93
  "step": 25
94
  },
95
  {
96
  "clip_ratio": 0.0,
97
- "completion_length": 599.6569450378418,
98
  "epoch": 0.511727078891258,
99
- "grad_norm": 23.480575561523438,
100
- "kl": 0.018339920043945312,
101
  "learning_rate": 1.6808050203829845e-06,
102
- "loss": 0.0385,
103
- "reward": 0.7616071805357933,
104
- "reward_std": 0.21710715610533954,
105
- "rewards/accuracy_reward": 0.7415178880095482,
106
- "rewards/format_reward": 0.02008928672876209,
107
  "step": 30
108
  },
109
  {
110
  "clip_ratio": 0.0,
111
- "completion_length": 584.1323913574219,
112
  "epoch": 0.5970149253731343,
113
- "grad_norm": 0.18863163888454437,
114
- "kl": 0.012631607055664063,
115
  "learning_rate": 1.2296174432791415e-06,
116
- "loss": 0.0322,
117
- "reward": 0.7970982536673545,
118
- "reward_std": 0.2728703811764717,
119
- "rewards/accuracy_reward": 0.7404018193483353,
120
- "rewards/format_reward": 0.056696431199088695,
121
  "step": 35
122
  },
123
  {
124
  "clip_ratio": 0.0,
125
- "completion_length": 575.9098495483398,
126
  "epoch": 0.6823027718550106,
127
- "grad_norm": 0.19670729339122772,
128
- "kl": 0.008755111694335937,
129
  "learning_rate": 8.029152419343472e-07,
130
- "loss": 0.0332,
131
- "reward": 0.8162946745753288,
132
- "reward_std": 0.2680498031899333,
133
- "rewards/accuracy_reward": 0.7600446820259095,
134
- "rewards/format_reward": 0.056250002863816914,
135
  "step": 40
136
  },
137
  {
138
  "clip_ratio": 0.0,
139
- "completion_length": 590.6379707336425,
140
  "epoch": 0.767590618336887,
141
- "grad_norm": 0.377120703458786,
142
- "kl": 0.013896942138671875,
143
  "learning_rate": 4.3933982822017883e-07,
144
- "loss": 0.0296,
145
- "reward": 0.8042411118745804,
146
- "reward_std": 0.3068106513470411,
147
- "rewards/accuracy_reward": 0.7363839611411095,
148
- "rewards/format_reward": 0.06785714637953788,
149
  "step": 45
150
  },
151
  {
152
  "clip_ratio": 0.0,
153
- "completion_length": 589.7989082336426,
154
  "epoch": 0.8528784648187633,
155
- "grad_norm": 0.6106008887290955,
156
- "kl": 0.013324737548828125,
157
  "learning_rate": 1.718159615201853e-07,
158
- "loss": 0.0311,
159
- "reward": 0.8252232551574707,
160
- "reward_std": 0.30773016847670076,
161
- "rewards/accuracy_reward": 0.7352678969502449,
162
- "rewards/format_reward": 0.08995536118745803,
163
  "step": 50
164
  },
165
  {
166
  "clip_ratio": 0.0,
167
- "completion_length": 581.0469017028809,
168
  "epoch": 0.9381663113006397,
169
- "grad_norm": 0.17101669311523438,
170
- "kl": 0.017275238037109376,
171
  "learning_rate": 2.4570139579284723e-08,
172
- "loss": 0.0434,
173
- "reward": 0.8486607581377029,
174
- "reward_std": 0.3164860598742962,
175
- "rewards/accuracy_reward": 0.7607143178582192,
176
- "rewards/format_reward": 0.08794643317814917,
177
  "step": 55
178
  },
179
  {
180
  "clip_ratio": 0.0,
181
- "completion_length": 572.9547259012858,
182
  "epoch": 0.9893390191897654,
183
- "kl": 0.011492411295572916,
184
- "reward": 0.8437500322858492,
185
- "reward_std": 0.3019534188012282,
186
- "rewards/accuracy_reward": 0.7477678855260214,
187
- "rewards/format_reward": 0.0959821470702688,
188
  "step": 58,
189
  "total_flos": 0.0,
190
- "train_loss": 0.04267983935002623,
191
- "train_runtime": 7907.107,
192
- "train_samples_per_second": 0.949,
193
  "train_steps_per_second": 0.007
194
  }
195
  ],
 
10
  "log_history": [
11
  {
12
  "clip_ratio": 0.0,
13
+ "completion_length": 607.8761520385742,
14
  "epoch": 0.017057569296375266,
15
+ "grad_norm": 0.21163204312324524,
16
  "kl": 0.0,
17
  "learning_rate": 5e-07,
18
+ "loss": 0.0347,
19
  "reward": 0.7064732536673546,
20
+ "reward_std": 0.2907280754297972,
21
+ "rewards/accuracy_reward": 0.6964286044239998,
22
+ "rewards/format_reward": 0.010044643422588706,
23
  "step": 1
24
  },
25
  {
26
  "clip_ratio": 0.0,
27
+ "completion_length": 598.3354082107544,
28
  "epoch": 0.08528784648187633,
29
+ "grad_norm": 0.22014343738555908,
30
+ "kl": 0.00016763806343078613,
31
  "learning_rate": 2.5e-06,
32
+ "loss": 0.0443,
33
+ "reward": 0.6668527107685804,
34
+ "reward_std": 0.30428835609927773,
35
+ "rewards/accuracy_reward": 0.6609933376312256,
36
+ "rewards/format_reward": 0.005859375291038305,
37
  "step": 5
38
  },
39
  {
40
  "clip_ratio": 0.0,
41
+ "completion_length": 602.3207862854003,
42
  "epoch": 0.17057569296375266,
43
+ "grad_norm": 0.1867409348487854,
44
+ "kl": 0.006086993217468262,
45
  "learning_rate": 2.956412726139078e-06,
46
+ "loss": 0.0561,
47
+ "reward": 0.6991071760654449,
48
+ "reward_std": 0.28411166220903394,
49
+ "rewards/accuracy_reward": 0.6957589611411095,
50
+ "rewards/format_reward": 0.0033482144586741923,
51
  "step": 10
52
  },
53
  {
54
  "clip_ratio": 0.0,
55
+ "completion_length": 591.975025177002,
56
  "epoch": 0.255863539445629,
57
+ "grad_norm": 0.272446870803833,
58
+ "kl": 0.006195259094238281,
59
  "learning_rate": 2.7836719084521715e-06,
60
+ "loss": 0.0687,
61
+ "reward": 0.7625000327825546,
62
+ "reward_std": 0.21445324290543794,
63
+ "rewards/accuracy_reward": 0.7611607477068901,
64
+ "rewards/format_reward": 0.0013392857741564511,
65
  "step": 15
66
  },
67
  {
68
  "clip_ratio": 0.0,
69
+ "completion_length": 580.5134178161621,
70
  "epoch": 0.3411513859275053,
71
+ "grad_norm": 0.4638945162296295,
72
+ "kl": 0.0028181076049804688,
73
  "learning_rate": 2.4946839873611927e-06,
74
+ "loss": 0.0489,
75
+ "reward": 0.7725446745753288,
76
+ "reward_std": 0.1972122782841325,
77
+ "rewards/accuracy_reward": 0.7714286059141159,
78
+ "rewards/format_reward": 0.0011160714784637094,
79
  "step": 20
80
  },
81
  {
82
  "clip_ratio": 0.0,
83
+ "completion_length": 586.8551612854004,
84
  "epoch": 0.42643923240938164,
85
+ "grad_norm": 0.13729101419448853,
86
+ "kl": 0.0033502578735351562,
87
  "learning_rate": 2.1156192081791355e-06,
88
+ "loss": 0.0425,
89
+ "reward": 0.7703125372529029,
90
+ "reward_std": 0.19164941012859343,
91
+ "rewards/accuracy_reward": 0.769866107404232,
92
+ "rewards/format_reward": 0.00044642859138548373,
93
  "step": 25
94
  },
95
  {
96
  "clip_ratio": 0.0,
97
+ "completion_length": 589.1917663574219,
98
  "epoch": 0.511727078891258,
99
+ "grad_norm": 0.10644034296274185,
100
+ "kl": 0.004234695434570312,
101
  "learning_rate": 1.6808050203829845e-06,
102
+ "loss": 0.034,
103
+ "reward": 0.7531250387430191,
104
+ "reward_std": 0.18948373831808568,
105
+ "rewards/accuracy_reward": 0.7511161044239998,
106
+ "rewards/format_reward": 0.002008928661234677,
107
  "step": 30
108
  },
109
  {
110
  "clip_ratio": 0.0,
111
+ "completion_length": 578.0815017700195,
112
  "epoch": 0.5970149253731343,
113
+ "grad_norm": 0.09403481334447861,
114
+ "kl": 0.004046249389648438,
115
  "learning_rate": 1.2296174432791415e-06,
116
+ "loss": 0.0411,
117
+ "reward": 0.7562500327825546,
118
+ "reward_std": 0.18806953858584166,
119
+ "rewards/accuracy_reward": 0.7495536088943482,
120
+ "rewards/format_reward": 0.006696428894065321,
121
  "step": 35
122
  },
123
  {
124
  "clip_ratio": 0.0,
125
+ "completion_length": 564.9935516357422,
126
  "epoch": 0.6823027718550106,
127
+ "grad_norm": 0.22079423069953918,
128
+ "kl": 0.00465545654296875,
129
  "learning_rate": 8.029152419343472e-07,
130
+ "loss": 0.0326,
131
+ "reward": 0.7790178969502449,
132
+ "reward_std": 0.18834841772913932,
133
+ "rewards/accuracy_reward": 0.7680803954601287,
134
+ "rewards/format_reward": 0.010937500628642738,
135
  "step": 40
136
  },
137
  {
138
  "clip_ratio": 0.0,
139
+ "completion_length": 582.787744140625,
140
  "epoch": 0.767590618336887,
141
+ "grad_norm": 0.16064484417438507,
142
+ "kl": 0.005054092407226563,
143
  "learning_rate": 4.3933982822017883e-07,
144
+ "loss": 0.0423,
145
+ "reward": 0.7665178939700127,
146
+ "reward_std": 0.22603920232504607,
147
+ "rewards/accuracy_reward": 0.7470982491970062,
148
+ "rewards/format_reward": 0.019419643795117736,
149
  "step": 45
150
  },
151
  {
152
  "clip_ratio": 0.0,
153
+ "completion_length": 580.5013664245605,
154
  "epoch": 0.8528784648187633,
155
+ "grad_norm": 0.18539777398109436,
156
+ "kl": 0.00699310302734375,
157
  "learning_rate": 1.718159615201853e-07,
158
+ "loss": 0.0301,
159
+ "reward": 0.7915178954601287,
160
+ "reward_std": 0.2188779940828681,
161
+ "rewards/accuracy_reward": 0.755357176065445,
162
+ "rewards/format_reward": 0.036160716065205635,
163
  "step": 50
164
  },
165
  {
166
  "clip_ratio": 0.0,
167
+ "completion_length": 580.8031463623047,
168
  "epoch": 0.9381663113006397,
169
+ "grad_norm": 0.2504195272922516,
170
+ "kl": 0.009212875366210937,
171
  "learning_rate": 2.4570139579284723e-08,
172
+ "loss": 0.038,
173
+ "reward": 0.8113839596509933,
174
+ "reward_std": 0.2401964722201228,
175
+ "rewards/accuracy_reward": 0.7790178954601288,
176
+ "rewards/format_reward": 0.03236607303842902,
177
  "step": 55
178
  },
179
  {
180
  "clip_ratio": 0.0,
181
+ "completion_length": 567.5023994445801,
182
  "epoch": 0.9893390191897654,
183
+ "kl": 0.011606852213541666,
184
+ "reward": 0.8020833730697632,
185
+ "reward_std": 0.23279494047164917,
186
+ "rewards/accuracy_reward": 0.7678571765621504,
187
+ "rewards/format_reward": 0.03422619208383063,
188
  "step": 58,
189
  "total_flos": 0.0,
190
+ "train_loss": 0.04428564981910689,
191
+ "train_runtime": 7921.619,
192
+ "train_samples_per_second": 0.947,
193
  "train_steps_per_second": 0.007
194
  }
195
  ],
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3af35ec78fd740b874622981c9d0bef9078fe3b060a798afef75bd2b09a2a4ce
3
  size 7992
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c472b231fb3fda1371ef81b463d7f886a525b30b76aebce96a985135998daf2
3
  size 7992