quantumfr commited on
Commit
db3ead9
·
verified ·
1 Parent(s): 84a1891

Model save

Browse files
README.md CHANGED
@@ -26,7 +26,7 @@ print(output["generated_text"])
26
 
27
  ## Training procedure
28
 
29
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/quantumcgx-sjtu/huggingface/runs/1i280yqx)
30
 
31
 
32
  This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models](https://huggingface.co/papers/2402.03300).
 
26
 
27
  ## Training procedure
28
 
29
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/quantumcgx-sjtu/huggingface/runs/q7992nd7)
30
 
31
 
32
  This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models](https://huggingface.co/papers/2402.03300).
all_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "total_flos": 0.0,
3
- "train_loss": 0.03796683862837481,
4
- "train_runtime": 12237.2268,
5
  "train_samples": 7500,
6
- "train_samples_per_second": 0.613,
7
  "train_steps_per_second": 0.005
8
  }
 
1
  {
2
  "total_flos": 0.0,
3
+ "train_loss": 0.04025471127784715,
4
+ "train_runtime": 12294.0923,
5
  "train_samples": 7500,
6
+ "train_samples_per_second": 0.61,
7
  "train_steps_per_second": 0.005
8
  }
model-00001-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5ca58a71519cc63005696ca2a2766476d42a1dc0611709f76d10eaa8d472f7cf
3
  size 4877660776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7f34a79b22d7146ceb012375790d8cf75921e2932374539663700d1a5724a4b
3
  size 4877660776
model-00002-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:687be6592543c8cf505c6ebddffa69d98186a673f4542948749e9989677e4263
3
  size 4932751008
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:79126fd3e643eaed94e367207aaf34c6a3dd1d38b711a309c2b457e4471f005b
3
  size 4932751008
model-00003-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3356de55d4d190c4856b36206c3f124f0bf4f7d7f270d60fe5a4f6959a23d217
3
  size 4330865200
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34075134a56e830a924933f0728a0c255399dcee0324ab7cf63d2a3cc5c9ad29
3
  size 4330865200
model-00004-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:555d5a4a5edc7b5e9c399086a329edc1a0101d01eac749df32cece6a0ad74c36
3
  size 1089994880
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e186cd77e5c087ab5c6ef72af5c054e5abb46082077a8f0d81ce0d095400dbf
3
  size 1089994880
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "total_flos": 0.0,
3
- "train_loss": 0.03796683862837481,
4
- "train_runtime": 12237.2268,
5
  "train_samples": 7500,
6
- "train_samples_per_second": 0.613,
7
  "train_steps_per_second": 0.005
8
  }
 
1
  {
2
  "total_flos": 0.0,
3
+ "train_loss": 0.04025471127784715,
4
+ "train_runtime": 12294.0923,
5
  "train_samples": 7500,
6
+ "train_samples_per_second": 0.61,
7
  "train_steps_per_second": 0.005
8
  }
trainer_state.json CHANGED
@@ -12,7 +12,7 @@
12
  "clip_ratio": 0.0,
13
  "completion_length": 594.6719055175781,
14
  "epoch": 0.017057569296375266,
15
- "grad_norm": 0.3894534707069397,
16
  "kl": 0.0,
17
  "learning_rate": 5e-07,
18
  "loss": -0.0015,
@@ -24,172 +24,172 @@
24
  },
25
  {
26
  "clip_ratio": 0.0,
27
- "completion_length": 606.3172702789307,
28
  "epoch": 0.08528784648187633,
29
- "grad_norm": 0.3337987959384918,
30
- "kl": 0.00029712915420532227,
31
  "learning_rate": 2.5e-06,
32
- "loss": 0.0289,
33
- "reward": 0.5892857406288385,
34
- "reward_std": 0.3550556660629809,
35
- "rewards/accuracy_reward": 0.5892857406288385,
36
  "rewards/format_reward": 0.0,
37
  "step": 5
38
  },
39
  {
40
  "clip_ratio": 0.0,
41
- "completion_length": 610.3469032287597,
42
  "epoch": 0.17057569296375266,
43
- "grad_norm": 0.4427908658981323,
44
- "kl": 0.015810632705688478,
45
  "learning_rate": 2.956412726139078e-06,
46
- "loss": 0.0597,
47
- "reward": 0.6875000283122062,
48
- "reward_std": 0.3029041964560747,
49
- "rewards/accuracy_reward": 0.6875000283122062,
50
  "rewards/format_reward": 0.0,
51
  "step": 10
52
  },
53
  {
54
  "clip_ratio": 0.0,
55
- "completion_length": 599.6522605895996,
56
  "epoch": 0.255863539445629,
57
- "grad_norm": 0.33527594804763794,
58
- "kl": 0.005640411376953125,
59
  "learning_rate": 2.7836719084521715e-06,
60
- "loss": 0.0684,
61
- "reward": 0.7642857536673546,
62
- "reward_std": 0.213128218986094,
63
- "rewards/accuracy_reward": 0.7642857536673546,
64
  "rewards/format_reward": 0.0,
65
  "step": 15
66
  },
67
  {
68
  "clip_ratio": 0.0,
69
- "completion_length": 590.715650177002,
70
  "epoch": 0.3411513859275053,
71
- "grad_norm": 0.08924940973520279,
72
- "kl": 0.006121444702148438,
73
  "learning_rate": 2.4946839873611927e-06,
74
- "loss": 0.057,
75
- "reward": 0.7723214641213417,
76
- "reward_std": 0.19404921345412732,
77
- "rewards/accuracy_reward": 0.7723214641213417,
78
  "rewards/format_reward": 0.0,
79
  "step": 20
80
  },
81
  {
82
  "clip_ratio": 0.0,
83
- "completion_length": 595.2143127441407,
84
  "epoch": 0.42643923240938164,
85
- "grad_norm": 0.2607589662075043,
86
- "kl": 0.0135772705078125,
87
  "learning_rate": 2.1156192081791355e-06,
88
- "loss": 0.0401,
89
- "reward": 0.7671875298023224,
90
- "reward_std": 0.18023715307936072,
91
- "rewards/accuracy_reward": 0.7671875298023224,
92
  "rewards/format_reward": 0.0,
93
  "step": 25
94
  },
95
  {
96
  "clip_ratio": 0.0,
97
- "completion_length": 598.6375213623047,
98
  "epoch": 0.511727078891258,
99
- "grad_norm": 0.22541268169879913,
100
- "kl": 0.007162857055664063,
101
  "learning_rate": 1.6808050203829845e-06,
102
- "loss": 0.0346,
103
- "reward": 0.7511161044239998,
104
- "reward_std": 0.17934485003352166,
105
- "rewards/accuracy_reward": 0.7511161044239998,
106
  "rewards/format_reward": 0.0,
107
  "step": 30
108
  },
109
  {
110
  "clip_ratio": 0.0,
111
- "completion_length": 586.3620780944824,
112
  "epoch": 0.5970149253731343,
113
- "grad_norm": 0.12512558698654175,
114
- "kl": 0.013501358032226563,
115
  "learning_rate": 1.2296174432791415e-06,
116
- "loss": 0.0307,
117
- "reward": 0.7457589611411095,
118
- "reward_std": 0.18167408434674143,
119
- "rewards/accuracy_reward": 0.7457589611411095,
120
  "rewards/format_reward": 0.0,
121
  "step": 35
122
  },
123
  {
124
  "clip_ratio": 0.0,
125
- "completion_length": 578.3451164245605,
126
  "epoch": 0.6823027718550106,
127
- "grad_norm": 0.15629476308822632,
128
- "kl": 0.00733642578125,
129
  "learning_rate": 8.029152419343472e-07,
130
- "loss": 0.0295,
131
- "reward": 0.7725446775555611,
132
- "reward_std": 0.16549305748194457,
133
- "rewards/accuracy_reward": 0.7725446775555611,
134
  "rewards/format_reward": 0.0,
135
  "step": 40
136
  },
137
  {
138
  "clip_ratio": 0.0,
139
- "completion_length": 590.9895362854004,
140
  "epoch": 0.767590618336887,
141
- "grad_norm": 0.11738639324903488,
142
- "kl": 0.016034698486328124,
143
  "learning_rate": 4.3933982822017883e-07,
144
- "loss": 0.0317,
145
- "reward": 0.7482143223285675,
146
- "reward_std": 0.1866074649617076,
147
- "rewards/accuracy_reward": 0.7482143223285675,
148
  "rewards/format_reward": 0.0,
149
  "step": 45
150
  },
151
  {
152
  "clip_ratio": 0.0,
153
- "completion_length": 595.0875236511231,
154
  "epoch": 0.8528784648187633,
155
- "grad_norm": 0.5059227347373962,
156
- "kl": 0.01042022705078125,
157
  "learning_rate": 1.718159615201853e-07,
158
- "loss": 0.0294,
159
- "reward": 0.7500000312924385,
160
- "reward_std": 0.18064689869061112,
161
- "rewards/accuracy_reward": 0.7500000312924385,
162
  "rewards/format_reward": 0.0,
163
  "step": 50
164
  },
165
  {
166
  "clip_ratio": 0.0,
167
- "completion_length": 581.5457832336426,
168
  "epoch": 0.9381663113006397,
169
- "grad_norm": 0.13856537640094757,
170
- "kl": 0.018731689453125,
171
  "learning_rate": 2.4570139579284723e-08,
172
- "loss": 0.0236,
173
- "reward": 0.7825893223285675,
174
- "reward_std": 0.18381581027060748,
175
- "rewards/accuracy_reward": 0.7825893223285675,
176
  "rewards/format_reward": 0.0,
177
  "step": 55
178
  },
179
  {
180
  "clip_ratio": 0.0,
181
- "completion_length": 580.0205459594727,
182
  "epoch": 0.9893390191897654,
183
- "kl": 0.008316675821940104,
184
- "reward": 0.7719494452079138,
185
- "reward_std": 0.17934072421242794,
186
- "rewards/accuracy_reward": 0.7719494452079138,
187
  "rewards/format_reward": 0.0,
188
  "step": 58,
189
  "total_flos": 0.0,
190
- "train_loss": 0.03796683862837481,
191
- "train_runtime": 12237.2268,
192
- "train_samples_per_second": 0.613,
193
  "train_steps_per_second": 0.005
194
  }
195
  ],
 
12
  "clip_ratio": 0.0,
13
  "completion_length": 594.6719055175781,
14
  "epoch": 0.017057569296375266,
15
+ "grad_norm": 0.39137744903564453,
16
  "kl": 0.0,
17
  "learning_rate": 5e-07,
18
  "loss": -0.0015,
 
24
  },
25
  {
26
  "clip_ratio": 0.0,
27
+ "completion_length": 607.5834541320801,
28
  "epoch": 0.08528784648187633,
29
+ "grad_norm": 1.0777415037155151,
30
+ "kl": 0.00038692355155944824,
31
  "learning_rate": 2.5e-06,
32
+ "loss": 0.0259,
33
+ "reward": 0.5962611874565482,
34
+ "reward_std": 0.36695866473019123,
35
+ "rewards/accuracy_reward": 0.5962611874565482,
36
  "rewards/format_reward": 0.0,
37
  "step": 5
38
  },
39
  {
40
  "clip_ratio": 0.0,
41
+ "completion_length": 606.7803886413574,
42
  "epoch": 0.17057569296375266,
43
+ "grad_norm": 1.1746565103530884,
44
+ "kl": 0.010793495178222656,
45
  "learning_rate": 2.956412726139078e-06,
46
+ "loss": 0.048,
47
+ "reward": 0.6758928880095482,
48
+ "reward_std": 0.30583293717354537,
49
+ "rewards/accuracy_reward": 0.6758928880095482,
50
  "rewards/format_reward": 0.0,
51
  "step": 10
52
  },
53
  {
54
  "clip_ratio": 0.0,
55
+ "completion_length": 602.4096214294434,
56
  "epoch": 0.255863539445629,
57
+ "grad_norm": 0.1635478436946869,
58
+ "kl": 0.008692169189453125,
59
  "learning_rate": 2.7836719084521715e-06,
60
+ "loss": 0.0619,
61
+ "reward": 0.755357176065445,
62
+ "reward_std": 0.22135366648435592,
63
+ "rewards/accuracy_reward": 0.755357176065445,
64
  "rewards/format_reward": 0.0,
65
  "step": 15
66
  },
67
  {
68
  "clip_ratio": 0.0,
69
+ "completion_length": 593.5116325378418,
70
  "epoch": 0.3411513859275053,
71
+ "grad_norm": 0.3694714605808258,
72
+ "kl": 0.00950794219970703,
73
  "learning_rate": 2.4946839873611927e-06,
74
+ "loss": 0.0543,
75
+ "reward": 0.7654018238186836,
76
+ "reward_std": 0.2084361480548978,
77
+ "rewards/accuracy_reward": 0.7654018238186836,
78
  "rewards/format_reward": 0.0,
79
  "step": 20
80
  },
81
  {
82
  "clip_ratio": 0.0,
83
+ "completion_length": 603.0078392028809,
84
  "epoch": 0.42643923240938164,
85
+ "grad_norm": 0.21624431014060974,
86
+ "kl": 0.004351234436035157,
87
  "learning_rate": 2.1156192081791355e-06,
88
+ "loss": 0.0513,
89
+ "reward": 0.7598214641213417,
90
+ "reward_std": 0.19899208266288043,
91
+ "rewards/accuracy_reward": 0.7598214641213417,
92
  "rewards/format_reward": 0.0,
93
  "step": 25
94
  },
95
  {
96
  "clip_ratio": 0.0,
97
+ "completion_length": 605.0727897644043,
98
  "epoch": 0.511727078891258,
99
+ "grad_norm": 0.26695239543914795,
100
+ "kl": 0.005370330810546875,
101
  "learning_rate": 1.6808050203829845e-06,
102
+ "loss": 0.0455,
103
+ "reward": 0.7566964611411094,
104
+ "reward_std": 0.18958494029939174,
105
+ "rewards/accuracy_reward": 0.7566964611411094,
106
  "rewards/format_reward": 0.0,
107
  "step": 30
108
  },
109
  {
110
  "clip_ratio": 0.0,
111
+ "completion_length": 592.4431076049805,
112
  "epoch": 0.5970149253731343,
113
+ "grad_norm": 0.12381298094987869,
114
+ "kl": 0.004366302490234375,
115
  "learning_rate": 1.2296174432791415e-06,
116
+ "loss": 0.0353,
117
+ "reward": 0.747098246216774,
118
+ "reward_std": 0.18222492672502993,
119
+ "rewards/accuracy_reward": 0.747098246216774,
120
  "rewards/format_reward": 0.0,
121
  "step": 35
122
  },
123
  {
124
  "clip_ratio": 0.0,
125
+ "completion_length": 583.5830627441406,
126
  "epoch": 0.6823027718550106,
127
+ "grad_norm": 0.15292133390903473,
128
+ "kl": 0.003928375244140625,
129
  "learning_rate": 8.029152419343472e-07,
130
+ "loss": 0.0394,
131
+ "reward": 0.7667411059141159,
132
+ "reward_std": 0.1841550744138658,
133
+ "rewards/accuracy_reward": 0.7667411059141159,
134
  "rewards/format_reward": 0.0,
135
  "step": 40
136
  },
137
  {
138
  "clip_ratio": 0.0,
139
+ "completion_length": 599.4895347595215,
140
  "epoch": 0.767590618336887,
141
+ "grad_norm": 0.2122126966714859,
142
+ "kl": 0.004204940795898437,
143
  "learning_rate": 4.3933982822017883e-07,
144
+ "loss": 0.0305,
145
+ "reward": 0.7553571775555611,
146
+ "reward_std": 0.1827343128155917,
147
+ "rewards/accuracy_reward": 0.7553571775555611,
148
  "rewards/format_reward": 0.0,
149
  "step": 45
150
  },
151
  {
152
  "clip_ratio": 0.0,
153
+ "completion_length": 596.6477951049804,
154
  "epoch": 0.8528784648187633,
155
+ "grad_norm": 0.10254240781068802,
156
+ "kl": 0.007993698120117188,
157
  "learning_rate": 1.718159615201853e-07,
158
+ "loss": 0.0327,
159
+ "reward": 0.7506696820259094,
160
+ "reward_std": 0.19283948028460146,
161
+ "rewards/accuracy_reward": 0.7506696820259094,
162
  "rewards/format_reward": 0.0,
163
  "step": 50
164
  },
165
  {
166
  "clip_ratio": 0.0,
167
+ "completion_length": 590.7453369140625,
168
  "epoch": 0.9381663113006397,
169
+ "grad_norm": 0.22994418442249298,
170
+ "kl": 0.00672760009765625,
171
  "learning_rate": 2.4570139579284723e-08,
172
+ "loss": 0.0323,
173
+ "reward": 0.7854911103844643,
174
+ "reward_std": 0.18603479415178298,
175
+ "rewards/accuracy_reward": 0.7854911103844643,
176
  "rewards/format_reward": 0.0,
177
  "step": 55
178
  },
179
  {
180
  "clip_ratio": 0.0,
181
+ "completion_length": 583.9811820983887,
182
  "epoch": 0.9893390191897654,
183
+ "kl": 0.004169464111328125,
184
+ "reward": 0.7682292014360428,
185
+ "reward_std": 0.18959872238337994,
186
+ "rewards/accuracy_reward": 0.7682292014360428,
187
  "rewards/format_reward": 0.0,
188
  "step": 58,
189
  "total_flos": 0.0,
190
+ "train_loss": 0.04025471127784715,
191
+ "train_runtime": 12294.0923,
192
+ "train_samples_per_second": 0.61,
193
  "train_steps_per_second": 0.005
194
  }
195
  ],
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d860344276bc120dcbed788ba38d2cb700cfc46846b3acf5a999df20e714ff13
3
- size 8056
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4c755ecba0508c9f40824f0fbf0d2922ab2bc7d3e57c93d27c3e5b830f38b179
3
+ size 7992