besimray commited on
Commit
bea7660
·
verified ·
1 Parent(s): e791322

Training in progress, step 40, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0bf014d1a50f271c41f7422261b08ca5ec84dc1faabd04c29231ef2836d36445
3
  size 45118424
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e41d69019d421204220964865c9c89fc7bacb60a8411908a8b29c0a12114b94a
3
  size 45118424
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b9606721fc8617b61d2e0dc2ec8042ef4e6afb22d62a62595ce9ec2026c0ad30
3
  size 23159290
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1928572a9ac56c6d98ea8953c7a7a8fdca570b8a20a3959f47990cc5be36ddc3
3
  size 23159290
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:490faae3574e0545627c6c066345113a5ec4be88337cd4484537a0d75c6be16a
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e565442590f72e02c81245f841f0d570f0816c276d51ef5abfd2bed9cc00d28
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1c297c5cf11a27c75d9f99f1df69752f78c3ad41b0275adf50cdd1b67f9d0bb3
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:062016b917f4cd81cef6ab15bfe81df4e94586c0afba8905b655deaa6fff468f
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "best_metric": 0.8169272541999817,
3
- "best_model_checkpoint": "miner_id_besimray/checkpoint-20",
4
- "epoch": 0.2572347266881029,
5
  "eval_steps": 20,
6
- "global_step": 20,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -163,6 +163,154 @@
163
  "eval_samples_per_second": 28.427,
164
  "eval_steps_per_second": 2.947,
165
  "step": 20
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
166
  }
167
  ],
168
  "logging_steps": 1,
@@ -191,7 +339,7 @@
191
  "attributes": {}
192
  }
193
  },
194
- "total_flos": 5141026150809600.0,
195
  "train_batch_size": 10,
196
  "trial_name": null,
197
  "trial_params": null
 
1
  {
2
+ "best_metric": 0.7905269861221313,
3
+ "best_model_checkpoint": "miner_id_besimray/checkpoint-40",
4
+ "epoch": 0.5144694533762058,
5
  "eval_steps": 20,
6
+ "global_step": 40,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
163
  "eval_samples_per_second": 28.427,
164
  "eval_steps_per_second": 2.947,
165
  "step": 20
166
+ },
167
+ {
168
+ "epoch": 0.27009646302250806,
169
+ "grad_norm": 0.20886819064617157,
170
+ "learning_rate": 0.0001969689520376687,
171
+ "loss": 0.7039,
172
+ "step": 21
173
+ },
174
+ {
175
+ "epoch": 0.2829581993569132,
176
+ "grad_norm": 0.2352903187274933,
177
+ "learning_rate": 0.00019639628606958533,
178
+ "loss": 0.77,
179
+ "step": 22
180
+ },
181
+ {
182
+ "epoch": 0.2958199356913183,
183
+ "grad_norm": 0.2201235145330429,
184
+ "learning_rate": 0.00019577508166849304,
185
+ "loss": 0.7533,
186
+ "step": 23
187
+ },
188
+ {
189
+ "epoch": 0.3086816720257235,
190
+ "grad_norm": 0.20297250151634216,
191
+ "learning_rate": 0.00019510565162951537,
192
+ "loss": 0.7303,
193
+ "step": 24
194
+ },
195
+ {
196
+ "epoch": 0.3215434083601286,
197
+ "grad_norm": 0.20995450019836426,
198
+ "learning_rate": 0.00019438833303083678,
199
+ "loss": 0.8122,
200
+ "step": 25
201
+ },
202
+ {
203
+ "epoch": 0.33440514469453375,
204
+ "grad_norm": 0.21005098521709442,
205
+ "learning_rate": 0.00019362348706397373,
206
+ "loss": 0.8264,
207
+ "step": 26
208
+ },
209
+ {
210
+ "epoch": 0.34726688102893893,
211
+ "grad_norm": 0.22008338570594788,
212
+ "learning_rate": 0.0001928114988519039,
213
+ "loss": 0.8181,
214
+ "step": 27
215
+ },
216
+ {
217
+ "epoch": 0.36012861736334406,
218
+ "grad_norm": 0.20176909863948822,
219
+ "learning_rate": 0.0001919527772551451,
220
+ "loss": 0.7487,
221
+ "step": 28
222
+ },
223
+ {
224
+ "epoch": 0.3729903536977492,
225
+ "grad_norm": 0.22451724112033844,
226
+ "learning_rate": 0.00019104775466588161,
227
+ "loss": 0.8272,
228
+ "step": 29
229
+ },
230
+ {
231
+ "epoch": 0.3858520900321543,
232
+ "grad_norm": 0.2383953183889389,
233
+ "learning_rate": 0.0001900968867902419,
234
+ "loss": 0.8406,
235
+ "step": 30
236
+ },
237
+ {
238
+ "epoch": 0.3987138263665595,
239
+ "grad_norm": 0.23490720987319946,
240
+ "learning_rate": 0.0001891006524188368,
241
+ "loss": 0.8325,
242
+ "step": 31
243
+ },
244
+ {
245
+ "epoch": 0.4115755627009646,
246
+ "grad_norm": 0.22093895077705383,
247
+ "learning_rate": 0.0001880595531856738,
248
+ "loss": 0.789,
249
+ "step": 32
250
+ },
251
+ {
252
+ "epoch": 0.42443729903536975,
253
+ "grad_norm": 0.2193409502506256,
254
+ "learning_rate": 0.00018697411331556956,
255
+ "loss": 0.7841,
256
+ "step": 33
257
+ },
258
+ {
259
+ "epoch": 0.43729903536977494,
260
+ "grad_norm": 0.2156234085559845,
261
+ "learning_rate": 0.00018584487936018661,
262
+ "loss": 0.7614,
263
+ "step": 34
264
+ },
265
+ {
266
+ "epoch": 0.45016077170418006,
267
+ "grad_norm": 0.20940746366977692,
268
+ "learning_rate": 0.00018467241992282843,
269
+ "loss": 0.7969,
270
+ "step": 35
271
+ },
272
+ {
273
+ "epoch": 0.4630225080385852,
274
+ "grad_norm": 0.21839483082294464,
275
+ "learning_rate": 0.00018345732537213027,
276
+ "loss": 0.7881,
277
+ "step": 36
278
+ },
279
+ {
280
+ "epoch": 0.4758842443729904,
281
+ "grad_norm": 0.2568162977695465,
282
+ "learning_rate": 0.00018220020754479102,
283
+ "loss": 0.9168,
284
+ "step": 37
285
+ },
286
+ {
287
+ "epoch": 0.4887459807073955,
288
+ "grad_norm": 0.21982234716415405,
289
+ "learning_rate": 0.00018090169943749476,
290
+ "loss": 0.8394,
291
+ "step": 38
292
+ },
293
+ {
294
+ "epoch": 0.5016077170418006,
295
+ "grad_norm": 0.22966712713241577,
296
+ "learning_rate": 0.00017956245488817812,
297
+ "loss": 0.8205,
298
+ "step": 39
299
+ },
300
+ {
301
+ "epoch": 0.5144694533762058,
302
+ "grad_norm": 0.22293612360954285,
303
+ "learning_rate": 0.000178183148246803,
304
+ "loss": 0.7895,
305
+ "step": 40
306
+ },
307
+ {
308
+ "epoch": 0.5144694533762058,
309
+ "eval_loss": 0.7905269861221313,
310
+ "eval_runtime": 7.4631,
311
+ "eval_samples_per_second": 21.975,
312
+ "eval_steps_per_second": 2.278,
313
+ "step": 40
314
  }
315
  ],
316
  "logging_steps": 1,
 
339
  "attributes": {}
340
  }
341
  },
342
+ "total_flos": 1.012328531755008e+16,
343
  "train_batch_size": 10,
344
  "trial_name": null,
345
  "trial_params": null