alessandronascimento commited on
Commit
1a74a9e
·
verified ·
1 Parent(s): d2d1c2d

Training in progress, epoch 4, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7fcef207828e97d334020207b7c4e0ce8b10911f91e884313cc27540bcf22215
3
  size 1713050034
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:790e59fe9066d0dd6221cfe2764f4275cd9c904136ac1745ccc0f3eb48a5ed1b
3
  size 1713050034
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a3ed2b4efa016d58457f3d8931b71a9ebaaae4412d1d353411975961be30dbac
3
  size 816721594
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df630dcd81b9b194bc0ac1be2a6a9fef794e8226952b0be6694b2f40bb7e1a82
3
  size 816721594
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6ba1195da48c8391083911b2a9cc3e15ab8d82f8f8936d669e0116eb962ce738
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8693c95f0f09ea2f300c00a9d97d4ecc249e7fd41ef6cc52cb95a982ef985bca
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:edebf099edd988740678361da4ee055b665f3729f5f42b95a748d6bab1b73604
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:414d81b0296706ee8544c28eaf6013748a0dfc834c3d7fbd1015bc57e89aeb34
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "best_metric": 6.0006157582392916e-05,
3
- "best_model_checkpoint": "ProtChem_ESM2_MolGen_Decoder/checkpoint-15889",
4
- "epoch": 4.0,
5
  "eval_steps": 500,
6
- "global_step": 15889,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -256,6 +256,70 @@
256
  "eval_samples_per_second": 15.238,
257
  "eval_steps_per_second": 0.952,
258
  "step": 15889
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
259
  }
260
  ],
261
  "logging_steps": 500,
@@ -284,7 +348,7 @@
284
  "attributes": {}
285
  }
286
  },
287
- "total_flos": 1.0669336571309322e+19,
288
  "train_batch_size": 32,
289
  "trial_name": null,
290
  "trial_params": null
 
1
  {
2
+ "best_metric": 5.9128127759322524e-05,
3
+ "best_model_checkpoint": "ProtChem_ESM2_MolGen_Decoder/checkpoint-19861",
4
+ "epoch": 4.999937063377179,
5
  "eval_steps": 500,
6
+ "global_step": 19861,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
256
  "eval_samples_per_second": 15.238,
257
  "eval_steps_per_second": 0.952,
258
  "step": 15889
259
+ },
260
+ {
261
+ "epoch": 4.027943860532444,
262
+ "grad_norm": 0.007171630859375,
263
+ "learning_rate": 1.393812617541025e-05,
264
+ "loss": 0.0001,
265
+ "step": 16000
266
+ },
267
+ {
268
+ "epoch": 4.153817106174083,
269
+ "grad_norm": 0.001190185546875,
270
+ "learning_rate": 1.3552038517004991e-05,
271
+ "loss": 0.0001,
272
+ "step": 16500
273
+ },
274
+ {
275
+ "epoch": 4.279690351815722,
276
+ "grad_norm": 0.003082275390625,
277
+ "learning_rate": 1.3159791843212542e-05,
278
+ "loss": 0.0001,
279
+ "step": 17000
280
+ },
281
+ {
282
+ "epoch": 4.405563597457361,
283
+ "grad_norm": 0.002593994140625,
284
+ "learning_rate": 1.2762066285537071e-05,
285
+ "loss": 0.0001,
286
+ "step": 17500
287
+ },
288
+ {
289
+ "epoch": 4.531436843099,
290
+ "grad_norm": 0.01129150390625,
291
+ "learning_rate": 1.2359551475529e-05,
292
+ "loss": 0.0001,
293
+ "step": 18000
294
+ },
295
+ {
296
+ "epoch": 4.657310088740639,
297
+ "grad_norm": 0.002166748046875,
298
+ "learning_rate": 1.1952945349006455e-05,
299
+ "loss": 0.0001,
300
+ "step": 18500
301
+ },
302
+ {
303
+ "epoch": 4.783183334382277,
304
+ "grad_norm": 0.006622314453125,
305
+ "learning_rate": 1.1542952935877703e-05,
306
+ "loss": 0.0001,
307
+ "step": 19000
308
+ },
309
+ {
310
+ "epoch": 4.909056580023916,
311
+ "grad_norm": 0.0029449462890625,
312
+ "learning_rate": 1.1130285137662811e-05,
313
+ "loss": 0.0001,
314
+ "step": 19500
315
+ },
316
+ {
317
+ "epoch": 4.999937063377179,
318
+ "eval_loss": 5.9128127759322524e-05,
319
+ "eval_runtime": 16790.9197,
320
+ "eval_samples_per_second": 15.24,
321
+ "eval_steps_per_second": 0.952,
322
+ "step": 19861
323
  }
324
  ],
325
  "logging_steps": 500,
 
348
  "attributes": {}
349
  }
350
  },
351
+ "total_flos": 1.3336670714136654e+19,
352
  "train_batch_size": 32,
353
  "trial_name": null,
354
  "trial_params": null