ribesstefano commited on
Commit
fe0f2a6
·
verified ·
1 Parent(s): a1072ce

Training in progress, step 30000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bf584d7532aca183ba3375c7e2b4cabba44bb42e2ada64266474944d71e8bc4b
3
  size 409608164
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:847704143a8fe1bda65db1fe2c65f7a5566407d3fa081a0f600bf66f16b8b3a0
3
  size 409608164
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2a5c2142f0e58a50a3708ce303ed8fdbdc0637809e5a24cca846334043ca8b38
3
  size 814647162
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c5957f98f8279a93ad8e35cd4746546f5160d187ed7070c5aa9bbdb47408fe1c
3
  size 814647162
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2d57b37b13c2760795173d6cc56195de8db4b64e0ea9a20349975e5a8a5a1457
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c65c2745f22d1c4db409b96f64b6f71c3a2ee8bc2f8c7ff26a84d73ab1a44cb8
3
  size 14244
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.5455028328611898,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_reduce-opt25-rand-smiles/checkpoint-10000",
4
- "epoch": 2.465969619254291,
5
  "eval_steps": 2500,
6
- "global_step": 25000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -483,6 +483,76 @@
483
  "learning_rate": 5e-05,
484
  "loss": 0.0015,
485
  "step": 25000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
486
  }
487
  ],
488
  "logging_steps": 500,
@@ -502,7 +572,7 @@
502
  "attributes": {}
503
  }
504
  },
505
- "total_flos": 2.1826558457597645e+17,
506
  "train_batch_size": 128,
507
  "trial_name": null,
508
  "trial_params": null
 
1
  {
2
  "best_metric": 0.5455028328611898,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_reduce-opt25-rand-smiles/checkpoint-10000",
4
+ "epoch": 2.959163543105149,
5
  "eval_steps": 2500,
6
+ "global_step": 30000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
483
  "learning_rate": 5e-05,
484
  "loss": 0.0015,
485
  "step": 25000
486
+ },
487
+ {
488
+ "epoch": 2.5152890116393767,
489
+ "grad_norm": 0.045403581112623215,
490
+ "learning_rate": 5e-05,
491
+ "loss": 0.0015,
492
+ "step": 25500
493
+ },
494
+ {
495
+ "epoch": 2.5646084040244626,
496
+ "grad_norm": 0.059790823608636856,
497
+ "learning_rate": 5e-05,
498
+ "loss": 0.0018,
499
+ "step": 26000
500
+ },
501
+ {
502
+ "epoch": 2.6139277964095484,
503
+ "grad_norm": 0.05007553473114967,
504
+ "learning_rate": 5e-05,
505
+ "loss": 0.0015,
506
+ "step": 26500
507
+ },
508
+ {
509
+ "epoch": 2.6632471887946343,
510
+ "grad_norm": 0.03945387527346611,
511
+ "learning_rate": 5e-05,
512
+ "loss": 0.0014,
513
+ "step": 27000
514
+ },
515
+ {
516
+ "epoch": 2.7125665811797197,
517
+ "grad_norm": 0.033401645720005035,
518
+ "learning_rate": 5e-05,
519
+ "loss": 0.0013,
520
+ "step": 27500
521
+ },
522
+ {
523
+ "epoch": 2.7618859735648056,
524
+ "grad_norm": 0.05006084218621254,
525
+ "learning_rate": 5e-05,
526
+ "loss": 0.0013,
527
+ "step": 28000
528
+ },
529
+ {
530
+ "epoch": 2.8112053659498915,
531
+ "grad_norm": 0.04564574733376503,
532
+ "learning_rate": 5e-05,
533
+ "loss": 0.0014,
534
+ "step": 28500
535
+ },
536
+ {
537
+ "epoch": 2.8605247583349773,
538
+ "grad_norm": 0.0373481847345829,
539
+ "learning_rate": 5e-05,
540
+ "loss": 0.0013,
541
+ "step": 29000
542
+ },
543
+ {
544
+ "epoch": 2.909844150720063,
545
+ "grad_norm": 0.030000342056155205,
546
+ "learning_rate": 5e-05,
547
+ "loss": 0.0019,
548
+ "step": 29500
549
+ },
550
+ {
551
+ "epoch": 2.959163543105149,
552
+ "grad_norm": 0.012932154349982738,
553
+ "learning_rate": 5e-05,
554
+ "loss": 0.0021,
555
+ "step": 30000
556
  }
557
  ],
558
  "logging_steps": 500,
 
572
  "attributes": {}
573
  }
574
  },
575
+ "total_flos": 2.619102484842793e+17,
576
  "train_batch_size": 128,
577
  "trial_name": null,
578
  "trial_params": null