ribesstefano commited on
Commit
a978981
·
verified ·
1 Parent(s): 3dfd790

Training in progress, step 40000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:680cdd8ce09ed81d5904b52f855191e264cb3a8777e419668f407b84fc55713a
3
  size 409608164
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:503ce751da80aa8b1f118a861d5757c844ce8de1db073fef17acad2029b59e2b
3
  size 409608164
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fe583685246694473b81b6159749d52ae9755ca3bd5d94c328cf706a1d1198e0
3
  size 814647162
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6fbaedfb0a4123dcead5a29fb791a66763b1c12ca90eb39e7c861a4cb61e314d
3
  size 814647162
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:86e4d6f5632dbc00793287b2f22a557709a70a15c41fc36f8e72a42acf0f499d
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:39417990b3af4b754d49ef8161883cbdefa4e1231e6e279b71d13d553fe650cf
3
  size 14244
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.5455028328611898,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_reduce-opt25-rand-smiles/checkpoint-10000",
4
- "epoch": 3.4523574669560073,
5
  "eval_steps": 2500,
6
- "global_step": 35000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -623,6 +623,76 @@
623
  "learning_rate": 5e-05,
624
  "loss": 0.0014,
625
  "step": 35000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
626
  }
627
  ],
628
  "logging_steps": 500,
@@ -642,7 +712,7 @@
642
  "attributes": {}
643
  }
644
  },
645
- "total_flos": 3.055328543636475e+17,
646
  "train_batch_size": 128,
647
  "trial_name": null,
648
  "trial_params": null
 
1
  {
2
  "best_metric": 0.5455028328611898,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_reduce-opt25-rand-smiles/checkpoint-10000",
4
+ "epoch": 3.945551390806865,
5
  "eval_steps": 2500,
6
+ "global_step": 40000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
623
  "learning_rate": 5e-05,
624
  "loss": 0.0014,
625
  "step": 35000
626
+ },
627
+ {
628
+ "epoch": 3.501676859341093,
629
+ "grad_norm": 0.021409129723906517,
630
+ "learning_rate": 5e-05,
631
+ "loss": 0.0014,
632
+ "step": 35500
633
+ },
634
+ {
635
+ "epoch": 3.5509962517261786,
636
+ "grad_norm": 0.01704758033156395,
637
+ "learning_rate": 5e-05,
638
+ "loss": 0.001,
639
+ "step": 36000
640
+ },
641
+ {
642
+ "epoch": 3.6003156441112645,
643
+ "grad_norm": 0.03872734308242798,
644
+ "learning_rate": 5e-05,
645
+ "loss": 0.001,
646
+ "step": 36500
647
+ },
648
+ {
649
+ "epoch": 3.6496350364963503,
650
+ "grad_norm": 0.030823158100247383,
651
+ "learning_rate": 5e-05,
652
+ "loss": 0.0012,
653
+ "step": 37000
654
+ },
655
+ {
656
+ "epoch": 3.698954428881436,
657
+ "grad_norm": 0.03794977068901062,
658
+ "learning_rate": 5e-05,
659
+ "loss": 0.001,
660
+ "step": 37500
661
+ },
662
+ {
663
+ "epoch": 3.748273821266522,
664
+ "grad_norm": 0.042292773723602295,
665
+ "learning_rate": 5e-05,
666
+ "loss": 0.0013,
667
+ "step": 38000
668
+ },
669
+ {
670
+ "epoch": 3.797593213651608,
671
+ "grad_norm": 0.040720898658037186,
672
+ "learning_rate": 5e-05,
673
+ "loss": 0.0015,
674
+ "step": 38500
675
+ },
676
+ {
677
+ "epoch": 3.8469126060366934,
678
+ "grad_norm": 0.017985910177230835,
679
+ "learning_rate": 5e-05,
680
+ "loss": 0.001,
681
+ "step": 39000
682
+ },
683
+ {
684
+ "epoch": 3.8962319984217793,
685
+ "grad_norm": 0.04241044074296951,
686
+ "learning_rate": 5e-05,
687
+ "loss": 0.0009,
688
+ "step": 39500
689
+ },
690
+ {
691
+ "epoch": 3.945551390806865,
692
+ "grad_norm": 0.02607624977827072,
693
+ "learning_rate": 5e-05,
694
+ "loss": 0.001,
695
+ "step": 40000
696
  }
697
  ],
698
  "logging_steps": 500,
 
712
  "attributes": {}
713
  }
714
  },
715
+ "total_flos": 3.4915036992864614e+17,
716
  "train_batch_size": 128,
717
  "trial_name": null,
718
  "trial_params": null