ribesstefano commited on
Commit
193837b
·
verified ·
1 Parent(s): 212964d

Training in progress, step 50000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b0c6cca2c6a9ac2adb2075297cbca19d4215769af04fc3b72aa6f7ffdaf1fbc7
3
  size 409608164
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:617b674267e2191b4addb7b6174d2a7b3b6be5e15b623c52cb202c65d1300fbc
3
  size 409608164
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ef0ddd76d8b4525cbbad4aefefb74d4c462f82436fa56c207a7d0c37484c235e
3
  size 814647162
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd83813d9dbd25af61d05bbd32798b812892a658328f22ee06646b799fb7b363
3
  size 814647162
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ec0ea160082cc39ac30942ebcddc73a0bcdea69205d0069eef8edae7fc1a9d72
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:11d87c5b913e7a5b1cabf70de94edcbb67847e2292b7550ae53ff3dda1196c2a
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dc36b8ae22f6cb108858cc00165c8e8672f7e966fabf0d057ddaa95d9238a73a
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:99fcfbd1683ab8df6f870c07c1a59edbf30ebbec5a46d6d0341019ab14692cb0
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.5462110481586402,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_cosine_restarts-opt25/checkpoint-10000",
4
- "epoch": 4.438745314657724,
5
  "eval_steps": 2500,
6
- "global_step": 45000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -763,6 +763,76 @@
763
  "learning_rate": 3.534411131835637e-05,
764
  "loss": 0.001,
765
  "step": 45000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
766
  }
767
  ],
768
  "logging_steps": 500,
@@ -782,7 +852,7 @@
782
  "attributes": {}
783
  }
784
  },
785
- "total_flos": 3.898212104814858e+17,
786
  "train_batch_size": 128,
787
  "trial_name": null,
788
  "trial_params": null
 
1
  {
2
  "best_metric": 0.5462110481586402,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_cosine_restarts-opt25/checkpoint-10000",
4
+ "epoch": 4.931939238508582,
5
  "eval_steps": 2500,
6
+ "global_step": 50000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
763
  "learning_rate": 3.534411131835637e-05,
764
  "loss": 0.001,
765
  "step": 45000
766
+ },
767
+ {
768
+ "epoch": 4.48806470704281,
769
+ "grad_norm": 0.019216470420360565,
770
+ "learning_rate": 2.3789123953690796e-05,
771
+ "loss": 0.0007,
772
+ "step": 45500
773
+ },
774
+ {
775
+ "epoch": 4.537384099427895,
776
+ "grad_norm": 0.02267817035317421,
777
+ "learning_rate": 1.2502285603117597e-05,
778
+ "loss": 0.0005,
779
+ "step": 46000
780
+ },
781
+ {
782
+ "epoch": 4.5867034918129805,
783
+ "grad_norm": 0.01543174497783184,
784
+ "learning_rate": 3.983071407096459e-06,
785
+ "loss": 0.0005,
786
+ "step": 46500
787
+ },
788
+ {
789
+ "epoch": 4.636022884198066,
790
+ "grad_norm": 0.020069127902388573,
791
+ "learning_rate": 1.180649630719205e-07,
792
+ "loss": 0.0004,
793
+ "step": 47000
794
+ },
795
+ {
796
+ "epoch": 4.685342276583152,
797
+ "grad_norm": 0.06315533071756363,
798
+ "learning_rate": 4.8236826412442815e-05,
799
+ "loss": 0.0016,
800
+ "step": 47500
801
+ },
802
+ {
803
+ "epoch": 4.734661668968238,
804
+ "grad_norm": 0.05815032124519348,
805
+ "learning_rate": 4.1445912722173056e-05,
806
+ "loss": 0.0011,
807
+ "step": 48000
808
+ },
809
+ {
810
+ "epoch": 4.783981061353324,
811
+ "grad_norm": 0.03868298977613449,
812
+ "learning_rate": 3.103609665799435e-05,
813
+ "loss": 0.0009,
814
+ "step": 48500
815
+ },
816
+ {
817
+ "epoch": 4.83330045373841,
818
+ "grad_norm": 0.026825400069355965,
819
+ "learning_rate": 1.9271702808963772e-05,
820
+ "loss": 0.0006,
821
+ "step": 49000
822
+ },
823
+ {
824
+ "epoch": 4.882619846123496,
825
+ "grad_norm": 0.022513261064887047,
826
+ "learning_rate": 8.775842802754727e-06,
827
+ "loss": 0.0005,
828
+ "step": 49500
829
+ },
830
+ {
831
+ "epoch": 4.931939238508582,
832
+ "grad_norm": 0.04292171820998192,
833
+ "learning_rate": 1.8728292899447775e-06,
834
+ "loss": 0.0004,
835
+ "step": 50000
836
  }
837
  ],
838
  "logging_steps": 500,
 
852
  "attributes": {}
853
  }
854
  },
855
+ "total_flos": 4.3312528608288154e+17,
856
  "train_batch_size": 128,
857
  "trial_name": null,
858
  "trial_params": null