ribesstefano commited on
Commit
3360b5a
·
verified ·
1 Parent(s): 7e74559

Training in progress, step 55000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bbd952fb7ca0f8ddc9b0e7256bf820a3fc04551b8ae426ac98c3094dee2ea254
3
  size 409608164
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:80a80ba1df1c841ee670f71a637fc2ff1df9f6633cdbb91661d95be8ee1a5bd3
3
  size 409608164
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0d7970449ed8a17f9c6c4492b954ad4964b40feccf1db02f9503f0b2b341dee5
3
  size 814647162
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:48e16df1d1836b7794f6a8635aea83ef2eb38c93ca6cc7736092c223e74883b7
3
  size 814647162
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:44f6a18681a37c28eaea9106736980f3cb0b876fb22eb09649d7485450746688
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ef35073d449d7a3d7a21c34c4d41c4e76bf9563ca58e18dcc098f99e20f38b0
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:129967ad4aacd30894639b032d8517110a4963d1cc93db81c93bfda673142ba8
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a8da913cdb87fa0da659fbfc8e8d4a175dea1907bb6d7b316ea1a236f774834f
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.5477160056657224,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_cosine-opt25/checkpoint-10000",
4
- "epoch": 4.931939238508582,
5
  "eval_steps": 2500,
6
- "global_step": 50000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -833,6 +833,76 @@
833
  "learning_rate": 2.504638484990372e-05,
834
  "loss": 0.0004,
835
  "step": 50000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
836
  }
837
  ],
838
  "logging_steps": 500,
@@ -852,7 +922,7 @@
852
  "attributes": {}
853
  }
854
  },
855
- "total_flos": 4.3312528608288154e+17,
856
  "train_batch_size": 128,
857
  "trial_name": null,
858
  "trial_params": null
 
1
  {
2
  "best_metric": 0.5477160056657224,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_cosine-opt25/checkpoint-10000",
4
+ "epoch": 5.425133162359439,
5
  "eval_steps": 2500,
6
+ "global_step": 55000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
833
  "learning_rate": 2.504638484990372e-05,
834
  "loss": 0.0004,
835
  "step": 50000
836
+ },
837
+ {
838
+ "epoch": 4.9812586308936675,
839
+ "grad_norm": 0.04032417759299278,
840
+ "learning_rate": 2.4653303816212188e-05,
841
+ "loss": 0.0005,
842
+ "step": 50500
843
+ },
844
+ {
845
+ "epoch": 5.030578023278753,
846
+ "grad_norm": 0.008274455554783344,
847
+ "learning_rate": 2.42603084959539e-05,
848
+ "loss": 0.0005,
849
+ "step": 51000
850
+ },
851
+ {
852
+ "epoch": 5.079897415663839,
853
+ "grad_norm": 0.028657300397753716,
854
+ "learning_rate": 2.386749604906234e-05,
855
+ "loss": 0.0004,
856
+ "step": 51500
857
+ },
858
+ {
859
+ "epoch": 5.129216808048925,
860
+ "grad_norm": 0.006416209042072296,
861
+ "learning_rate": 2.3474963590259334e-05,
862
+ "loss": 0.0004,
863
+ "step": 52000
864
+ },
865
+ {
866
+ "epoch": 5.178536200434011,
867
+ "grad_norm": 0.014230284839868546,
868
+ "learning_rate": 2.3082808165045468e-05,
869
+ "loss": 0.0004,
870
+ "step": 52500
871
+ },
872
+ {
873
+ "epoch": 5.227855592819097,
874
+ "grad_norm": 0.014515167102217674,
875
+ "learning_rate": 2.2691126725707605e-05,
876
+ "loss": 0.0004,
877
+ "step": 53000
878
+ },
879
+ {
880
+ "epoch": 5.277174985204182,
881
+ "grad_norm": 0.0125635527074337,
882
+ "learning_rate": 2.2300797694611534e-05,
883
+ "loss": 0.0004,
884
+ "step": 53500
885
+ },
886
+ {
887
+ "epoch": 5.326494377589268,
888
+ "grad_norm": 0.022368459030985832,
889
+ "learning_rate": 2.191035315978832e-05,
890
+ "loss": 0.0004,
891
+ "step": 54000
892
+ },
893
+ {
894
+ "epoch": 5.375813769974354,
895
+ "grad_norm": 0.007136535365134478,
896
+ "learning_rate": 2.1520672475998373e-05,
897
+ "loss": 0.0004,
898
+ "step": 54500
899
+ },
900
+ {
901
+ "epoch": 5.425133162359439,
902
+ "grad_norm": 0.00662201177328825,
903
+ "learning_rate": 2.1131851983700107e-05,
904
+ "loss": 0.0004,
905
+ "step": 55000
906
  }
907
  ],
908
  "logging_steps": 500,
 
922
  "attributes": {}
923
  }
924
  },
925
+ "total_flos": 4.7645419624832256e+17,
926
  "train_batch_size": 128,
927
  "trial_name": null,
928
  "trial_params": null