ribesstefano commited on
Commit
a702bc3
·
verified ·
1 Parent(s): 5aba80b

Training in progress, step 55000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:617b674267e2191b4addb7b6174d2a7b3b6be5e15b623c52cb202c65d1300fbc
3
  size 409608164
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7492c7bf922a3eb50cfa4081fec5e81f8de3ef95639132c619339ba07befabc1
3
  size 409608164
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fd83813d9dbd25af61d05bbd32798b812892a658328f22ee06646b799fb7b363
3
  size 814647162
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bfd83dccf1b07606afabe5fdded3e70c80b15a1881b1ff52715b129c8766202f
3
  size 814647162
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:11d87c5b913e7a5b1cabf70de94edcbb67847e2292b7550ae53ff3dda1196c2a
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2915ca4a6fbfe4a8c8b438dcf3502a8e4e6cc49eb1760c5072daaa5ca3a37cb
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:99fcfbd1683ab8df6f870c07c1a59edbf30ebbec5a46d6d0341019ab14692cb0
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:409aa698ded6143db955deb4e5b7664a1227e243cdfc2f5415f3c7211cca6584
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.5462110481586402,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_cosine_restarts-opt25/checkpoint-10000",
4
- "epoch": 4.931939238508582,
5
  "eval_steps": 2500,
6
- "global_step": 50000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -833,6 +833,76 @@
833
  "learning_rate": 1.8728292899447775e-06,
834
  "loss": 0.0004,
835
  "step": 50000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
836
  }
837
  ],
838
  "logging_steps": 500,
@@ -852,7 +922,7 @@
852
  "attributes": {}
853
  }
854
  },
855
- "total_flos": 4.3312528608288154e+17,
856
  "train_batch_size": 128,
857
  "trial_name": null,
858
  "trial_params": null
 
1
  {
2
  "best_metric": 0.5462110481586402,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_cosine_restarts-opt25/checkpoint-10000",
4
+ "epoch": 5.425133162359439,
5
  "eval_steps": 2500,
6
+ "global_step": 55000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
833
  "learning_rate": 1.8728292899447775e-06,
834
  "loss": 0.0004,
835
  "step": 50000
836
+ },
837
+ {
838
+ "epoch": 4.9812586308936675,
839
+ "grad_norm": 0.09393975138664246,
840
+ "learning_rate": 4.9908662458888896e-05,
841
+ "loss": 0.0005,
842
+ "step": 50500
843
+ },
844
+ {
845
+ "epoch": 5.030578023278753,
846
+ "grad_norm": 0.019034981727600098,
847
+ "learning_rate": 4.61993375092424e-05,
848
+ "loss": 0.0019,
849
+ "step": 51000
850
+ },
851
+ {
852
+ "epoch": 5.079897415663839,
853
+ "grad_norm": 0.008647086098790169,
854
+ "learning_rate": 3.7791398273089494e-05,
855
+ "loss": 0.0009,
856
+ "step": 51500
857
+ },
858
+ {
859
+ "epoch": 5.129216808048925,
860
+ "grad_norm": 0.013575778342783451,
861
+ "learning_rate": 2.655079846342193e-05,
862
+ "loss": 0.0006,
863
+ "step": 52000
864
+ },
865
+ {
866
+ "epoch": 5.178536200434011,
867
+ "grad_norm": 0.009713593870401382,
868
+ "learning_rate": 1.496677367620188e-05,
869
+ "loss": 0.0005,
870
+ "step": 52500
871
+ },
872
+ {
873
+ "epoch": 5.227855592819097,
874
+ "grad_norm": 0.024831971153616905,
875
+ "learning_rate": 5.604611114346173e-06,
876
+ "loss": 0.0004,
877
+ "step": 53000
878
+ },
879
+ {
880
+ "epoch": 5.277174985204182,
881
+ "grad_norm": 0.005304230377078056,
882
+ "learning_rate": 5.37565648481686e-07,
883
+ "loss": 0.0004,
884
+ "step": 53500
885
+ },
886
+ {
887
+ "epoch": 5.326494377589268,
888
+ "grad_norm": 0.055664077401161194,
889
+ "learning_rate": 4.911226335360363e-05,
890
+ "loss": 0.0012,
891
+ "step": 54000
892
+ },
893
+ {
894
+ "epoch": 5.375813769974354,
895
+ "grad_norm": 0.050345055758953094,
896
+ "learning_rate": 4.342242140793571e-05,
897
+ "loss": 0.0011,
898
+ "step": 54500
899
+ },
900
+ {
901
+ "epoch": 5.425133162359439,
902
+ "grad_norm": 0.015438569709658623,
903
+ "learning_rate": 3.3652926426937384e-05,
904
+ "loss": 0.0008,
905
+ "step": 55000
906
  }
907
  ],
908
  "logging_steps": 500,
 
922
  "attributes": {}
923
  }
924
  },
925
+ "total_flos": 4.7645419624832256e+17,
926
  "train_batch_size": 128,
927
  "trial_name": null,
928
  "trial_params": null