ribesstefano commited on
Commit
95c313f
·
verified ·
1 Parent(s): 163bb44

Training in progress, step 65000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:27d41ba63ba593f2cfce363802b440927332dce61aafea879db261ae77cd52d0
3
  size 409608164
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b765806c32a71debbe5fcd75a03b9d683e507d65faff94e01c59c7581ed7ac5f
3
  size 409608164
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d742eeb8d925c1a166af3e09a3e0fd4ebadea870bd4919c2660b4864e7654fcd
3
  size 814647162
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:973086399ebf7dfee4ca3fae615f7dbbcc6764ef896837b42c878ef225b08ff0
3
  size 814647162
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ac455c3f391185463b9c69f4fb71b00665f813b40d3fca5439822d6487ecd3b5
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d94c3d1ced5d6eccb982de07377a5bac97c708bdc2a2e2ec8462b5f4c3224ec7
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8abde5ea6d7508710178712fe236e737431b4a593f0d18f03c2b4dfe63cb7155
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ef1f0d7258e92b9f9871d81d52ebcc946c37081ef596ae367fa8c041eca7a05
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.5389518413597734,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_cosine-opt25-rand-smiles/checkpoint-10000",
4
- "epoch": 5.918327086210298,
5
  "eval_steps": 2500,
6
- "global_step": 60000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -973,6 +973,76 @@
973
  "learning_rate": 1.7313466151363385e-05,
974
  "loss": 0.0004,
975
  "step": 60000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
976
  }
977
  ],
978
  "logging_steps": 500,
@@ -992,7 +1062,7 @@
992
  "attributes": {}
993
  }
994
  },
995
- "total_flos": 5.2375401437785344e+17,
996
  "train_batch_size": 128,
997
  "trial_name": null,
998
  "trial_params": null
 
1
  {
2
  "best_metric": 0.5389518413597734,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_cosine-opt25-rand-smiles/checkpoint-10000",
4
+ "epoch": 6.411521010061156,
5
  "eval_steps": 2500,
6
+ "global_step": 65000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
973
  "learning_rate": 1.7313466151363385e-05,
974
  "loss": 0.0004,
975
  "step": 60000
976
+ },
977
+ {
978
+ "epoch": 5.967646478595384,
979
+ "grad_norm": 0.011568223126232624,
980
+ "learning_rate": 1.6941124932601034e-05,
981
+ "loss": 0.0005,
982
+ "step": 60500
983
+ },
984
+ {
985
+ "epoch": 6.01696587098047,
986
+ "grad_norm": 0.04736906662583351,
987
+ "learning_rate": 1.6570027997553185e-05,
988
+ "loss": 0.0004,
989
+ "step": 61000
990
+ },
991
+ {
992
+ "epoch": 6.066285263365556,
993
+ "grad_norm": 0.01865927129983902,
994
+ "learning_rate": 1.6201015198057866e-05,
995
+ "loss": 0.0004,
996
+ "step": 61500
997
+ },
998
+ {
999
+ "epoch": 6.115604655750641,
1000
+ "grad_norm": 0.014169979840517044,
1001
+ "learning_rate": 1.5834177764868226e-05,
1002
+ "loss": 0.0004,
1003
+ "step": 62000
1004
+ },
1005
+ {
1006
+ "epoch": 6.164924048135727,
1007
+ "grad_norm": 0.009675228968262672,
1008
+ "learning_rate": 1.5469606390923247e-05,
1009
+ "loss": 0.0004,
1010
+ "step": 62500
1011
+ },
1012
+ {
1013
+ "epoch": 6.2142434405208125,
1014
+ "grad_norm": 0.00304154260084033,
1015
+ "learning_rate": 1.5107391208925827e-05,
1016
+ "loss": 0.0004,
1017
+ "step": 63000
1018
+ },
1019
+ {
1020
+ "epoch": 6.263562832905898,
1021
+ "grad_norm": 0.004374227486550808,
1022
+ "learning_rate": 1.474762176905931e-05,
1023
+ "loss": 0.0004,
1024
+ "step": 63500
1025
+ },
1026
+ {
1027
+ "epoch": 6.312882225290984,
1028
+ "grad_norm": 0.006118281278759241,
1029
+ "learning_rate": 1.439038701684809e-05,
1030
+ "loss": 0.0004,
1031
+ "step": 64000
1032
+ },
1033
+ {
1034
+ "epoch": 6.36220161767607,
1035
+ "grad_norm": 0.006160255055874586,
1036
+ "learning_rate": 1.4035775271167664e-05,
1037
+ "loss": 0.0004,
1038
+ "step": 64500
1039
+ },
1040
+ {
1041
+ "epoch": 6.411521010061156,
1042
+ "grad_norm": 0.009528923779726028,
1043
+ "learning_rate": 1.3684575241351546e-05,
1044
+ "loss": 0.0004,
1045
+ "step": 65000
1046
  }
1047
  ],
1048
  "logging_steps": 500,
 
1062
  "attributes": {}
1063
  }
1064
  },
1065
+ "total_flos": 5.674161087565732e+17,
1066
  "train_batch_size": 128,
1067
  "trial_name": null,
1068
  "trial_params": null