ribesstefano commited on
Commit
83385ea
·
verified ·
1 Parent(s): e192670

Training in progress, step 65000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a0716e05b7be0802eaf3e025833847e8a3cae599b796d778d586c13f3d195761
3
  size 409608164
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5eb7417fd05987e94e77dc15f6dd9ec3188574d18a6f308eef6ee01d32c516aa
3
  size 409608164
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ac850fbcce728867ed8902fa223176fea14017c30527b0b8a62896ea10e17974
3
  size 814647162
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:536d9d4ac5a4e3b8f8430f128ced2edec56ca8ac219ebb21b6d3a5db6375c3ad
3
  size 814647162
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cc90a600027e1aba1caa11eec2a0f46a3006ea83630bb9dfb9aa44a091d77fcf
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b8012fbd0d7e4287f476308b3defe2cd698e9d90522ea0a0958949b79663329
3
  size 14244
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.5455028328611898,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_reduce-opt25-rand-smiles/checkpoint-10000",
4
- "epoch": 5.918327086210298,
5
  "eval_steps": 2500,
6
- "global_step": 60000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -973,6 +973,76 @@
973
  "learning_rate": 5e-05,
974
  "loss": 0.0007,
975
  "step": 60000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
976
  }
977
  ],
978
  "logging_steps": 500,
@@ -992,7 +1062,7 @@
992
  "attributes": {}
993
  }
994
  },
995
- "total_flos": 5.2375401437785344e+17,
996
  "train_batch_size": 128,
997
  "trial_name": null,
998
  "trial_params": null
 
1
  {
2
  "best_metric": 0.5455028328611898,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_reduce-opt25-rand-smiles/checkpoint-10000",
4
+ "epoch": 6.411521010061156,
5
  "eval_steps": 2500,
6
+ "global_step": 65000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
973
  "learning_rate": 5e-05,
974
  "loss": 0.0007,
975
  "step": 60000
976
+ },
977
+ {
978
+ "epoch": 5.967646478595384,
979
+ "grad_norm": 0.0135957645252347,
980
+ "learning_rate": 5e-05,
981
+ "loss": 0.0007,
982
+ "step": 60500
983
+ },
984
+ {
985
+ "epoch": 6.01696587098047,
986
+ "grad_norm": 0.016737064346671104,
987
+ "learning_rate": 5e-05,
988
+ "loss": 0.0007,
989
+ "step": 61000
990
+ },
991
+ {
992
+ "epoch": 6.066285263365556,
993
+ "grad_norm": 0.0470641553401947,
994
+ "learning_rate": 5e-05,
995
+ "loss": 0.0008,
996
+ "step": 61500
997
+ },
998
+ {
999
+ "epoch": 6.115604655750641,
1000
+ "grad_norm": 0.037444982677698135,
1001
+ "learning_rate": 5e-05,
1002
+ "loss": 0.0009,
1003
+ "step": 62000
1004
+ },
1005
+ {
1006
+ "epoch": 6.164924048135727,
1007
+ "grad_norm": 0.03800535574555397,
1008
+ "learning_rate": 5e-05,
1009
+ "loss": 0.0006,
1010
+ "step": 62500
1011
+ },
1012
+ {
1013
+ "epoch": 6.2142434405208125,
1014
+ "grad_norm": 0.02700197882950306,
1015
+ "learning_rate": 5e-05,
1016
+ "loss": 0.0006,
1017
+ "step": 63000
1018
+ },
1019
+ {
1020
+ "epoch": 6.263562832905898,
1021
+ "grad_norm": 0.03783607482910156,
1022
+ "learning_rate": 5e-05,
1023
+ "loss": 0.0008,
1024
+ "step": 63500
1025
+ },
1026
+ {
1027
+ "epoch": 6.312882225290984,
1028
+ "grad_norm": 0.03808404505252838,
1029
+ "learning_rate": 5e-05,
1030
+ "loss": 0.0011,
1031
+ "step": 64000
1032
+ },
1033
+ {
1034
+ "epoch": 6.36220161767607,
1035
+ "grad_norm": 0.016872331500053406,
1036
+ "learning_rate": 5e-05,
1037
+ "loss": 0.0011,
1038
+ "step": 64500
1039
+ },
1040
+ {
1041
+ "epoch": 6.411521010061156,
1042
+ "grad_norm": 0.032465968281030655,
1043
+ "learning_rate": 5e-05,
1044
+ "loss": 0.0005,
1045
+ "step": 65000
1046
  }
1047
  ],
1048
  "logging_steps": 500,
 
1062
  "attributes": {}
1063
  }
1064
  },
1065
+ "total_flos": 5.674161087565732e+17,
1066
  "train_batch_size": 128,
1067
  "trial_name": null,
1068
  "trial_params": null