ribesstefano commited on
Commit
a9f55a4
·
verified ·
1 Parent(s): 6803e6a

Training in progress, step 60000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:80a80ba1df1c841ee670f71a637fc2ff1df9f6633cdbb91661d95be8ee1a5bd3
3
  size 409608164
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6a123f0f4c5a11cb72b99972714ce78828980a28225a86c5c52efa0dfd4317d
3
  size 409608164
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:48e16df1d1836b7794f6a8635aea83ef2eb38c93ca6cc7736092c223e74883b7
3
  size 814647162
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15ee547f4d1e2effb0117ee97acb3ccd85c719ade258b1abcb9b39428c1fc01e
3
  size 814647162
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1ef35073d449d7a3d7a21c34c4d41c4e76bf9563ca58e18dcc098f99e20f38b0
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3146ea4f58eeba70ef991619520fba1220f5a6969c7140f3e94b740993fab985
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a8da913cdb87fa0da659fbfc8e8d4a175dea1907bb6d7b316ea1a236f774834f
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8abde5ea6d7508710178712fe236e737431b4a593f0d18f03c2b4dfe63cb7155
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.5477160056657224,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_cosine-opt25/checkpoint-10000",
4
- "epoch": 5.425133162359439,
5
  "eval_steps": 2500,
6
- "global_step": 55000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -903,6 +903,76 @@
903
  "learning_rate": 2.1131851983700107e-05,
904
  "loss": 0.0004,
905
  "step": 55000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
906
  }
907
  ],
908
  "logging_steps": 500,
@@ -922,7 +992,7 @@
922
  "attributes": {}
923
  }
924
  },
925
- "total_flos": 4.7645419624832256e+17,
926
  "train_batch_size": 128,
927
  "trial_name": null,
928
  "trial_params": null
 
1
  {
2
  "best_metric": 0.5477160056657224,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_cosine-opt25/checkpoint-10000",
4
+ "epoch": 5.918327086210298,
5
  "eval_steps": 2500,
6
+ "global_step": 60000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
903
  "learning_rate": 2.1131851983700107e-05,
904
  "loss": 0.0004,
905
  "step": 55000
906
+ },
907
+ {
908
+ "epoch": 5.474452554744525,
909
+ "grad_norm": 0.004521891940385103,
910
+ "learning_rate": 2.074398781068747e-05,
911
+ "loss": 0.0004,
912
+ "step": 55500
913
+ },
914
+ {
915
+ "epoch": 5.523771947129611,
916
+ "grad_norm": 0.010427785106003284,
917
+ "learning_rate": 2.0357175848324307e-05,
918
+ "loss": 0.0004,
919
+ "step": 56000
920
+ },
921
+ {
922
+ "epoch": 5.573091339514697,
923
+ "grad_norm": 0.01812533475458622,
924
+ "learning_rate": 1.997228184710801e-05,
925
+ "loss": 0.0004,
926
+ "step": 56500
927
+ },
928
+ {
929
+ "epoch": 5.622410731899783,
930
+ "grad_norm": 0.01055191457271576,
931
+ "learning_rate": 1.958785833460854e-05,
932
+ "loss": 0.0004,
933
+ "step": 57000
934
+ },
935
+ {
936
+ "epoch": 5.671730124284869,
937
+ "grad_norm": 0.02802406996488571,
938
+ "learning_rate": 1.9204772861767664e-05,
939
+ "loss": 0.0004,
940
+ "step": 57500
941
+ },
942
+ {
943
+ "epoch": 5.721049516669955,
944
+ "grad_norm": 0.02212759293615818,
945
+ "learning_rate": 1.8823120138514845e-05,
946
+ "loss": 0.0004,
947
+ "step": 58000
948
+ },
949
+ {
950
+ "epoch": 5.7703689090550405,
951
+ "grad_norm": 0.020376645028591156,
952
+ "learning_rate": 1.8442994520561926e-05,
953
+ "loss": 0.0004,
954
+ "step": 58500
955
+ },
956
+ {
957
+ "epoch": 5.819688301440126,
958
+ "grad_norm": 0.004514740779995918,
959
+ "learning_rate": 1.8064489986075706e-05,
960
+ "loss": 0.0004,
961
+ "step": 59000
962
+ },
963
+ {
964
+ "epoch": 5.869007693825212,
965
+ "grad_norm": 0.014004266820847988,
966
+ "learning_rate": 1.768845191897044e-05,
967
+ "loss": 0.0004,
968
+ "step": 59500
969
+ },
970
+ {
971
+ "epoch": 5.918327086210298,
972
+ "grad_norm": 0.020170176401734352,
973
+ "learning_rate": 1.7313466151363385e-05,
974
+ "loss": 0.0004,
975
+ "step": 60000
976
  }
977
  ],
978
  "logging_steps": 500,
 
992
  "attributes": {}
993
  }
994
  },
995
+ "total_flos": 5.1976814397455616e+17,
996
  "train_batch_size": 128,
997
  "trial_name": null,
998
  "trial_params": null