ribesstefano commited on
Commit
379db46
·
verified ·
1 Parent(s): 875ff93

Training in progress, step 50000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:99d99ef005366c768182dbd5fbbed197e6391e6fc6d0c746b5feb7b2fd2a2764
3
  size 409608164
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe75c7d1d2a9fb1383fa16ca04fa9582f79bce9c2c6b619ccc5bf8f41a7349d3
3
  size 409608164
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8f6b17853dc9aaa706ba3ed378aef704b5d5535f20af1ae291d3e1c2ee5cb5bc
3
  size 814647162
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c16c4ea2aeaf96ed945f3ff8d93677b434f48d1a77151570a2c01992d79bde54
3
  size 814647162
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:76ef189c202797a47ce62889d21e7c97c6a4ce5ace6abf783017722de7ad98b2
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1cdbc5184733c39b403d000033558f016107ac570eaa5b3b5be67cc987ed4902
3
  size 14244
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.5455028328611898,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_reduce-opt25-rand-smiles/checkpoint-10000",
4
- "epoch": 4.438745314657724,
5
  "eval_steps": 2500,
6
- "global_step": 45000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -763,6 +763,76 @@
763
  "learning_rate": 5e-05,
764
  "loss": 0.0009,
765
  "step": 45000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
766
  }
767
  ],
768
  "logging_steps": 500,
@@ -782,7 +852,7 @@
782
  "attributes": {}
783
  }
784
  },
785
- "total_flos": 3.927828479328522e+17,
786
  "train_batch_size": 128,
787
  "trial_name": null,
788
  "trial_params": null
 
1
  {
2
  "best_metric": 0.5455028328611898,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_reduce-opt25-rand-smiles/checkpoint-10000",
4
+ "epoch": 4.931939238508582,
5
  "eval_steps": 2500,
6
+ "global_step": 50000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
763
  "learning_rate": 5e-05,
764
  "loss": 0.0009,
765
  "step": 45000
766
+ },
767
+ {
768
+ "epoch": 4.48806470704281,
769
+ "grad_norm": 0.03474525734782219,
770
+ "learning_rate": 5e-05,
771
+ "loss": 0.0008,
772
+ "step": 45500
773
+ },
774
+ {
775
+ "epoch": 4.537384099427895,
776
+ "grad_norm": 0.036050453782081604,
777
+ "learning_rate": 5e-05,
778
+ "loss": 0.0009,
779
+ "step": 46000
780
+ },
781
+ {
782
+ "epoch": 4.5867034918129805,
783
+ "grad_norm": 0.027342744171619415,
784
+ "learning_rate": 5e-05,
785
+ "loss": 0.0012,
786
+ "step": 46500
787
+ },
788
+ {
789
+ "epoch": 4.636022884198066,
790
+ "grad_norm": 0.03911557421088219,
791
+ "learning_rate": 5e-05,
792
+ "loss": 0.0008,
793
+ "step": 47000
794
+ },
795
+ {
796
+ "epoch": 4.685342276583152,
797
+ "grad_norm": 0.04079333692789078,
798
+ "learning_rate": 5e-05,
799
+ "loss": 0.0009,
800
+ "step": 47500
801
+ },
802
+ {
803
+ "epoch": 4.734661668968238,
804
+ "grad_norm": 0.03646528348326683,
805
+ "learning_rate": 5e-05,
806
+ "loss": 0.0008,
807
+ "step": 48000
808
+ },
809
+ {
810
+ "epoch": 4.783981061353324,
811
+ "grad_norm": 0.0461641363799572,
812
+ "learning_rate": 5e-05,
813
+ "loss": 0.001,
814
+ "step": 48500
815
+ },
816
+ {
817
+ "epoch": 4.83330045373841,
818
+ "grad_norm": 0.050128430128097534,
819
+ "learning_rate": 5e-05,
820
+ "loss": 0.0009,
821
+ "step": 49000
822
+ },
823
+ {
824
+ "epoch": 4.882619846123496,
825
+ "grad_norm": 0.02025548741221428,
826
+ "learning_rate": 5e-05,
827
+ "loss": 0.0008,
828
+ "step": 49500
829
+ },
830
+ {
831
+ "epoch": 4.931939238508582,
832
+ "grad_norm": 0.03685050085186958,
833
+ "learning_rate": 5e-05,
834
+ "loss": 0.0008,
835
+ "step": 50000
836
  }
837
  ],
838
  "logging_steps": 500,
 
852
  "attributes": {}
853
  }
854
  },
855
+ "total_flos": 4.3642997987236454e+17,
856
  "train_batch_size": 128,
857
  "trial_name": null,
858
  "trial_params": null