ribesstefano commited on
Commit
e72f654
·
verified ·
1 Parent(s): 722f088

Training in progress, step 50000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:87fd26095cc52b1b1fcb61aa241c45efbc7b650471ca5a054f91cf66241e51c2
3
  size 409608164
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bbd952fb7ca0f8ddc9b0e7256bf820a3fc04551b8ae426ac98c3094dee2ea254
3
  size 409608164
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c07bb32cd46c9fb53458473e6c056fc27d5f5b5b0c76311e6ffe2ee4772ef387
3
  size 814647162
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d7970449ed8a17f9c6c4492b954ad4964b40feccf1db02f9503f0b2b341dee5
3
  size 814647162
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:482495ef335bd0cb4d91c4031a90b31c1acd8f32c3166de3bdbd2f81cf638e6a
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44f6a18681a37c28eaea9106736980f3cb0b876fb22eb09649d7485450746688
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:950318ed157dfaa6d9d1d461c0e1fb839308bf3fb8549cac2587ce9ecd744d04
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:129967ad4aacd30894639b032d8517110a4963d1cc93db81c93bfda673142ba8
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.5477160056657224,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_cosine-opt25/checkpoint-10000",
4
- "epoch": 4.438745314657724,
5
  "eval_steps": 2500,
6
- "global_step": 45000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -763,6 +763,76 @@
763
  "learning_rate": 2.8960549998903897e-05,
764
  "loss": 0.0005,
765
  "step": 45000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
766
  }
767
  ],
768
  "logging_steps": 500,
@@ -782,7 +852,7 @@
782
  "attributes": {}
783
  }
784
  },
785
- "total_flos": 3.898212104814858e+17,
786
  "train_batch_size": 128,
787
  "trial_name": null,
788
  "trial_params": null
 
1
  {
2
  "best_metric": 0.5477160056657224,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_cosine-opt25/checkpoint-10000",
4
+ "epoch": 4.931939238508582,
5
  "eval_steps": 2500,
6
+ "global_step": 50000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
763
  "learning_rate": 2.8960549998903897e-05,
764
  "loss": 0.0005,
765
  "step": 45000
766
+ },
767
+ {
768
+ "epoch": 4.48806470704281,
769
+ "grad_norm": 0.06772090494632721,
770
+ "learning_rate": 2.857194840237535e-05,
771
+ "loss": 0.0005,
772
+ "step": 45500
773
+ },
774
+ {
775
+ "epoch": 4.537384099427895,
776
+ "grad_norm": 0.05354917794466019,
777
+ "learning_rate": 2.818246371576595e-05,
778
+ "loss": 0.0005,
779
+ "step": 46000
780
+ },
781
+ {
782
+ "epoch": 4.5867034918129805,
783
+ "grad_norm": 0.034519363194704056,
784
+ "learning_rate": 2.7792192231077896e-05,
785
+ "loss": 0.0005,
786
+ "step": 46500
787
+ },
788
+ {
789
+ "epoch": 4.636022884198066,
790
+ "grad_norm": 0.02402963489294052,
791
+ "learning_rate": 2.740123043483287e-05,
792
+ "loss": 0.0005,
793
+ "step": 47000
794
+ },
795
+ {
796
+ "epoch": 4.685342276583152,
797
+ "grad_norm": 0.03265873342752457,
798
+ "learning_rate": 2.7009674984217738e-05,
799
+ "loss": 0.0005,
800
+ "step": 47500
801
+ },
802
+ {
803
+ "epoch": 4.734661668968238,
804
+ "grad_norm": 0.04922989383339882,
805
+ "learning_rate": 2.661762268318811e-05,
806
+ "loss": 0.0005,
807
+ "step": 48000
808
+ },
809
+ {
810
+ "epoch": 4.783981061353324,
811
+ "grad_norm": 0.02939719334244728,
812
+ "learning_rate": 2.6225170458535525e-05,
813
+ "loss": 0.0005,
814
+ "step": 48500
815
+ },
816
+ {
817
+ "epoch": 4.83330045373841,
818
+ "grad_norm": 0.043153587728738785,
819
+ "learning_rate": 2.5832415335924354e-05,
820
+ "loss": 0.0005,
821
+ "step": 49000
822
+ },
823
+ {
824
+ "epoch": 4.882619846123496,
825
+ "grad_norm": 0.024349920451641083,
826
+ "learning_rate": 2.5439454415904175e-05,
827
+ "loss": 0.0005,
828
+ "step": 49500
829
+ },
830
+ {
831
+ "epoch": 4.931939238508582,
832
+ "grad_norm": 0.04341769590973854,
833
+ "learning_rate": 2.504638484990372e-05,
834
+ "loss": 0.0004,
835
+ "step": 50000
836
  }
837
  ],
838
  "logging_steps": 500,
 
852
  "attributes": {}
853
  }
854
  },
855
+ "total_flos": 4.3312528608288154e+17,
856
  "train_batch_size": 128,
857
  "trial_name": null,
858
  "trial_params": null