ribesstefano commited on
Commit
31bc121
·
verified ·
1 Parent(s): 9ab4950

Training in progress, step 50000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5e76e2b1a2a56e66260a8b24222402fb6411337a94fc1ff21dd2fed0f6176b3e
3
  size 409608164
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06bc14bbbd27d253e78bd89df250f8a3729fff02beb78d421639b61a114eebe7
3
  size 409608164
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c90a12cb152d8a5960479824e00fe42bc60c63a3a82282630a1fa6788db5ca6a
3
  size 814647162
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32d96311d632155e63ab79cdb6ae5d13eac2ccb218ce46d58a4f410d578c97c9
3
  size 814647162
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:af53b8b54dc8f23640a2242053c5d32aa1e77f56497336f222b5262feb7ef177
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eddbb822fd8c77685fb0b027be7260417141436b826497b6aee7db3cba27626a
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9299d1257fb2340390546c30a874d076b53bb55c63dc6155e353a5f573ba664d
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:129967ad4aacd30894639b032d8517110a4963d1cc93db81c93bfda673142ba8
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.5389518413597734,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_cosine-opt25-rand-smiles/checkpoint-10000",
4
- "epoch": 4.438745314657724,
5
  "eval_steps": 2500,
6
- "global_step": 45000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -763,6 +763,76 @@
763
  "learning_rate": 2.8958997479041255e-05,
764
  "loss": 0.0008,
765
  "step": 45000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
766
  }
767
  ],
768
  "logging_steps": 500,
@@ -782,7 +852,7 @@
782
  "attributes": {}
783
  }
784
  },
785
- "total_flos": 3.927828479328522e+17,
786
  "train_batch_size": 128,
787
  "trial_name": null,
788
  "trial_params": null
 
1
  {
2
  "best_metric": 0.5389518413597734,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_cosine-opt25-rand-smiles/checkpoint-10000",
4
+ "epoch": 4.931939238508582,
5
  "eval_steps": 2500,
6
+ "global_step": 50000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
763
  "learning_rate": 2.8958997479041255e-05,
764
  "loss": 0.0008,
765
  "step": 45000
766
+ },
767
+ {
768
+ "epoch": 4.48806470704281,
769
+ "grad_norm": 0.05904557555913925,
770
+ "learning_rate": 2.857039215862017e-05,
771
+ "loss": 0.0007,
772
+ "step": 45500
773
+ },
774
+ {
775
+ "epoch": 4.537384099427895,
776
+ "grad_norm": 0.025458356365561485,
777
+ "learning_rate": 2.8180904132867185e-05,
778
+ "loss": 0.0007,
779
+ "step": 46000
780
+ },
781
+ {
782
+ "epoch": 4.5867034918129805,
783
+ "grad_norm": 0.025726348161697388,
784
+ "learning_rate": 2.7791410964224228e-05,
785
+ "loss": 0.0009,
786
+ "step": 46500
787
+ },
788
+ {
789
+ "epoch": 4.636022884198066,
790
+ "grad_norm": 0.01956965960562229,
791
+ "learning_rate": 2.740123043483287e-05,
792
+ "loss": 0.0007,
793
+ "step": 47000
794
+ },
795
+ {
796
+ "epoch": 4.685342276583152,
797
+ "grad_norm": 0.038180191069841385,
798
+ "learning_rate": 2.7009674984217738e-05,
799
+ "loss": 0.0007,
800
+ "step": 47500
801
+ },
802
+ {
803
+ "epoch": 4.734661668968238,
804
+ "grad_norm": 0.0360223725438118,
805
+ "learning_rate": 2.661762268318811e-05,
806
+ "loss": 0.0007,
807
+ "step": 48000
808
+ },
809
+ {
810
+ "epoch": 4.783981061353324,
811
+ "grad_norm": 0.03571155667304993,
812
+ "learning_rate": 2.6225170458535525e-05,
813
+ "loss": 0.0007,
814
+ "step": 48500
815
+ },
816
+ {
817
+ "epoch": 4.83330045373841,
818
+ "grad_norm": 0.012124253436923027,
819
+ "learning_rate": 2.5832415335924354e-05,
820
+ "loss": 0.0007,
821
+ "step": 49000
822
+ },
823
+ {
824
+ "epoch": 4.882619846123496,
825
+ "grad_norm": 0.029337838292121887,
826
+ "learning_rate": 2.5439454415904175e-05,
827
+ "loss": 0.0006,
828
+ "step": 49500
829
+ },
830
+ {
831
+ "epoch": 4.931939238508582,
832
+ "grad_norm": 0.04516094923019409,
833
+ "learning_rate": 2.504638484990372e-05,
834
+ "loss": 0.0006,
835
+ "step": 50000
836
  }
837
  ],
838
  "logging_steps": 500,
 
852
  "attributes": {}
853
  }
854
  },
855
+ "total_flos": 4.3642997987236454e+17,
856
  "train_batch_size": 128,
857
  "trial_name": null,
858
  "trial_params": null