ribesstefano commited on
Commit
fa81ef8
·
verified ·
1 Parent(s): ae7fd05

Training in progress, step 45000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3a6c87098dee6f04ae2bceaaa2c0f617952d030c4e279874a01eb06fdd42c41e
3
  size 409608164
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87fd26095cc52b1b1fcb61aa241c45efbc7b650471ca5a054f91cf66241e51c2
3
  size 409608164
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5faf6a3e2939d75a948aba9447459c4e343a50ef20460c74841f4d26332215cf
3
  size 814647162
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c07bb32cd46c9fb53458473e6c056fc27d5f5b5b0c76311e6ffe2ee4772ef387
3
  size 814647162
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eeb57d6a99029070321c5cbe53e1de318ca2dbd673dc5c838164e987bf464c4d
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:482495ef335bd0cb4d91c4031a90b31c1acd8f32c3166de3bdbd2f81cf638e6a
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:08e1d8ba2414508e50f3f6e207e18b230dc5f1405db079b8fc4d29a0585a5ea6
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:950318ed157dfaa6d9d1d461c0e1fb839308bf3fb8549cac2587ce9ecd744d04
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.5477160056657224,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_cosine-opt25/checkpoint-10000",
4
- "epoch": 3.945551390806865,
5
  "eval_steps": 2500,
6
- "global_step": 40000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -693,6 +693,76 @@
693
  "learning_rate": 3.27755039594663e-05,
694
  "loss": 0.0006,
695
  "step": 40000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
696
  }
697
  ],
698
  "logging_steps": 500,
@@ -712,7 +782,7 @@
712
  "attributes": {}
713
  }
714
  },
715
- "total_flos": 3.465342568466058e+17,
716
  "train_batch_size": 128,
717
  "trial_name": null,
718
  "trial_params": null
 
1
  {
2
  "best_metric": 0.5477160056657224,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_cosine-opt25/checkpoint-10000",
4
+ "epoch": 4.438745314657724,
5
  "eval_steps": 2500,
6
+ "global_step": 45000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
693
  "learning_rate": 3.27755039594663e-05,
694
  "loss": 0.0006,
695
  "step": 40000
696
+ },
697
+ {
698
+ "epoch": 3.994870783191951,
699
+ "grad_norm": 0.028563622385263443,
700
+ "learning_rate": 3.2400962082482306e-05,
701
+ "loss": 0.0007,
702
+ "step": 40500
703
+ },
704
+ {
705
+ "epoch": 4.044190175577037,
706
+ "grad_norm": 0.012048379518091679,
707
+ "learning_rate": 3.2024590471283946e-05,
708
+ "loss": 0.0006,
709
+ "step": 41000
710
+ },
711
+ {
712
+ "epoch": 4.093509567962123,
713
+ "grad_norm": 0.01336103118956089,
714
+ "learning_rate": 3.164648217593777e-05,
715
+ "loss": 0.0006,
716
+ "step": 41500
717
+ },
718
+ {
719
+ "epoch": 4.142828960347209,
720
+ "grad_norm": 0.04073134437203407,
721
+ "learning_rate": 3.1266730675869434e-05,
722
+ "loss": 0.0006,
723
+ "step": 42000
724
+ },
725
+ {
726
+ "epoch": 4.1921483527322945,
727
+ "grad_norm": 0.03634357079863548,
728
+ "learning_rate": 3.088542985675283e-05,
729
+ "loss": 0.0006,
730
+ "step": 42500
731
+ },
732
+ {
733
+ "epoch": 4.24146774511738,
734
+ "grad_norm": 0.03682945668697357,
735
+ "learning_rate": 3.0502673987298837e-05,
736
+ "loss": 0.0007,
737
+ "step": 43000
738
+ },
739
+ {
740
+ "epoch": 4.290787137502466,
741
+ "grad_norm": 0.030003152787685394,
742
+ "learning_rate": 3.0119327223166655e-05,
743
+ "loss": 0.0007,
744
+ "step": 43500
745
+ },
746
+ {
747
+ "epoch": 4.340106529887552,
748
+ "grad_norm": 0.025118093937635422,
749
+ "learning_rate": 2.973471986909034e-05,
750
+ "loss": 0.0008,
751
+ "step": 44000
752
+ },
753
+ {
754
+ "epoch": 4.389425922272638,
755
+ "grad_norm": 0.01078966073691845,
756
+ "learning_rate": 2.9348172431675086e-05,
757
+ "loss": 0.0005,
758
+ "step": 44500
759
+ },
760
+ {
761
+ "epoch": 4.438745314657724,
762
+ "grad_norm": 0.020486511290073395,
763
+ "learning_rate": 2.8960549998903897e-05,
764
+ "loss": 0.0005,
765
+ "step": 45000
766
  }
767
  ],
768
  "logging_steps": 500,
 
782
  "attributes": {}
783
  }
784
  },
785
+ "total_flos": 3.898212104814858e+17,
786
  "train_batch_size": 128,
787
  "trial_name": null,
788
  "trial_params": null