ribesstefano commited on
Commit
196abcb
·
verified ·
1 Parent(s): 83c873b

Training in progress, step 45000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:503ce751da80aa8b1f118a861d5757c844ce8de1db073fef17acad2029b59e2b
3
  size 409608164
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:99d99ef005366c768182dbd5fbbed197e6391e6fc6d0c746b5feb7b2fd2a2764
3
  size 409608164
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6fbaedfb0a4123dcead5a29fb791a66763b1c12ca90eb39e7c861a4cb61e314d
3
  size 814647162
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f6b17853dc9aaa706ba3ed378aef704b5d5535f20af1ae291d3e1c2ee5cb5bc
3
  size 814647162
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:39417990b3af4b754d49ef8161883cbdefa4e1231e6e279b71d13d553fe650cf
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76ef189c202797a47ce62889d21e7c97c6a4ce5ace6abf783017722de7ad98b2
3
  size 14244
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.5455028328611898,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_reduce-opt25-rand-smiles/checkpoint-10000",
4
- "epoch": 3.945551390806865,
5
  "eval_steps": 2500,
6
- "global_step": 40000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -693,6 +693,76 @@
693
  "learning_rate": 5e-05,
694
  "loss": 0.001,
695
  "step": 40000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
696
  }
697
  ],
698
  "logging_steps": 500,
@@ -712,7 +782,7 @@
712
  "attributes": {}
713
  }
714
  },
715
- "total_flos": 3.4915036992864614e+17,
716
  "train_batch_size": 128,
717
  "trial_name": null,
718
  "trial_params": null
 
1
  {
2
  "best_metric": 0.5455028328611898,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_reduce-opt25-rand-smiles/checkpoint-10000",
4
+ "epoch": 4.438745314657724,
5
  "eval_steps": 2500,
6
+ "global_step": 45000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
693
  "learning_rate": 5e-05,
694
  "loss": 0.001,
695
  "step": 40000
696
+ },
697
+ {
698
+ "epoch": 3.994870783191951,
699
+ "grad_norm": 0.02784532681107521,
700
+ "learning_rate": 5e-05,
701
+ "loss": 0.0011,
702
+ "step": 40500
703
+ },
704
+ {
705
+ "epoch": 4.044190175577037,
706
+ "grad_norm": 0.03348594531416893,
707
+ "learning_rate": 5e-05,
708
+ "loss": 0.001,
709
+ "step": 41000
710
+ },
711
+ {
712
+ "epoch": 4.093509567962123,
713
+ "grad_norm": 0.06848876923322678,
714
+ "learning_rate": 5e-05,
715
+ "loss": 0.0009,
716
+ "step": 41500
717
+ },
718
+ {
719
+ "epoch": 4.142828960347209,
720
+ "grad_norm": 0.023778628557920456,
721
+ "learning_rate": 5e-05,
722
+ "loss": 0.0009,
723
+ "step": 42000
724
+ },
725
+ {
726
+ "epoch": 4.1921483527322945,
727
+ "grad_norm": 0.014797426760196686,
728
+ "learning_rate": 5e-05,
729
+ "loss": 0.0009,
730
+ "step": 42500
731
+ },
732
+ {
733
+ "epoch": 4.24146774511738,
734
+ "grad_norm": 0.019765177741646767,
735
+ "learning_rate": 5e-05,
736
+ "loss": 0.0009,
737
+ "step": 43000
738
+ },
739
+ {
740
+ "epoch": 4.290787137502466,
741
+ "grad_norm": 0.03716424107551575,
742
+ "learning_rate": 5e-05,
743
+ "loss": 0.001,
744
+ "step": 43500
745
+ },
746
+ {
747
+ "epoch": 4.340106529887552,
748
+ "grad_norm": 0.04304041340947151,
749
+ "learning_rate": 5e-05,
750
+ "loss": 0.0009,
751
+ "step": 44000
752
+ },
753
+ {
754
+ "epoch": 4.389425922272638,
755
+ "grad_norm": 0.07148012518882751,
756
+ "learning_rate": 5e-05,
757
+ "loss": 0.0011,
758
+ "step": 44500
759
+ },
760
+ {
761
+ "epoch": 4.438745314657724,
762
+ "grad_norm": 0.025658031925559044,
763
+ "learning_rate": 5e-05,
764
+ "loss": 0.0009,
765
+ "step": 45000
766
  }
767
  ],
768
  "logging_steps": 500,
 
782
  "attributes": {}
783
  }
784
  },
785
+ "total_flos": 3.927828479328522e+17,
786
  "train_batch_size": 128,
787
  "trial_name": null,
788
  "trial_params": null