ribesstefano commited on
Commit
85c7e01
·
verified ·
1 Parent(s): 1341439

Training in progress, step 55000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:06bc14bbbd27d253e78bd89df250f8a3729fff02beb78d421639b61a114eebe7
3
  size 409608164
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93519581a984b3faeda34e9c51ea07efcda77f7a9d84f527caa96e5b89bf10a9
3
  size 409608164
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:32d96311d632155e63ab79cdb6ae5d13eac2ccb218ce46d58a4f410d578c97c9
3
  size 814647162
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f6444aa8631df227f0ec6227d520da42c363645fbe759571e8f81c553923cf6
3
  size 814647162
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eddbb822fd8c77685fb0b027be7260417141436b826497b6aee7db3cba27626a
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f2699a9b77fb475f94e0e7f2090ac73edfc720b8337803d7322ea61cd0cdd49e
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:129967ad4aacd30894639b032d8517110a4963d1cc93db81c93bfda673142ba8
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a8da913cdb87fa0da659fbfc8e8d4a175dea1907bb6d7b316ea1a236f774834f
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.5389518413597734,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_cosine-opt25-rand-smiles/checkpoint-10000",
4
- "epoch": 4.931939238508582,
5
  "eval_steps": 2500,
6
- "global_step": 50000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -833,6 +833,76 @@
833
  "learning_rate": 2.504638484990372e-05,
834
  "loss": 0.0006,
835
  "step": 50000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
836
  }
837
  ],
838
  "logging_steps": 500,
@@ -852,7 +922,7 @@
852
  "attributes": {}
853
  }
854
  },
855
- "total_flos": 4.3642997987236454e+17,
856
  "train_batch_size": 128,
857
  "trial_name": null,
858
  "trial_params": null
 
1
  {
2
  "best_metric": 0.5389518413597734,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_cosine-opt25-rand-smiles/checkpoint-10000",
4
+ "epoch": 5.425133162359439,
5
  "eval_steps": 2500,
6
+ "global_step": 55000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
833
  "learning_rate": 2.504638484990372e-05,
834
  "loss": 0.0006,
835
  "step": 50000
836
+ },
837
+ {
838
+ "epoch": 4.9812586308936675,
839
+ "grad_norm": 0.015657415613532066,
840
+ "learning_rate": 2.4653303816212188e-05,
841
+ "loss": 0.0006,
842
+ "step": 50500
843
+ },
844
+ {
845
+ "epoch": 5.030578023278753,
846
+ "grad_norm": 0.009272584691643715,
847
+ "learning_rate": 2.42603084959539e-05,
848
+ "loss": 0.0006,
849
+ "step": 51000
850
+ },
851
+ {
852
+ "epoch": 5.079897415663839,
853
+ "grad_norm": 0.016377432271838188,
854
+ "learning_rate": 2.386749604906234e-05,
855
+ "loss": 0.0005,
856
+ "step": 51500
857
+ },
858
+ {
859
+ "epoch": 5.129216808048925,
860
+ "grad_norm": 0.02414529025554657,
861
+ "learning_rate": 2.3475748311231863e-05,
862
+ "loss": 0.0005,
863
+ "step": 52000
864
+ },
865
+ {
866
+ "epoch": 5.178536200434011,
867
+ "grad_norm": 0.023295147344470024,
868
+ "learning_rate": 2.3083592035157375e-05,
869
+ "loss": 0.0005,
870
+ "step": 52500
871
+ },
872
+ {
873
+ "epoch": 5.227855592819097,
874
+ "grad_norm": 0.020559698343276978,
875
+ "learning_rate": 2.2691909551163278e-05,
876
+ "loss": 0.0005,
877
+ "step": 53000
878
+ },
879
+ {
880
+ "epoch": 5.277174985204182,
881
+ "grad_norm": 0.022504296153783798,
882
+ "learning_rate": 2.2300797694611534e-05,
883
+ "loss": 0.0005,
884
+ "step": 53500
885
+ },
886
+ {
887
+ "epoch": 5.326494377589268,
888
+ "grad_norm": 0.006661570630967617,
889
+ "learning_rate": 2.191035315978832e-05,
890
+ "loss": 0.0005,
891
+ "step": 54000
892
+ },
893
+ {
894
+ "epoch": 5.375813769974354,
895
+ "grad_norm": 0.011046066880226135,
896
+ "learning_rate": 2.1520672475998373e-05,
897
+ "loss": 0.0005,
898
+ "step": 54500
899
+ },
900
+ {
901
+ "epoch": 5.425133162359439,
902
+ "grad_norm": 0.02395332045853138,
903
+ "learning_rate": 2.1131851983700107e-05,
904
+ "loss": 0.0005,
905
+ "step": 55000
906
  }
907
  ],
908
  "logging_steps": 500,
 
922
  "attributes": {}
923
  }
924
  },
925
+ "total_flos": 4.800723300014085e+17,
926
  "train_batch_size": 128,
927
  "trial_name": null,
928
  "trial_params": null