ribesstefano commited on
Commit
c777f9c
·
verified ·
1 Parent(s): 8b8dc3f

Training in progress, step 40000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:039842f075c59d724d698a37730d17724d4fd364d3e15cc20b3f1179fde1bca8
3
  size 409608164
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3df87380109d29dca4926961e9f97b620aee35f53d36656b38807372b537c9be
3
  size 409608164
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:89eeaf45dca11e93dc975549990f3cbcb0ebdaffb489fed2b2340133a55bedd8
3
  size 814647162
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5fc238c5d795a910b52830af54cc0780bbea05ab4a8c991e81b09a6c2de7e3b
3
  size 814647162
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1f37fe3424153bca4b7fc59f99bd2c0b98d90aaf8d45d5b05abebae1157a8e7e
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b0963f4ce0cdba7b48e5ad52354a99a0b710c9dba5d5be3f901641a7087746ca
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fd12236a2b04dcc8172c63d077f003702ceca140fc00b62558c7c08cdf884e4b
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:595e7bd6b84a3e5df0100574e8bff7a6926e9fbf535eeac8a208de6027ce6ce4
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.5389518413597734,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_cosine-opt25-rand-smiles/checkpoint-10000",
4
- "epoch": 3.4523574669560073,
5
  "eval_steps": 2500,
6
- "global_step": 35000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -623,6 +623,76 @@
623
  "learning_rate": 3.639806991778405e-05,
624
  "loss": 0.0012,
625
  "step": 35000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
626
  }
627
  ],
628
  "logging_steps": 500,
@@ -642,7 +712,7 @@
642
  "attributes": {}
643
  }
644
  },
645
- "total_flos": 3.055328543636475e+17,
646
  "train_batch_size": 128,
647
  "trial_name": null,
648
  "trial_params": null
 
1
  {
2
  "best_metric": 0.5389518413597734,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_cosine-opt25-rand-smiles/checkpoint-10000",
4
+ "epoch": 3.945551390806865,
5
  "eval_steps": 2500,
6
+ "global_step": 40000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
623
  "learning_rate": 3.639806991778405e-05,
624
  "loss": 0.0012,
625
  "step": 35000
626
+ },
627
+ {
628
+ "epoch": 3.501676859341093,
629
+ "grad_norm": 0.027150381356477737,
630
+ "learning_rate": 3.6046815733494614e-05,
631
+ "loss": 0.0011,
632
+ "step": 35500
633
+ },
634
+ {
635
+ "epoch": 3.5509962517261786,
636
+ "grad_norm": 0.034229960292577744,
637
+ "learning_rate": 3.5692830453371675e-05,
638
+ "loss": 0.001,
639
+ "step": 36000
640
+ },
641
+ {
642
+ "epoch": 3.6003156441112645,
643
+ "grad_norm": 0.04263359308242798,
644
+ "learning_rate": 3.533620159292621e-05,
645
+ "loss": 0.001,
646
+ "step": 36500
647
+ },
648
+ {
649
+ "epoch": 3.6496350364963503,
650
+ "grad_norm": 0.026501759886741638,
651
+ "learning_rate": 3.497701732123957e-05,
652
+ "loss": 0.0011,
653
+ "step": 37000
654
+ },
655
+ {
656
+ "epoch": 3.698954428881436,
657
+ "grad_norm": 0.014317413792014122,
658
+ "learning_rate": 3.4615366439165435e-05,
659
+ "loss": 0.001,
660
+ "step": 37500
661
+ },
662
+ {
663
+ "epoch": 3.748273821266522,
664
+ "grad_norm": 0.04115162417292595,
665
+ "learning_rate": 3.425133835737574e-05,
666
+ "loss": 0.001,
667
+ "step": 38000
668
+ },
669
+ {
670
+ "epoch": 3.797593213651608,
671
+ "grad_norm": 0.01732572540640831,
672
+ "learning_rate": 3.388502307425566e-05,
673
+ "loss": 0.0011,
674
+ "step": 38500
675
+ },
676
+ {
677
+ "epoch": 3.8469126060366934,
678
+ "grad_norm": 0.02579200640320778,
679
+ "learning_rate": 3.3516511153653423e-05,
680
+ "loss": 0.0009,
681
+ "step": 39000
682
+ },
683
+ {
684
+ "epoch": 3.8962319984217793,
685
+ "grad_norm": 0.023984914645552635,
686
+ "learning_rate": 3.314663697789466e-05,
687
+ "loss": 0.0009,
688
+ "step": 39500
689
+ },
690
+ {
691
+ "epoch": 3.945551390806865,
692
+ "grad_norm": 0.05176176130771637,
693
+ "learning_rate": 3.2774009559667783e-05,
694
+ "loss": 0.001,
695
+ "step": 40000
696
  }
697
  ],
698
  "logging_steps": 500,
 
712
  "attributes": {}
713
  }
714
  },
715
+ "total_flos": 3.4915036992864614e+17,
716
  "train_batch_size": 128,
717
  "trial_name": null,
718
  "trial_params": null