ribesstefano commited on
Commit
03aa27e
·
verified ·
1 Parent(s): 19ca090

Training in progress, step 40000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:483ae222a11ad65143d340dfeb9c147487c0e9d7650c854f6fa4e6913822f4ac
3
  size 409608164
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4eb13fae68b4da2fd0470aa5e44d87722b7f922b14cc45570ba15d7332a098a3
3
  size 409608164
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a57b7ac202a308e621e5caa585c7bf49d7660d1bf4e2e4136fe4ce9cfc2f2612
3
  size 814647162
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b4fb83a31c5f551867b73e4a04fb0792b6dcb81cfab8e97c9c7c1be40261673
3
  size 814647162
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:484914ffc89ec808f83bedc7e13af0a69fde5616be4a86784457fcd8d54f74ae
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:67ac518124b0903d77f498378aa78563fdd32b2c504118e8b67305a05ba8b32e
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:253cb0c7aa8e83fabe20fc0de3d6da7c690f3ee9b7052b32a92753a48a7bfa6d
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b3d29e9416426ecd9d3676908bb6dd359c1ff73bc4326e4d53164f0d162e2cb
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.5462110481586402,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_cosine_restarts-opt25/checkpoint-10000",
4
- "epoch": 3.4523574669560073,
5
  "eval_steps": 2500,
6
- "global_step": 35000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -623,6 +623,76 @@
623
  "learning_rate": 3.6959108562141376e-05,
624
  "loss": 0.0012,
625
  "step": 35000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
626
  }
627
  ],
628
  "logging_steps": 500,
@@ -642,7 +712,7 @@
642
  "attributes": {}
643
  }
644
  },
645
- "total_flos": 3.032375853388385e+17,
646
  "train_batch_size": 128,
647
  "trial_name": null,
648
  "trial_params": null
 
1
  {
2
  "best_metric": 0.5462110481586402,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_cosine_restarts-opt25/checkpoint-10000",
4
+ "epoch": 3.945551390806865,
5
  "eval_steps": 2500,
6
+ "global_step": 40000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
623
  "learning_rate": 3.6959108562141376e-05,
624
  "loss": 0.0012,
625
  "step": 35000
626
+ },
627
+ {
628
+ "epoch": 3.501676859341093,
629
+ "grad_norm": 0.021087462082505226,
630
+ "learning_rate": 2.559374318373061e-05,
631
+ "loss": 0.0009,
632
+ "step": 35500
633
+ },
634
+ {
635
+ "epoch": 3.5509962517261786,
636
+ "grad_norm": 0.022866034880280495,
637
+ "learning_rate": 1.409689312542472e-05,
638
+ "loss": 0.0007,
639
+ "step": 36000
640
+ },
641
+ {
642
+ "epoch": 3.6003156441112645,
643
+ "grad_norm": 0.042014122009277344,
644
+ "learning_rate": 5.014540709444851e-06,
645
+ "loss": 0.0006,
646
+ "step": 36500
647
+ },
648
+ {
649
+ "epoch": 3.6496350364963503,
650
+ "grad_norm": 0.012297836132347584,
651
+ "learning_rate": 3.5797673092126393e-07,
652
+ "loss": 0.0005,
653
+ "step": 37000
654
+ },
655
+ {
656
+ "epoch": 3.698954428881436,
657
+ "grad_norm": 0.06086747720837593,
658
+ "learning_rate": 4.884160074185018e-05,
659
+ "loss": 0.002,
660
+ "step": 37500
661
+ },
662
+ {
663
+ "epoch": 3.748273821266522,
664
+ "grad_norm": 0.03979231417179108,
665
+ "learning_rate": 4.279484393132507e-05,
666
+ "loss": 0.0017,
667
+ "step": 38000
668
+ },
669
+ {
670
+ "epoch": 3.797593213651608,
671
+ "grad_norm": 0.020537426695227623,
672
+ "learning_rate": 3.2793161758248515e-05,
673
+ "loss": 0.0011,
674
+ "step": 38500
675
+ },
676
+ {
677
+ "epoch": 3.8469126060366934,
678
+ "grad_norm": 0.012818257324397564,
679
+ "learning_rate": 2.1065680608053368e-05,
680
+ "loss": 0.0007,
681
+ "step": 39000
682
+ },
683
+ {
684
+ "epoch": 3.8962319984217793,
685
+ "grad_norm": 0.009783191606402397,
686
+ "learning_rate": 1.0209456156090146e-05,
687
+ "loss": 0.0006,
688
+ "step": 39500
689
+ },
690
+ {
691
+ "epoch": 3.945551390806865,
692
+ "grad_norm": 0.01907140202820301,
693
+ "learning_rate": 2.628603912310604e-06,
694
+ "loss": 0.0005,
695
+ "step": 40000
696
  }
697
  ],
698
  "logging_steps": 500,
 
712
  "attributes": {}
713
  }
714
  },
715
+ "total_flos": 3.465342568466058e+17,
716
  "train_batch_size": 128,
717
  "trial_name": null,
718
  "trial_params": null