ribesstefano commited on
Commit
ceed202
·
verified ·
1 Parent(s): 94c25a4

Training in progress, step 40000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9b082085d9aa1da3efa2ad33dd1002ddb5511bf8eb3c2067b8fc30888db68b37
3
  size 409608164
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a6c87098dee6f04ae2bceaaa2c0f617952d030c4e279874a01eb06fdd42c41e
3
  size 409608164
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:053d966c702b26ff46dde77ec6b0cff280c687a0dd1a4474041cda54900e0bf9
3
  size 814647162
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5faf6a3e2939d75a948aba9447459c4e343a50ef20460c74841f4d26332215cf
3
  size 814647162
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:375e808faf0ee5ae5d50ffe1d1a4fc8f21560fc6eeedbd7f2acd27833efa40af
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eeb57d6a99029070321c5cbe53e1de318ca2dbd673dc5c838164e987bf464c4d
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fd12236a2b04dcc8172c63d077f003702ceca140fc00b62558c7c08cdf884e4b
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08e1d8ba2414508e50f3f6e207e18b230dc5f1405db079b8fc4d29a0585a5ea6
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.5477160056657224,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_cosine-opt25/checkpoint-10000",
4
- "epoch": 3.4523574669560073,
5
  "eval_steps": 2500,
6
- "global_step": 35000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -623,6 +623,76 @@
623
  "learning_rate": 3.639806991778405e-05,
624
  "loss": 0.0009,
625
  "step": 35000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
626
  }
627
  ],
628
  "logging_steps": 500,
@@ -642,7 +712,7 @@
642
  "attributes": {}
643
  }
644
  },
645
- "total_flos": 3.032375853388385e+17,
646
  "train_batch_size": 128,
647
  "trial_name": null,
648
  "trial_params": null
 
1
  {
2
  "best_metric": 0.5477160056657224,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_cosine-opt25/checkpoint-10000",
4
+ "epoch": 3.945551390806865,
5
  "eval_steps": 2500,
6
+ "global_step": 40000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
623
  "learning_rate": 3.639806991778405e-05,
624
  "loss": 0.0009,
625
  "step": 35000
626
+ },
627
+ {
628
+ "epoch": 3.501676859341093,
629
+ "grad_norm": 0.023237833753228188,
630
+ "learning_rate": 3.604752099655735e-05,
631
+ "loss": 0.0009,
632
+ "step": 35500
633
+ },
634
+ {
635
+ "epoch": 3.5509962517261786,
636
+ "grad_norm": 0.034029532223939896,
637
+ "learning_rate": 3.569354109150707e-05,
638
+ "loss": 0.0006,
639
+ "step": 36000
640
+ },
641
+ {
642
+ "epoch": 3.6003156441112645,
643
+ "grad_norm": 0.04032694920897484,
644
+ "learning_rate": 3.533691743044375e-05,
645
+ "loss": 0.0007,
646
+ "step": 36500
647
+ },
648
+ {
649
+ "epoch": 3.6496350364963503,
650
+ "grad_norm": 0.03434913605451584,
651
+ "learning_rate": 3.497773818116328e-05,
652
+ "loss": 0.0007,
653
+ "step": 37000
654
+ },
655
+ {
656
+ "epoch": 3.698954428881436,
657
+ "grad_norm": 0.012562757357954979,
658
+ "learning_rate": 3.461609214327767e-05,
659
+ "loss": 0.0007,
660
+ "step": 37500
661
+ },
662
+ {
663
+ "epoch": 3.748273821266522,
664
+ "grad_norm": 0.03795718029141426,
665
+ "learning_rate": 3.4253529436582336e-05,
666
+ "loss": 0.0012,
667
+ "step": 38000
668
+ },
669
+ {
670
+ "epoch": 3.797593213651608,
671
+ "grad_norm": 0.036550700664520264,
672
+ "learning_rate": 3.3887227607163154e-05,
673
+ "loss": 0.0007,
674
+ "step": 38500
675
+ },
676
+ {
677
+ "epoch": 3.8469126060366934,
678
+ "grad_norm": 0.009776749648153782,
679
+ "learning_rate": 3.3518728595236816e-05,
680
+ "loss": 0.0006,
681
+ "step": 39000
682
+ },
683
+ {
684
+ "epoch": 3.8962319984217793,
685
+ "grad_norm": 0.010442009195685387,
686
+ "learning_rate": 3.314812350453317e-05,
687
+ "loss": 0.0006,
688
+ "step": 39500
689
+ },
690
+ {
691
+ "epoch": 3.945551390806865,
692
+ "grad_norm": 0.023100633174180984,
693
+ "learning_rate": 3.27755039594663e-05,
694
+ "loss": 0.0006,
695
+ "step": 40000
696
  }
697
  ],
698
  "logging_steps": 500,
 
712
  "attributes": {}
713
  }
714
  },
715
+ "total_flos": 3.465342568466058e+17,
716
  "train_batch_size": 128,
717
  "trial_name": null,
718
  "trial_params": null