ribesstefano commited on
Commit
7023b48
·
verified ·
1 Parent(s): 03a5661

Training in progress, step 35000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:847704143a8fe1bda65db1fe2c65f7a5566407d3fa081a0f600bf66f16b8b3a0
3
  size 409608164
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:680cdd8ce09ed81d5904b52f855191e264cb3a8777e419668f407b84fc55713a
3
  size 409608164
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c5957f98f8279a93ad8e35cd4746546f5160d187ed7070c5aa9bbdb47408fe1c
3
  size 814647162
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe583685246694473b81b6159749d52ae9755ca3bd5d94c328cf706a1d1198e0
3
  size 814647162
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c65c2745f22d1c4db409b96f64b6f71c3a2ee8bc2f8c7ff26a84d73ab1a44cb8
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86e4d6f5632dbc00793287b2f22a557709a70a15c41fc36f8e72a42acf0f499d
3
  size 14244
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.5455028328611898,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_reduce-opt25-rand-smiles/checkpoint-10000",
4
- "epoch": 2.959163543105149,
5
  "eval_steps": 2500,
6
- "global_step": 30000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -553,6 +553,76 @@
553
  "learning_rate": 5e-05,
554
  "loss": 0.0021,
555
  "step": 30000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
556
  }
557
  ],
558
  "logging_steps": 500,
@@ -572,7 +642,7 @@
572
  "attributes": {}
573
  }
574
  },
575
- "total_flos": 2.619102484842793e+17,
576
  "train_batch_size": 128,
577
  "trial_name": null,
578
  "trial_params": null
 
1
  {
2
  "best_metric": 0.5455028328611898,
3
  "best_model_checkpoint": "/mimer/NOBACKUP/groups/naiss2023-6-290/stefano/models//PROTAC-Splitter-EncoderDecoder-lr_reduce-opt25-rand-smiles/checkpoint-10000",
4
+ "epoch": 3.4523574669560073,
5
  "eval_steps": 2500,
6
+ "global_step": 35000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
553
  "learning_rate": 5e-05,
554
  "loss": 0.0021,
555
  "step": 30000
556
+ },
557
+ {
558
+ "epoch": 3.008482935490235,
559
+ "grad_norm": 0.029712537303566933,
560
+ "learning_rate": 5e-05,
561
+ "loss": 0.0011,
562
+ "step": 30500
563
+ },
564
+ {
565
+ "epoch": 3.0578023278753204,
566
+ "grad_norm": 0.022233586758375168,
567
+ "learning_rate": 5e-05,
568
+ "loss": 0.001,
569
+ "step": 31000
570
+ },
571
+ {
572
+ "epoch": 3.1071217202604062,
573
+ "grad_norm": 0.04128130525350571,
574
+ "learning_rate": 5e-05,
575
+ "loss": 0.0011,
576
+ "step": 31500
577
+ },
578
+ {
579
+ "epoch": 3.156441112645492,
580
+ "grad_norm": 0.031225621700286865,
581
+ "learning_rate": 5e-05,
582
+ "loss": 0.0012,
583
+ "step": 32000
584
+ },
585
+ {
586
+ "epoch": 3.205760505030578,
587
+ "grad_norm": 0.04397075995802879,
588
+ "learning_rate": 5e-05,
589
+ "loss": 0.0011,
590
+ "step": 32500
591
+ },
592
+ {
593
+ "epoch": 3.255079897415664,
594
+ "grad_norm": 0.03178331255912781,
595
+ "learning_rate": 5e-05,
596
+ "loss": 0.0011,
597
+ "step": 33000
598
+ },
599
+ {
600
+ "epoch": 3.3043992898007497,
601
+ "grad_norm": 0.04911533370614052,
602
+ "learning_rate": 5e-05,
603
+ "loss": 0.0012,
604
+ "step": 33500
605
+ },
606
+ {
607
+ "epoch": 3.3537186821858356,
608
+ "grad_norm": 0.036495912820100784,
609
+ "learning_rate": 5e-05,
610
+ "loss": 0.0011,
611
+ "step": 34000
612
+ },
613
+ {
614
+ "epoch": 3.4030380745709214,
615
+ "grad_norm": 0.0601472444832325,
616
+ "learning_rate": 5e-05,
617
+ "loss": 0.0012,
618
+ "step": 34500
619
+ },
620
+ {
621
+ "epoch": 3.4523574669560073,
622
+ "grad_norm": 0.035126738250255585,
623
+ "learning_rate": 5e-05,
624
+ "loss": 0.0014,
625
+ "step": 35000
626
  }
627
  ],
628
  "logging_steps": 500,
 
642
  "attributes": {}
643
  }
644
  },
645
+ "total_flos": 3.055328543636475e+17,
646
  "train_batch_size": 128,
647
  "trial_name": null,
648
  "trial_params": null