irishprancer commited on
Commit
95765d2
·
verified ·
1 Parent(s): 7213d8a

Training in progress, step 750, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ba7511c2605cd169f9d4925ef16d0aab1204009d184f031c3e7da4d154628430
3
  size 527048968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3cd890db48dacc3d3b3df0971ab6f506777ee6324b5fe98afa58345c1314ac67
3
  size 527048968
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4450e064cd4e1c2b71c64918ea5ef37084002b46be2aef8983814714f6dfcb31
3
  size 1054135994
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f50b123a4aeda0cacc062ede7b0235cc81f9613e4e55cf6e5de746a463a52fea
3
  size 1054135994
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b3db7d037567f9907c679104cc622f6046bdf2f33143ca3a29365b2fa4072ce8
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ca9142fcbd976a2b9880762578e5776f18d9cad34016a627060f41ab78ec47d
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b5ff073b2d9654300593116ff77249d188bd5752ecbe3d0cb800964ea5d21c80
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8974fc0327a603be8a90ee235d3ef81c4ee4af2818c873bdeb9bfb38f108e1a
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.7166430950164795,
3
  "best_model_checkpoint": "./output/checkpoint-450",
4
- "epoch": 26.08695652173913,
5
  "eval_steps": 150,
6
- "global_step": 600,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -587,6 +587,151 @@
587
  "EMA_steps_per_second": 23.585,
588
  "epoch": 26.08695652173913,
589
  "step": 600
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
590
  }
591
  ],
592
  "logging_steps": 10,
@@ -606,7 +751,7 @@
606
  "attributes": {}
607
  }
608
  },
609
- "total_flos": 1.5363065839583232e+16,
610
  "train_batch_size": 4,
611
  "trial_name": null,
612
  "trial_params": null
 
1
  {
2
  "best_metric": 0.7166430950164795,
3
  "best_model_checkpoint": "./output/checkpoint-450",
4
+ "epoch": 32.608695652173914,
5
  "eval_steps": 150,
6
+ "global_step": 750,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
587
  "EMA_steps_per_second": 23.585,
588
  "epoch": 26.08695652173913,
589
  "step": 600
590
+ },
591
+ {
592
+ "epoch": 26.52173913043478,
593
+ "grad_norm": 1.0925955772399902,
594
+ "learning_rate": 2.9992268514282142e-05,
595
+ "loss": 0.5116,
596
+ "step": 610
597
+ },
598
+ {
599
+ "epoch": 26.956521739130434,
600
+ "grad_norm": 1.343130111694336,
601
+ "learning_rate": 2.999196237360452e-05,
602
+ "loss": 0.4316,
603
+ "step": 620
604
+ },
605
+ {
606
+ "epoch": 27.391304347826086,
607
+ "grad_norm": 1.203368902206421,
608
+ "learning_rate": 2.9991650290592016e-05,
609
+ "loss": 0.4756,
610
+ "step": 630
611
+ },
612
+ {
613
+ "epoch": 27.82608695652174,
614
+ "grad_norm": 1.3935104608535767,
615
+ "learning_rate": 2.999133226536832e-05,
616
+ "loss": 0.501,
617
+ "step": 640
618
+ },
619
+ {
620
+ "epoch": 28.26086956521739,
621
+ "grad_norm": 1.412856936454773,
622
+ "learning_rate": 2.9991008298059493e-05,
623
+ "loss": 0.4107,
624
+ "step": 650
625
+ },
626
+ {
627
+ "epoch": 28.695652173913043,
628
+ "grad_norm": 1.5606491565704346,
629
+ "learning_rate": 2.9990678388793944e-05,
630
+ "loss": 0.5064,
631
+ "step": 660
632
+ },
633
+ {
634
+ "epoch": 29.130434782608695,
635
+ "grad_norm": 1.315181016921997,
636
+ "learning_rate": 2.999034253770244e-05,
637
+ "loss": 0.4347,
638
+ "step": 670
639
+ },
640
+ {
641
+ "epoch": 29.565217391304348,
642
+ "grad_norm": 1.077691674232483,
643
+ "learning_rate": 2.9990000744918097e-05,
644
+ "loss": 0.4705,
645
+ "step": 680
646
+ },
647
+ {
648
+ "epoch": 30.0,
649
+ "grad_norm": 2.501568078994751,
650
+ "learning_rate": 2.9989653010576392e-05,
651
+ "loss": 0.4145,
652
+ "step": 690
653
+ },
654
+ {
655
+ "epoch": 30.434782608695652,
656
+ "grad_norm": 1.3340367078781128,
657
+ "learning_rate": 2.9989299334815158e-05,
658
+ "loss": 0.4764,
659
+ "step": 700
660
+ },
661
+ {
662
+ "epoch": 30.869565217391305,
663
+ "grad_norm": 1.6282958984375,
664
+ "learning_rate": 2.9988939717774578e-05,
665
+ "loss": 0.4118,
666
+ "step": 710
667
+ },
668
+ {
669
+ "epoch": 31.304347826086957,
670
+ "grad_norm": 0.9019575119018555,
671
+ "learning_rate": 2.9988574159597194e-05,
672
+ "loss": 0.4244,
673
+ "step": 720
674
+ },
675
+ {
676
+ "epoch": 31.73913043478261,
677
+ "grad_norm": 1.6408599615097046,
678
+ "learning_rate": 2.9988202660427907e-05,
679
+ "loss": 0.4821,
680
+ "step": 730
681
+ },
682
+ {
683
+ "epoch": 32.17391304347826,
684
+ "grad_norm": 1.1973698139190674,
685
+ "learning_rate": 2.9987825220413958e-05,
686
+ "loss": 0.4385,
687
+ "step": 740
688
+ },
689
+ {
690
+ "epoch": 32.608695652173914,
691
+ "grad_norm": 1.7692193984985352,
692
+ "learning_rate": 2.998744183970496e-05,
693
+ "loss": 0.4738,
694
+ "step": 750
695
+ },
696
+ {
697
+ "epoch": 32.608695652173914,
698
+ "eval_loss": 0.731913149356842,
699
+ "eval_runtime": 0.5751,
700
+ "eval_samples_per_second": 17.389,
701
+ "eval_steps_per_second": 17.389,
702
+ "step": 750
703
+ },
704
+ {
705
+ "Start_State_loss": 0.8601926565170288,
706
+ "Start_State_runtime": 0.4166,
707
+ "Start_State_samples_per_second": 24.004,
708
+ "Start_State_steps_per_second": 24.004,
709
+ "epoch": 32.608695652173914,
710
+ "step": 750
711
+ },
712
+ {
713
+ "Raw_Model_loss": 0.731913149356842,
714
+ "Raw_Model_runtime": 0.402,
715
+ "Raw_Model_samples_per_second": 24.875,
716
+ "Raw_Model_steps_per_second": 24.875,
717
+ "epoch": 32.608695652173914,
718
+ "step": 750
719
+ },
720
+ {
721
+ "SWA_loss": 0.7290045022964478,
722
+ "SWA_runtime": 0.4054,
723
+ "SWA_samples_per_second": 24.666,
724
+ "SWA_steps_per_second": 24.666,
725
+ "epoch": 32.608695652173914,
726
+ "step": 750
727
+ },
728
+ {
729
+ "EMA_loss": 0.8603373765945435,
730
+ "EMA_runtime": 0.4026,
731
+ "EMA_samples_per_second": 24.839,
732
+ "EMA_steps_per_second": 24.839,
733
+ "epoch": 32.608695652173914,
734
+ "step": 750
735
  }
736
  ],
737
  "logging_steps": 10,
 
751
  "attributes": {}
752
  }
753
  },
754
+ "total_flos": 1.9231766387195904e+16,
755
  "train_batch_size": 4,
756
  "trial_name": null,
757
  "trial_params": null