irishprancer commited on
Commit
2d5a1b6
·
verified ·
1 Parent(s): de7e42e

Training in progress, step 1050, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0983150c1715310f5c2da69abbae72b1cf5ae96dde2a1a3b792fa14cb62311ea
3
  size 1054440872
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b8e949e38a08ba0f4413bafb410ed70d7a41a41b2df06b83f7b9b860bc8aa10
3
  size 1054440872
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9ee576b632d9191a1a1b656e19d65bdddfcd6fc083641530f2da5f3c5efd1e76
3
  size 2041777658
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df8a2bcf283e2aa935d4a4101eb908bed381b3a7d8411b8f0471da588f009acf
3
  size 2041777658
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a492449a8db3f0894cc4cb5d785b2553cbbcc51daceca3e199e1b1c9c7bb1cad
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b3d4f33ba1cd66fae2678da6d84e2417aca4d64da851f275fb6304b99a966c0
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:60a827078c85f0fd366049a88039c9056197c970413749f5c2cbfe85bd3bd7d6
3
  size 1256
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9cf334b45c59ec3117af6235b0cb6b8da2aab8bbc78c388edc1a1925c6731983
3
  size 1256
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 1.3389599323272705,
3
  "best_model_checkpoint": "./output/checkpoint-150",
4
- "epoch": 0.002053926994306058,
5
  "eval_steps": 150,
6
- "global_step": 900,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -685,6 +685,119 @@
685
  "eval_samples_per_second": 10.874,
686
  "eval_steps_per_second": 10.874,
687
  "step": 900
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
688
  }
689
  ],
690
  "logging_steps": 10,
@@ -704,7 +817,7 @@
704
  "attributes": {}
705
  }
706
  },
707
- "total_flos": 2.9630718300266496e+16,
708
  "train_batch_size": 4,
709
  "trial_name": null,
710
  "trial_params": null
 
1
  {
2
  "best_metric": 1.3389599323272705,
3
  "best_model_checkpoint": "./output/checkpoint-150",
4
+ "epoch": 0.0023962481600237344,
5
  "eval_steps": 150,
6
+ "global_step": 1050,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
685
  "eval_samples_per_second": 10.874,
686
  "eval_steps_per_second": 10.874,
687
  "step": 900
688
+ },
689
+ {
690
+ "epoch": 0.002076748405353903,
691
+ "grad_norm": 34.646400451660156,
692
+ "learning_rate": 9.340773121260896e-05,
693
+ "loss": 1.6514,
694
+ "step": 910
695
+ },
696
+ {
697
+ "epoch": 0.002099569816401748,
698
+ "grad_norm": 10.080218315124512,
699
+ "learning_rate": 9.32477428972347e-05,
700
+ "loss": 2.0668,
701
+ "step": 920
702
+ },
703
+ {
704
+ "epoch": 0.0021223912274495934,
705
+ "grad_norm": 16.587017059326172,
706
+ "learning_rate": 9.308597683653978e-05,
707
+ "loss": 1.8328,
708
+ "step": 930
709
+ },
710
+ {
711
+ "epoch": 0.002145212638497438,
712
+ "grad_norm": 40.35968780517578,
713
+ "learning_rate": 9.292243968009333e-05,
714
+ "loss": 2.6786,
715
+ "step": 940
716
+ },
717
+ {
718
+ "epoch": 0.0021680340495452833,
719
+ "grad_norm": 25.870718002319336,
720
+ "learning_rate": 9.275713815026734e-05,
721
+ "loss": 2.0659,
722
+ "step": 950
723
+ },
724
+ {
725
+ "epoch": 0.0021908554605931285,
726
+ "grad_norm": 18.393529891967773,
727
+ "learning_rate": 9.259007904196024e-05,
728
+ "loss": 1.5834,
729
+ "step": 960
730
+ },
731
+ {
732
+ "epoch": 0.0022136768716409737,
733
+ "grad_norm": 15.957837104797363,
734
+ "learning_rate": 9.242126922231766e-05,
735
+ "loss": 1.5815,
736
+ "step": 970
737
+ },
738
+ {
739
+ "epoch": 0.0022364982826888185,
740
+ "grad_norm": 43.54640579223633,
741
+ "learning_rate": 9.225071563045009e-05,
742
+ "loss": 1.8949,
743
+ "step": 980
744
+ },
745
+ {
746
+ "epoch": 0.0022593196937366637,
747
+ "grad_norm": 86.7748794555664,
748
+ "learning_rate": 9.207842527714768e-05,
749
+ "loss": 1.8539,
750
+ "step": 990
751
+ },
752
+ {
753
+ "epoch": 0.002282141104784509,
754
+ "grad_norm": 40.725074768066406,
755
+ "learning_rate": 9.190440524459205e-05,
756
+ "loss": 2.2359,
757
+ "step": 1000
758
+ },
759
+ {
760
+ "epoch": 0.002304962515832354,
761
+ "grad_norm": 17.94641876220703,
762
+ "learning_rate": 9.172866268606516e-05,
763
+ "loss": 1.6349,
764
+ "step": 1010
765
+ },
766
+ {
767
+ "epoch": 0.002327783926880199,
768
+ "grad_norm": 37.57284164428711,
769
+ "learning_rate": 9.155120482565522e-05,
770
+ "loss": 2.2269,
771
+ "step": 1020
772
+ },
773
+ {
774
+ "epoch": 0.002350605337928044,
775
+ "grad_norm": 67.7833480834961,
776
+ "learning_rate": 9.137203895795986e-05,
777
+ "loss": 1.754,
778
+ "step": 1030
779
+ },
780
+ {
781
+ "epoch": 0.002373426748975889,
782
+ "grad_norm": 91.33174133300781,
783
+ "learning_rate": 9.11911724477861e-05,
784
+ "loss": 1.5477,
785
+ "step": 1040
786
+ },
787
+ {
788
+ "epoch": 0.0023962481600237344,
789
+ "grad_norm": 35.31724166870117,
790
+ "learning_rate": 9.100861272984782e-05,
791
+ "loss": 2.1628,
792
+ "step": 1050
793
+ },
794
+ {
795
+ "epoch": 0.0023962481600237344,
796
+ "eval_loss": 1.4954074621200562,
797
+ "eval_runtime": 45.7096,
798
+ "eval_samples_per_second": 10.939,
799
+ "eval_steps_per_second": 10.939,
800
+ "step": 1050
801
  }
802
  ],
803
  "logging_steps": 10,
 
817
  "attributes": {}
818
  }
819
  },
820
+ "total_flos": 3.399070698941645e+16,
821
  "train_batch_size": 4,
822
  "trial_name": null,
823
  "trial_params": null