alicegoesdown commited on
Commit
fcbfab9
·
verified ·
1 Parent(s): 77e0f19

Training in progress, step 1050, checkpoint

Browse files
last-checkpoint/lora_top/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ccb49d5271c7b037d22265d5e378c053f9675b00afee51edf29fbed95fb5adc2
3
  size 6299784
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:445dfc8c9a44767a9d29209856d539d1af1c5ffbfea88ff27342cc41a2123780
3
  size 6299784
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:44602cc112706b95fd2b9283dfbb0b273ed1485a42866dd738c66879694b2f35
3
  size 12623930
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:015db8111d5fb0b2719936aa426d41571826aa24cdb6e8099ed8132fa831fed6
3
  size 12623930
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:252022ef9767c17747fa4dd69ffdfd1cd40ac5486666b3d512fd26da71aae964
3
  size 14180
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:494ad52bdfa5ee43ee076b397c9e670106a4c42645a5460ea78ea0cc6f64f64d
3
  size 14180
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fb3c381fb6d3e2736d65db7e2c8174fa727eb1f4bfb816b1d696eaf8cf976d72
3
  size 1256
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:40989f422992086de56b7ab56f6bba81cac547e6785f52be59f14989b370dd0e
3
  size 1256
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "best_metric": 4.2297749519348145,
3
- "best_model_checkpoint": "./output/checkpoint-900",
4
- "epoch": 1.3888888888888888,
5
  "eval_steps": 150,
6
- "global_step": 900,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -685,6 +685,119 @@
685
  "eval_samples_per_second": 40.027,
686
  "eval_steps_per_second": 40.027,
687
  "step": 900
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
688
  }
689
  ],
690
  "logging_steps": 10,
@@ -704,7 +817,7 @@
704
  "attributes": {}
705
  }
706
  },
707
- "total_flos": 3448026176667648.0,
708
  "train_batch_size": 16,
709
  "trial_name": null,
710
  "trial_params": null
 
1
  {
2
+ "best_metric": 4.179497241973877,
3
+ "best_model_checkpoint": "./output/checkpoint-1050",
4
+ "epoch": 1.6203703703703702,
5
  "eval_steps": 150,
6
+ "global_step": 1050,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
685
  "eval_samples_per_second": 40.027,
686
  "eval_steps_per_second": 40.027,
687
  "step": 900
688
+ },
689
+ {
690
+ "epoch": 1.404320987654321,
691
+ "grad_norm": 2.6041247844696045,
692
+ "learning_rate": 9.34077312126089e-05,
693
+ "loss": 3.9427,
694
+ "step": 910
695
+ },
696
+ {
697
+ "epoch": 1.4197530864197532,
698
+ "grad_norm": 2.9636037349700928,
699
+ "learning_rate": 9.324774289723465e-05,
700
+ "loss": 4.0945,
701
+ "step": 920
702
+ },
703
+ {
704
+ "epoch": 1.4351851851851851,
705
+ "grad_norm": 3.2570748329162598,
706
+ "learning_rate": 9.308597683653974e-05,
707
+ "loss": 4.0807,
708
+ "step": 930
709
+ },
710
+ {
711
+ "epoch": 1.4506172839506173,
712
+ "grad_norm": 3.080496072769165,
713
+ "learning_rate": 9.292243968009328e-05,
714
+ "loss": 4.0598,
715
+ "step": 940
716
+ },
717
+ {
718
+ "epoch": 1.4660493827160495,
719
+ "grad_norm": 2.9787585735321045,
720
+ "learning_rate": 9.27571381502673e-05,
721
+ "loss": 4.1266,
722
+ "step": 950
723
+ },
724
+ {
725
+ "epoch": 1.4814814814814814,
726
+ "grad_norm": 2.558279514312744,
727
+ "learning_rate": 9.25900790419602e-05,
728
+ "loss": 4.1626,
729
+ "step": 960
730
+ },
731
+ {
732
+ "epoch": 1.4969135802469136,
733
+ "grad_norm": 3.3579540252685547,
734
+ "learning_rate": 9.24212692223176e-05,
735
+ "loss": 4.0621,
736
+ "step": 970
737
+ },
738
+ {
739
+ "epoch": 1.5123456790123457,
740
+ "grad_norm": 2.9373960494995117,
741
+ "learning_rate": 9.225071563045005e-05,
742
+ "loss": 4.0351,
743
+ "step": 980
744
+ },
745
+ {
746
+ "epoch": 1.5277777777777777,
747
+ "grad_norm": 2.6689186096191406,
748
+ "learning_rate": 9.207842527714764e-05,
749
+ "loss": 4.076,
750
+ "step": 990
751
+ },
752
+ {
753
+ "epoch": 1.5432098765432098,
754
+ "grad_norm": 3.0218076705932617,
755
+ "learning_rate": 9.1904405244592e-05,
756
+ "loss": 3.9422,
757
+ "step": 1000
758
+ },
759
+ {
760
+ "epoch": 1.558641975308642,
761
+ "grad_norm": 3.2610349655151367,
762
+ "learning_rate": 9.172866268606511e-05,
763
+ "loss": 4.0026,
764
+ "step": 1010
765
+ },
766
+ {
767
+ "epoch": 1.574074074074074,
768
+ "grad_norm": 2.5781984329223633,
769
+ "learning_rate": 9.155120482565518e-05,
770
+ "loss": 4.0294,
771
+ "step": 1020
772
+ },
773
+ {
774
+ "epoch": 1.5895061728395061,
775
+ "grad_norm": 2.750657320022583,
776
+ "learning_rate": 9.13720389579598e-05,
777
+ "loss": 4.0155,
778
+ "step": 1030
779
+ },
780
+ {
781
+ "epoch": 1.6049382716049383,
782
+ "grad_norm": 2.2791829109191895,
783
+ "learning_rate": 9.119117244778605e-05,
784
+ "loss": 4.1632,
785
+ "step": 1040
786
+ },
787
+ {
788
+ "epoch": 1.6203703703703702,
789
+ "grad_norm": 3.793689012527466,
790
+ "learning_rate": 9.100861272984777e-05,
791
+ "loss": 4.1136,
792
+ "step": 1050
793
+ },
794
+ {
795
+ "epoch": 1.6203703703703702,
796
+ "eval_loss": 4.179497241973877,
797
+ "eval_runtime": 12.4924,
798
+ "eval_samples_per_second": 40.024,
799
+ "eval_steps_per_second": 40.024,
800
+ "step": 1050
801
  }
802
  ],
803
  "logging_steps": 10,
 
817
  "attributes": {}
818
  }
819
  },
820
+ "total_flos": 4032037194940416.0,
821
  "train_batch_size": 16,
822
  "trial_name": null,
823
  "trial_params": null