irishprancer commited on
Commit
65263b9
·
verified ·
1 Parent(s): a448f10

Training in progress, step 1050, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:40e24560228e7d2978eabd16e3650531b45d9a750bc9d07e5c6599b62ea323f2
3
  size 888392856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:082057aa43ad4ab7720869b29ef30158762f116e91bb5a5c5ba20dd343d39b13
3
  size 888392856
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:62f00fcd6458929b08666bcc1161c1c5bb3616c73b0923ac024450838f886472
3
  size 1724014074
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b5a5de184b932a90ab2a229e1bc76a1cf6ebc20738808f656f5d2e4d0b48c01
3
  size 1724014074
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:795ff3ab1a3fad7a4fadc5b5b41d290cb0cea6133c8206ce725785c7d25c85ec
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b3fca934b31e66071a287bca1241ed82b21fe605c83ad79042a2a656622d0a9
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:afcf237ffb833ad28294e886af7d2dc67a409568829cbb127d7df34e2d3d6579
3
  size 1256
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d142240d9d34452809b9cf34c6ae579fd7cc4f0b4e78197fa4853e45b45884a6
3
  size 1256
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 1.1937243938446045,
3
  "best_model_checkpoint": "./output/checkpoint-150",
4
- "epoch": 0.002053926994306058,
5
  "eval_steps": 150,
6
- "global_step": 900,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -685,6 +685,119 @@
685
  "eval_samples_per_second": 11.765,
686
  "eval_steps_per_second": 11.765,
687
  "step": 900
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
688
  }
689
  ],
690
  "logging_steps": 10,
@@ -704,7 +817,7 @@
704
  "attributes": {}
705
  }
706
  },
707
- "total_flos": 3.124423356101376e+16,
708
  "train_batch_size": 4,
709
  "trial_name": null,
710
  "trial_params": null
 
1
  {
2
  "best_metric": 1.1937243938446045,
3
  "best_model_checkpoint": "./output/checkpoint-150",
4
+ "epoch": 0.0023962481600237344,
5
  "eval_steps": 150,
6
+ "global_step": 1050,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
685
  "eval_samples_per_second": 11.765,
686
  "eval_steps_per_second": 11.765,
687
  "step": 900
688
+ },
689
+ {
690
+ "epoch": 0.002076748405353903,
691
+ "grad_norm": 61.62263488769531,
692
+ "learning_rate": 2.0886603661542245e-05,
693
+ "loss": 1.4012,
694
+ "step": 910
695
+ },
696
+ {
697
+ "epoch": 0.002099569816401748,
698
+ "grad_norm": 59.046051025390625,
699
+ "learning_rate": 2.0850829186663994e-05,
700
+ "loss": 2.1143,
701
+ "step": 920
702
+ },
703
+ {
704
+ "epoch": 0.0021223912274495934,
705
+ "grad_norm": 25.007780075073242,
706
+ "learning_rate": 2.0814657195847375e-05,
707
+ "loss": 1.5044,
708
+ "step": 930
709
+ },
710
+ {
711
+ "epoch": 0.002145212638497438,
712
+ "grad_norm": 42.121761322021484,
713
+ "learning_rate": 2.077808917598125e-05,
714
+ "loss": 1.6489,
715
+ "step": 940
716
+ },
717
+ {
718
+ "epoch": 0.0021680340495452833,
719
+ "grad_norm": 62.19163131713867,
720
+ "learning_rate": 2.0741126630233687e-05,
721
+ "loss": 1.5162,
722
+ "step": 950
723
+ },
724
+ {
725
+ "epoch": 0.0021908554605931285,
726
+ "grad_norm": 25.229106903076172,
727
+ "learning_rate": 2.070377107799017e-05,
728
+ "loss": 1.7089,
729
+ "step": 960
730
+ },
731
+ {
732
+ "epoch": 0.0022136768716409737,
733
+ "grad_norm": 41.89048385620117,
734
+ "learning_rate": 2.0666024054791137e-05,
735
+ "loss": 1.6282,
736
+ "step": 970
737
+ },
738
+ {
739
+ "epoch": 0.0022364982826888185,
740
+ "grad_norm": 39.0805549621582,
741
+ "learning_rate": 2.0627887112268875e-05,
742
+ "loss": 1.614,
743
+ "step": 980
744
+ },
745
+ {
746
+ "epoch": 0.0022593196937366637,
747
+ "grad_norm": 35.681884765625,
748
+ "learning_rate": 2.0589361818083712e-05,
749
+ "loss": 2.1457,
750
+ "step": 990
751
+ },
752
+ {
753
+ "epoch": 0.002282141104784509,
754
+ "grad_norm": 36.427616119384766,
755
+ "learning_rate": 2.0550449755859598e-05,
756
+ "loss": 1.257,
757
+ "step": 1000
758
+ },
759
+ {
760
+ "epoch": 0.002304962515832354,
761
+ "grad_norm": 85.96231079101562,
762
+ "learning_rate": 2.0511152525119014e-05,
763
+ "loss": 1.6372,
764
+ "step": 1010
765
+ },
766
+ {
767
+ "epoch": 0.002327783926880199,
768
+ "grad_norm": 67.32917785644531,
769
+ "learning_rate": 2.0471471741217183e-05,
770
+ "loss": 1.4432,
771
+ "step": 1020
772
+ },
773
+ {
774
+ "epoch": 0.002350605337928044,
775
+ "grad_norm": 82.99893188476562,
776
+ "learning_rate": 2.0431409035275724e-05,
777
+ "loss": 1.925,
778
+ "step": 1030
779
+ },
780
+ {
781
+ "epoch": 0.002373426748975889,
782
+ "grad_norm": 79.15724182128906,
783
+ "learning_rate": 2.0390966054115558e-05,
784
+ "loss": 1.2961,
785
+ "step": 1040
786
+ },
787
+ {
788
+ "epoch": 0.0023962481600237344,
789
+ "grad_norm": 29.493953704833984,
790
+ "learning_rate": 2.035014446018924e-05,
791
+ "loss": 1.6837,
792
+ "step": 1050
793
+ },
794
+ {
795
+ "epoch": 0.0023962481600237344,
796
+ "eval_loss": 1.3128489255905151,
797
+ "eval_runtime": 44.2241,
798
+ "eval_samples_per_second": 11.306,
799
+ "eval_steps_per_second": 11.306,
800
+ "step": 1050
801
  }
802
  ],
803
  "logging_steps": 10,
 
817
  "attributes": {}
818
  }
819
  },
820
+ "total_flos": 3.58409393634432e+16,
821
  "train_batch_size": 4,
822
  "trial_name": null,
823
  "trial_params": null