texanrangee commited on
Commit
a1dcf6d
·
verified ·
1 Parent(s): 322ae05

Training in progress, step 1200, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e9728d9ea717659640b33eac820e8f2635cee3b307ef358a0ff4945f21e9ceee
3
  size 339512200
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6a9215a28ac872da52f55aa19af9b4ee906373701fc6e5b55f68e973faffe57
3
  size 339512200
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2b96feeb44097cd300a1f7d72f7938fb3ab57e54288c1f21baff0c575f25a963
3
  size 659562874
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05ad7d0fbd10afe05f5e34de56623ac3ae28854fd1e0a22f3a69eb3d6bbd6439
3
  size 659562874
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7cab2295487651294553c4caf49c38f12ef02ee011d6e686596a41aeb1665f09
3
  size 14180
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:176b20eb86b16b21531afeae0fcad43474b9a44ba3c085cf5ebf304a89fb5f43
3
  size 14180
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:40989f422992086de56b7ab56f6bba81cac547e6785f52be59f14989b370dd0e
3
  size 1256
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:80737f34367bed0f31ec1eeecad8be2c3717a20421ad2a8f693a5747cb780b5c
3
  size 1256
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.7156569361686707,
3
  "best_model_checkpoint": "./output/checkpoint-300",
4
- "epoch": 3.3333333333333335,
5
  "eval_steps": 150,
6
- "global_step": 1050,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -798,6 +798,119 @@
798
  "eval_samples_per_second": 13.193,
799
  "eval_steps_per_second": 13.193,
800
  "step": 1050
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
801
  }
802
  ],
803
  "logging_steps": 10,
@@ -817,7 +930,7 @@
817
  "attributes": {}
818
  }
819
  },
820
- "total_flos": 2.199995203854336e+16,
821
  "train_batch_size": 8,
822
  "trial_name": null,
823
  "trial_params": null
 
1
  {
2
  "best_metric": 0.7156569361686707,
3
  "best_model_checkpoint": "./output/checkpoint-300",
4
+ "epoch": 3.8095238095238093,
5
  "eval_steps": 150,
6
+ "global_step": 1200,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
798
  "eval_samples_per_second": 13.193,
799
  "eval_steps_per_second": 13.193,
800
  "step": 1050
801
+ },
802
+ {
803
+ "epoch": 3.365079365079365,
804
+ "grad_norm": 4.088830947875977,
805
+ "learning_rate": 9.082436730845992e-05,
806
+ "loss": 0.2478,
807
+ "step": 1060
808
+ },
809
+ {
810
+ "epoch": 3.3968253968253967,
811
+ "grad_norm": 3.7507574558258057,
812
+ "learning_rate": 9.063844375723012e-05,
813
+ "loss": 0.1793,
814
+ "step": 1070
815
+ },
816
+ {
817
+ "epoch": 3.4285714285714284,
818
+ "grad_norm": 4.122824668884277,
819
+ "learning_rate": 9.045084971874735e-05,
820
+ "loss": 0.2135,
821
+ "step": 1080
822
+ },
823
+ {
824
+ "epoch": 3.4603174603174605,
825
+ "grad_norm": 8.127439498901367,
826
+ "learning_rate": 9.026159290426779e-05,
827
+ "loss": 0.1958,
828
+ "step": 1090
829
+ },
830
+ {
831
+ "epoch": 3.492063492063492,
832
+ "grad_norm": 3.515589952468872,
833
+ "learning_rate": 9.007068109339781e-05,
834
+ "loss": 0.184,
835
+ "step": 1100
836
+ },
837
+ {
838
+ "epoch": 3.5238095238095237,
839
+ "grad_norm": 4.518334865570068,
840
+ "learning_rate": 8.987812213377421e-05,
841
+ "loss": 0.2064,
842
+ "step": 1110
843
+ },
844
+ {
845
+ "epoch": 3.5555555555555554,
846
+ "grad_norm": 3.4224295616149902,
847
+ "learning_rate": 8.968392394074161e-05,
848
+ "loss": 0.1771,
849
+ "step": 1120
850
+ },
851
+ {
852
+ "epoch": 3.5873015873015874,
853
+ "grad_norm": 5.875314235687256,
854
+ "learning_rate": 8.94880944970271e-05,
855
+ "loss": 0.3107,
856
+ "step": 1130
857
+ },
858
+ {
859
+ "epoch": 3.619047619047619,
860
+ "grad_norm": 6.066832065582275,
861
+ "learning_rate": 8.92906418524121e-05,
862
+ "loss": 0.2793,
863
+ "step": 1140
864
+ },
865
+ {
866
+ "epoch": 3.6507936507936507,
867
+ "grad_norm": 0.021297266706824303,
868
+ "learning_rate": 8.909157412340148e-05,
869
+ "loss": 0.2369,
870
+ "step": 1150
871
+ },
872
+ {
873
+ "epoch": 3.682539682539683,
874
+ "grad_norm": 8.33176040649414,
875
+ "learning_rate": 8.889089949288984e-05,
876
+ "loss": 0.2405,
877
+ "step": 1160
878
+ },
879
+ {
880
+ "epoch": 3.7142857142857144,
881
+ "grad_norm": 4.250291347503662,
882
+ "learning_rate": 8.868862620982532e-05,
883
+ "loss": 0.1879,
884
+ "step": 1170
885
+ },
886
+ {
887
+ "epoch": 3.746031746031746,
888
+ "grad_norm": 7.068946361541748,
889
+ "learning_rate": 8.848476258887028e-05,
890
+ "loss": 0.2618,
891
+ "step": 1180
892
+ },
893
+ {
894
+ "epoch": 3.7777777777777777,
895
+ "grad_norm": 7.198329448699951,
896
+ "learning_rate": 8.827931701005971e-05,
897
+ "loss": 0.2264,
898
+ "step": 1190
899
+ },
900
+ {
901
+ "epoch": 3.8095238095238093,
902
+ "grad_norm": 8.922143936157227,
903
+ "learning_rate": 8.80722979184567e-05,
904
+ "loss": 0.3599,
905
+ "step": 1200
906
+ },
907
+ {
908
+ "epoch": 3.8095238095238093,
909
+ "eval_loss": 0.8716204762458801,
910
+ "eval_runtime": 21.3108,
911
+ "eval_samples_per_second": 13.139,
912
+ "eval_steps_per_second": 13.139,
913
+ "step": 1200
914
  }
915
  ],
916
  "logging_steps": 10,
 
930
  "attributes": {}
931
  }
932
  },
933
+ "total_flos": 2.511073958137037e+16,
934
  "train_batch_size": 8,
935
  "trial_name": null,
936
  "trial_params": null