besimray commited on
Commit
fe57ba1
·
verified ·
1 Parent(s): 3d16d7f

Training in progress, step 120, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:72d67265c9a443fe69b3354dfa4ae158837e22dced7d2197ec76e9fc1416d452
3
  size 45118424
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85fd3349fb0b6d9fe33609fd627bbfa198712a033c54895c31d1954fff998d2f
3
  size 45118424
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:18e24e09b1e55b470e3716dd0556376e6189a9a5ee317a63b79489e372305eff
3
  size 23159290
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ef667d6928d98b9252d546a69290d43eb4b3268ece09ea8d4b8985820baaa05
3
  size 23159290
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0fa2e51ab7a64ce18512351c5dbd00e07cf7826b590c6de414616260bdbb9965
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f9fd0216f0d6187ef1dc8511dbb510792521c2906a3f52d90a5f66d2aa05fa7a
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7736f824b0a772b3806e37af9e860068207311e879196bc19f8a76d97eaf6bce
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61c2b4927e3039b26d377375be782c03ce853d193f96b5868ccf559441e84af9
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "best_metric": 0.7664662599563599,
3
- "best_model_checkpoint": "miner_id_besimray/checkpoint-100",
4
- "epoch": 1.2861736334405145,
5
  "eval_steps": 20,
6
- "global_step": 100,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -755,6 +755,154 @@
755
  "eval_samples_per_second": 21.558,
756
  "eval_steps_per_second": 2.235,
757
  "step": 100
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
758
  }
759
  ],
760
  "logging_steps": 1,
@@ -783,7 +931,7 @@
783
  "attributes": {}
784
  }
785
  },
786
- "total_flos": 2.394546334728192e+16,
787
  "train_batch_size": 10,
788
  "trial_name": null,
789
  "trial_params": null
 
1
  {
2
+ "best_metric": 0.7635705471038818,
3
+ "best_model_checkpoint": "miner_id_besimray/checkpoint-120",
4
+ "epoch": 1.5434083601286175,
5
  "eval_steps": 20,
6
+ "global_step": 120,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
755
  "eval_samples_per_second": 21.558,
756
  "eval_steps_per_second": 2.235,
757
  "step": 100
758
+ },
759
+ {
760
+ "epoch": 1.2990353697749195,
761
+ "grad_norm": 0.27163776755332947,
762
+ "learning_rate": 5.4600950026045326e-05,
763
+ "loss": 0.8166,
764
+ "step": 101
765
+ },
766
+ {
767
+ "epoch": 1.3118971061093248,
768
+ "grad_norm": 0.24301646649837494,
769
+ "learning_rate": 5.261313375270014e-05,
770
+ "loss": 0.7311,
771
+ "step": 102
772
+ },
773
+ {
774
+ "epoch": 1.32475884244373,
775
+ "grad_norm": 0.2601422369480133,
776
+ "learning_rate": 5.0649178193565314e-05,
777
+ "loss": 0.8192,
778
+ "step": 103
779
+ },
780
+ {
781
+ "epoch": 1.337620578778135,
782
+ "grad_norm": 0.28303226828575134,
783
+ "learning_rate": 4.87100722594094e-05,
784
+ "loss": 0.7946,
785
+ "step": 104
786
+ },
787
+ {
788
+ "epoch": 1.3504823151125402,
789
+ "grad_norm": 0.25355738401412964,
790
+ "learning_rate": 4.6796792348466356e-05,
791
+ "loss": 0.7246,
792
+ "step": 105
793
+ },
794
+ {
795
+ "epoch": 1.3633440514469453,
796
+ "grad_norm": 0.2478146255016327,
797
+ "learning_rate": 4.491030185478976e-05,
798
+ "loss": 0.8159,
799
+ "step": 106
800
+ },
801
+ {
802
+ "epoch": 1.3762057877813505,
803
+ "grad_norm": 0.22157523036003113,
804
+ "learning_rate": 4.305155068315481e-05,
805
+ "loss": 0.6485,
806
+ "step": 107
807
+ },
808
+ {
809
+ "epoch": 1.3890675241157555,
810
+ "grad_norm": 0.28142282366752625,
811
+ "learning_rate": 4.12214747707527e-05,
812
+ "loss": 0.7363,
813
+ "step": 108
814
+ },
815
+ {
816
+ "epoch": 1.4019292604501608,
817
+ "grad_norm": 0.21578700840473175,
818
+ "learning_rate": 3.942099561591802e-05,
819
+ "loss": 0.6999,
820
+ "step": 109
821
+ },
822
+ {
823
+ "epoch": 1.414790996784566,
824
+ "grad_norm": 0.24640053510665894,
825
+ "learning_rate": 3.7651019814126654e-05,
826
+ "loss": 0.7388,
827
+ "step": 110
828
+ },
829
+ {
830
+ "epoch": 1.427652733118971,
831
+ "grad_norm": 0.21803902089595795,
832
+ "learning_rate": 3.591243860149759e-05,
833
+ "loss": 0.6631,
834
+ "step": 111
835
+ },
836
+ {
837
+ "epoch": 1.4405144694533762,
838
+ "grad_norm": 0.2544996440410614,
839
+ "learning_rate": 3.4206127406028745e-05,
840
+ "loss": 0.6946,
841
+ "step": 112
842
+ },
843
+ {
844
+ "epoch": 1.4533762057877815,
845
+ "grad_norm": 0.25184303522109985,
846
+ "learning_rate": 3.253294540679257e-05,
847
+ "loss": 0.6731,
848
+ "step": 113
849
+ },
850
+ {
851
+ "epoch": 1.4662379421221865,
852
+ "grad_norm": 0.25497835874557495,
853
+ "learning_rate": 3.089373510131354e-05,
854
+ "loss": 0.7296,
855
+ "step": 114
856
+ },
857
+ {
858
+ "epoch": 1.4790996784565915,
859
+ "grad_norm": 0.273090660572052,
860
+ "learning_rate": 2.9289321881345254e-05,
861
+ "loss": 0.7861,
862
+ "step": 115
863
+ },
864
+ {
865
+ "epoch": 1.4919614147909968,
866
+ "grad_norm": 0.278880774974823,
867
+ "learning_rate": 2.7720513617260856e-05,
868
+ "loss": 0.7271,
869
+ "step": 116
870
+ },
871
+ {
872
+ "epoch": 1.504823151125402,
873
+ "grad_norm": 0.2511500120162964,
874
+ "learning_rate": 2.6188100251265945e-05,
875
+ "loss": 0.6752,
876
+ "step": 117
877
+ },
878
+ {
879
+ "epoch": 1.517684887459807,
880
+ "grad_norm": 0.2742422819137573,
881
+ "learning_rate": 2.4692853399638917e-05,
882
+ "loss": 0.7954,
883
+ "step": 118
884
+ },
885
+ {
886
+ "epoch": 1.5305466237942122,
887
+ "grad_norm": 0.274632066488266,
888
+ "learning_rate": 2.323552596419889e-05,
889
+ "loss": 0.8,
890
+ "step": 119
891
+ },
892
+ {
893
+ "epoch": 1.5434083601286175,
894
+ "grad_norm": 0.25790005922317505,
895
+ "learning_rate": 2.181685175319702e-05,
896
+ "loss": 0.8212,
897
+ "step": 120
898
+ },
899
+ {
900
+ "epoch": 1.5434083601286175,
901
+ "eval_loss": 0.7635705471038818,
902
+ "eval_runtime": 5.7735,
903
+ "eval_samples_per_second": 28.406,
904
+ "eval_steps_per_second": 2.945,
905
+ "step": 120
906
  }
907
  ],
908
  "logging_steps": 1,
 
931
  "attributes": {}
932
  }
933
  },
934
+ "total_flos": 2.858372738187264e+16,
935
  "train_batch_size": 10,
936
  "trial_name": null,
937
  "trial_params": null