besimray commited on
Commit
0578027
·
verified ·
1 Parent(s): bac49f4

Training in progress, step 120, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:96ccc5b9ac087dba0f10fdc4d5342763320b3d393d47476a061e07710ee9b638
3
  size 45118424
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:833eb01138783f900186e10d4e4abdce2331f847c08833bd0fa504402085a60b
3
  size 45118424
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b7e67d1ef70988eb681b3fa8d5f97f7cf8dd017881f8271c13d1e14c3a983051
3
  size 23159290
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be261bb291f9b4a5d55868833557a7b0f1e34d282ee624617485ad6c262a8639
3
  size 23159290
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0432b165814871079f9ee89b9acfe77d274309a01b9cb7e1f0b9e05d72efa489
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d57cb21cc7fa77e1d0615ec039ca096efa9b134639fd673e38d6b4c2602fef16
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7736f824b0a772b3806e37af9e860068207311e879196bc19f8a76d97eaf6bce
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61c2b4927e3039b26d377375be782c03ce853d193f96b5868ccf559441e84af9
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "best_metric": 0.09471629559993744,
3
- "best_model_checkpoint": "miner_id_besimray/checkpoint-100",
4
- "epoch": 1.2861736334405145,
5
  "eval_steps": 20,
6
- "global_step": 100,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -755,6 +755,154 @@
755
  "eval_samples_per_second": 23.861,
756
  "eval_steps_per_second": 2.473,
757
  "step": 100
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
758
  }
759
  ],
760
  "logging_steps": 1,
@@ -783,7 +931,7 @@
783
  "attributes": {}
784
  }
785
  },
786
- "total_flos": 2.394546334728192e+16,
787
  "train_batch_size": 10,
788
  "trial_name": null,
789
  "trial_params": null
 
1
  {
2
+ "best_metric": 0.07654841244220734,
3
+ "best_model_checkpoint": "miner_id_besimray/checkpoint-120",
4
+ "epoch": 1.5434083601286175,
5
  "eval_steps": 20,
6
+ "global_step": 120,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
755
  "eval_samples_per_second": 23.861,
756
  "eval_steps_per_second": 2.473,
757
  "step": 100
758
+ },
759
+ {
760
+ "epoch": 1.2990353697749195,
761
+ "grad_norm": 0.821508526802063,
762
+ "learning_rate": 5.4600950026045326e-05,
763
+ "loss": 0.0432,
764
+ "step": 101
765
+ },
766
+ {
767
+ "epoch": 1.3118971061093248,
768
+ "grad_norm": 2.061079740524292,
769
+ "learning_rate": 5.261313375270014e-05,
770
+ "loss": 0.0979,
771
+ "step": 102
772
+ },
773
+ {
774
+ "epoch": 1.32475884244373,
775
+ "grad_norm": 1.4887949228286743,
776
+ "learning_rate": 5.0649178193565314e-05,
777
+ "loss": 0.0788,
778
+ "step": 103
779
+ },
780
+ {
781
+ "epoch": 1.337620578778135,
782
+ "grad_norm": 0.8588582873344421,
783
+ "learning_rate": 4.87100722594094e-05,
784
+ "loss": 0.0554,
785
+ "step": 104
786
+ },
787
+ {
788
+ "epoch": 1.3504823151125402,
789
+ "grad_norm": 0.8016582131385803,
790
+ "learning_rate": 4.6796792348466356e-05,
791
+ "loss": 0.065,
792
+ "step": 105
793
+ },
794
+ {
795
+ "epoch": 1.3633440514469453,
796
+ "grad_norm": 1.197761058807373,
797
+ "learning_rate": 4.491030185478976e-05,
798
+ "loss": 0.1194,
799
+ "step": 106
800
+ },
801
+ {
802
+ "epoch": 1.3762057877813505,
803
+ "grad_norm": 0.5511653423309326,
804
+ "learning_rate": 4.305155068315481e-05,
805
+ "loss": 0.032,
806
+ "step": 107
807
+ },
808
+ {
809
+ "epoch": 1.3890675241157555,
810
+ "grad_norm": 1.1377793550491333,
811
+ "learning_rate": 4.12214747707527e-05,
812
+ "loss": 0.0882,
813
+ "step": 108
814
+ },
815
+ {
816
+ "epoch": 1.4019292604501608,
817
+ "grad_norm": 0.49581602215766907,
818
+ "learning_rate": 3.942099561591802e-05,
819
+ "loss": 0.0292,
820
+ "step": 109
821
+ },
822
+ {
823
+ "epoch": 1.414790996784566,
824
+ "grad_norm": 1.2391666173934937,
825
+ "learning_rate": 3.7651019814126654e-05,
826
+ "loss": 0.0838,
827
+ "step": 110
828
+ },
829
+ {
830
+ "epoch": 1.427652733118971,
831
+ "grad_norm": 0.8613415360450745,
832
+ "learning_rate": 3.591243860149759e-05,
833
+ "loss": 0.059,
834
+ "step": 111
835
+ },
836
+ {
837
+ "epoch": 1.4405144694533762,
838
+ "grad_norm": 1.378289818763733,
839
+ "learning_rate": 3.4206127406028745e-05,
840
+ "loss": 0.0894,
841
+ "step": 112
842
+ },
843
+ {
844
+ "epoch": 1.4533762057877815,
845
+ "grad_norm": 1.2210605144500732,
846
+ "learning_rate": 3.253294540679257e-05,
847
+ "loss": 0.1327,
848
+ "step": 113
849
+ },
850
+ {
851
+ "epoch": 1.4662379421221865,
852
+ "grad_norm": 1.610064148902893,
853
+ "learning_rate": 3.089373510131354e-05,
854
+ "loss": 0.1172,
855
+ "step": 114
856
+ },
857
+ {
858
+ "epoch": 1.4790996784565915,
859
+ "grad_norm": 0.9988131523132324,
860
+ "learning_rate": 2.9289321881345254e-05,
861
+ "loss": 0.0927,
862
+ "step": 115
863
+ },
864
+ {
865
+ "epoch": 1.4919614147909968,
866
+ "grad_norm": 0.8481462597846985,
867
+ "learning_rate": 2.7720513617260856e-05,
868
+ "loss": 0.0406,
869
+ "step": 116
870
+ },
871
+ {
872
+ "epoch": 1.504823151125402,
873
+ "grad_norm": 0.4117134213447571,
874
+ "learning_rate": 2.6188100251265945e-05,
875
+ "loss": 0.0235,
876
+ "step": 117
877
+ },
878
+ {
879
+ "epoch": 1.517684887459807,
880
+ "grad_norm": 1.1421048641204834,
881
+ "learning_rate": 2.4692853399638917e-05,
882
+ "loss": 0.0818,
883
+ "step": 118
884
+ },
885
+ {
886
+ "epoch": 1.5305466237942122,
887
+ "grad_norm": 0.8076781630516052,
888
+ "learning_rate": 2.323552596419889e-05,
889
+ "loss": 0.053,
890
+ "step": 119
891
+ },
892
+ {
893
+ "epoch": 1.5434083601286175,
894
+ "grad_norm": 1.6947088241577148,
895
+ "learning_rate": 2.181685175319702e-05,
896
+ "loss": 0.1237,
897
+ "step": 120
898
+ },
899
+ {
900
+ "epoch": 1.5434083601286175,
901
+ "eval_loss": 0.07654841244220734,
902
+ "eval_runtime": 6.4533,
903
+ "eval_samples_per_second": 25.413,
904
+ "eval_steps_per_second": 2.634,
905
+ "step": 120
906
  }
907
  ],
908
  "logging_steps": 1,
 
931
  "attributes": {}
932
  }
933
  },
934
+ "total_flos": 2.858372738187264e+16,
935
  "train_batch_size": 10,
936
  "trial_name": null,
937
  "trial_params": null