texanrangee commited on
Commit
29df206
·
verified ·
1 Parent(s): d75fc4d

Training in progress, step 1050, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7ff8c43c5d0a70fdd3c5a092f45b576ce4d6695039dd5efa0b3372bf9de2b4bc
3
  size 1302756720
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0e95a395e4b276d4c7eacb38473e906143acd581ad5443468b2f990e935ddfa
3
  size 1302756720
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f88ef668a708a27bef57939446f0bd0a3ccfb9b33174df84a17271afc84ba685
3
  size 2565803322
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a2654c6ae5ddd64e2c312c0b2a422dfd011eb63a7c92d74501c71b32ea30675
3
  size 2565803322
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4252aa9c1b46b4624b57d2f922441631bdb1287ae4a706e82d33f1ad66a031db
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:425026aaf9fa84dc0943287ca090b73e61643088d53bfd5fc139a1654ad51203
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:60a827078c85f0fd366049a88039c9056197c970413749f5c2cbfe85bd3bd7d6
3
  size 1256
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9cf334b45c59ec3117af6235b0cb6b8da2aab8bbc78c388edc1a1925c6731983
3
  size 1256
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 1.324646234512329,
3
  "best_model_checkpoint": "./output/checkpoint-150",
4
- "epoch": 0.027973766823112547,
5
  "eval_steps": 150,
6
- "global_step": 900,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -685,6 +685,119 @@
685
  "eval_samples_per_second": 11.895,
686
  "eval_steps_per_second": 11.895,
687
  "step": 900
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
688
  }
689
  ],
690
  "logging_steps": 10,
@@ -704,7 +817,7 @@
704
  "attributes": {}
705
  }
706
  },
707
- "total_flos": 8.218529161795584e+16,
708
  "train_batch_size": 4,
709
  "trial_name": null,
710
  "trial_params": null
 
1
  {
2
  "best_metric": 1.324646234512329,
3
  "best_model_checkpoint": "./output/checkpoint-150",
4
+ "epoch": 0.03263606129363131,
5
  "eval_steps": 150,
6
+ "global_step": 1050,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
685
  "eval_samples_per_second": 11.895,
686
  "eval_steps_per_second": 11.895,
687
  "step": 900
688
+ },
689
+ {
690
+ "epoch": 0.028284586454480464,
691
+ "grad_norm": 29.72920036315918,
692
+ "learning_rate": 9.340773121260896e-05,
693
+ "loss": 1.1198,
694
+ "step": 910
695
+ },
696
+ {
697
+ "epoch": 0.02859540608584838,
698
+ "grad_norm": 26.55568504333496,
699
+ "learning_rate": 9.32477428972347e-05,
700
+ "loss": 1.206,
701
+ "step": 920
702
+ },
703
+ {
704
+ "epoch": 0.0289062257172163,
705
+ "grad_norm": 17.740917205810547,
706
+ "learning_rate": 9.308597683653978e-05,
707
+ "loss": 1.2149,
708
+ "step": 930
709
+ },
710
+ {
711
+ "epoch": 0.029217045348584218,
712
+ "grad_norm": 16.225730895996094,
713
+ "learning_rate": 9.292243968009333e-05,
714
+ "loss": 1.1254,
715
+ "step": 940
716
+ },
717
+ {
718
+ "epoch": 0.029527864979952134,
719
+ "grad_norm": 29.286996841430664,
720
+ "learning_rate": 9.275713815026734e-05,
721
+ "loss": 1.3235,
722
+ "step": 950
723
+ },
724
+ {
725
+ "epoch": 0.02983868461132005,
726
+ "grad_norm": 22.0376033782959,
727
+ "learning_rate": 9.259007904196024e-05,
728
+ "loss": 1.4885,
729
+ "step": 960
730
+ },
731
+ {
732
+ "epoch": 0.030149504242687968,
733
+ "grad_norm": 26.638015747070312,
734
+ "learning_rate": 9.242126922231766e-05,
735
+ "loss": 1.2483,
736
+ "step": 970
737
+ },
738
+ {
739
+ "epoch": 0.030460323874055884,
740
+ "grad_norm": 12.953681945800781,
741
+ "learning_rate": 9.225071563045009e-05,
742
+ "loss": 1.5445,
743
+ "step": 980
744
+ },
745
+ {
746
+ "epoch": 0.0307711435054238,
747
+ "grad_norm": 19.948116302490234,
748
+ "learning_rate": 9.207842527714768e-05,
749
+ "loss": 1.2768,
750
+ "step": 990
751
+ },
752
+ {
753
+ "epoch": 0.03108196313679172,
754
+ "grad_norm": 16.087295532226562,
755
+ "learning_rate": 9.190440524459205e-05,
756
+ "loss": 1.0367,
757
+ "step": 1000
758
+ },
759
+ {
760
+ "epoch": 0.031392782768159634,
761
+ "grad_norm": 20.46804428100586,
762
+ "learning_rate": 9.172866268606516e-05,
763
+ "loss": 1.1743,
764
+ "step": 1010
765
+ },
766
+ {
767
+ "epoch": 0.031703602399527554,
768
+ "grad_norm": 10.569546699523926,
769
+ "learning_rate": 9.155120482565522e-05,
770
+ "loss": 1.2312,
771
+ "step": 1020
772
+ },
773
+ {
774
+ "epoch": 0.032014422030895474,
775
+ "grad_norm": 23.994094848632812,
776
+ "learning_rate": 9.137203895795986e-05,
777
+ "loss": 1.2055,
778
+ "step": 1030
779
+ },
780
+ {
781
+ "epoch": 0.03232524166226339,
782
+ "grad_norm": 18.373334884643555,
783
+ "learning_rate": 9.11911724477861e-05,
784
+ "loss": 1.4074,
785
+ "step": 1040
786
+ },
787
+ {
788
+ "epoch": 0.03263606129363131,
789
+ "grad_norm": 18.44525718688965,
790
+ "learning_rate": 9.100861272984782e-05,
791
+ "loss": 1.1465,
792
+ "step": 1050
793
+ },
794
+ {
795
+ "epoch": 0.03263606129363131,
796
+ "eval_loss": 1.3763073682785034,
797
+ "eval_runtime": 44.4721,
798
+ "eval_samples_per_second": 11.243,
799
+ "eval_steps_per_second": 11.243,
800
+ "step": 1050
801
  }
802
  ],
803
  "logging_steps": 10,
 
817
  "attributes": {}
818
  }
819
  },
820
+ "total_flos": 9.590225761995571e+16,
821
  "train_batch_size": 4,
822
  "trial_name": null,
823
  "trial_params": null