rootxhacker commited on
Commit
57fb636
·
verified ·
1 Parent(s): 54b497e

Training in progress, step 5500, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a0c1e03b37e306af29004a56bbc2e3cfe78ac4558d064ded4954032563a506c0
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce807f113eae8af3f76aad00110eac4aaa67725db83293608e360d0ae6f2138c
3
  size 36730224
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0d76b1ed26fb441fb296b2bc10e92b9546db047308ea5940e5ff2ef749dcfb42
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b62d615ea1686f1839d2072736549e7fb167c1df54c146af3bd0e3c76f28b705
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a263bfee7c334e6fc36b6c4750e79a70eeb042f841889c6eb0ea501e4fa6ac73
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:452f9d1449e365d91136837164af087a289325b1bf94b381d6eeb9d18ffa0513
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1bbcc820499d7de99bfade4a29f6e9516d1e9cb1c8de79befcf1f0d7f2a4da15
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f428dc4c0936653e8741f49c8b481e2cc2f0d2a1ce2182ef1df7e775a7d4db5
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7200a3e7cdec9309704535acc69d439ad04e127a63c3c13c5ca0a94f0236ea31
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b342bb583b996f802c3b66f99264c907dce5eb33ed829269a76eaa3a6c0e850c
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": 5000,
3
  "best_metric": 4.414160251617432,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-5000",
5
- "epoch": 0.3845858010922237,
6
  "eval_steps": 250,
7
- "global_step": 5000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -868,6 +868,92 @@
868
  "eval_samples_per_second": 53.604,
869
  "eval_steps_per_second": 13.401,
870
  "step": 5000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
871
  }
872
  ],
873
  "logging_steps": 50,
 
2
  "best_global_step": 5000,
3
  "best_metric": 4.414160251617432,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-5000",
5
+ "epoch": 0.42304438120144605,
6
  "eval_steps": 250,
7
+ "global_step": 5500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
868
  "eval_samples_per_second": 53.604,
869
  "eval_steps_per_second": 13.401,
870
  "step": 5000
871
+ },
872
+ {
873
+ "epoch": 0.3884316591031459,
874
+ "grad_norm": 2.658412456512451,
875
+ "learning_rate": 0.00011455387669570971,
876
+ "loss": 4.5453,
877
+ "step": 5050
878
+ },
879
+ {
880
+ "epoch": 0.39227751711406816,
881
+ "grad_norm": 2.231886148452759,
882
+ "learning_rate": 0.00011429653321569059,
883
+ "loss": 4.5524,
884
+ "step": 5100
885
+ },
886
+ {
887
+ "epoch": 0.3961233751249904,
888
+ "grad_norm": 4.202503204345703,
889
+ "learning_rate": 0.00011403918973567147,
890
+ "loss": 4.5274,
891
+ "step": 5150
892
+ },
893
+ {
894
+ "epoch": 0.39996923313591265,
895
+ "grad_norm": 2.8525800704956055,
896
+ "learning_rate": 0.00011378184625565236,
897
+ "loss": 4.5095,
898
+ "step": 5200
899
+ },
900
+ {
901
+ "epoch": 0.40381509114683484,
902
+ "grad_norm": 3.2517142295837402,
903
+ "learning_rate": 0.00011352964964523362,
904
+ "loss": 4.5043,
905
+ "step": 5250
906
+ },
907
+ {
908
+ "epoch": 0.40381509114683484,
909
+ "eval_loss": 4.595612525939941,
910
+ "eval_runtime": 18.9024,
911
+ "eval_samples_per_second": 52.903,
912
+ "eval_steps_per_second": 13.226,
913
+ "step": 5250
914
+ },
915
+ {
916
+ "epoch": 0.4076609491577571,
917
+ "grad_norm": 5.091184616088867,
918
+ "learning_rate": 0.00011327745303481488,
919
+ "loss": 4.4768,
920
+ "step": 5300
921
+ },
922
+ {
923
+ "epoch": 0.4115068071686793,
924
+ "grad_norm": 6.631587028503418,
925
+ "learning_rate": 0.00011302010955479578,
926
+ "loss": 4.5572,
927
+ "step": 5350
928
+ },
929
+ {
930
+ "epoch": 0.41535266517960157,
931
+ "grad_norm": 3.529118299484253,
932
+ "learning_rate": 0.00011276276607477666,
933
+ "loss": 4.6685,
934
+ "step": 5400
935
+ },
936
+ {
937
+ "epoch": 0.4191985231905238,
938
+ "grad_norm": 3.1017537117004395,
939
+ "learning_rate": 0.00011250542259475754,
940
+ "loss": 4.4271,
941
+ "step": 5450
942
+ },
943
+ {
944
+ "epoch": 0.42304438120144605,
945
+ "grad_norm": 3.930664300918579,
946
+ "learning_rate": 0.00011224807911473842,
947
+ "loss": 4.5501,
948
+ "step": 5500
949
+ },
950
+ {
951
+ "epoch": 0.42304438120144605,
952
+ "eval_loss": 4.486245632171631,
953
+ "eval_runtime": 18.9209,
954
+ "eval_samples_per_second": 52.851,
955
+ "eval_steps_per_second": 13.213,
956
+ "step": 5500
957
  }
958
  ],
959
  "logging_steps": 50,