rootxhacker commited on
Commit
548016f
·
verified ·
1 Parent(s): 85c2206

Training in progress, step 6000, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ce807f113eae8af3f76aad00110eac4aaa67725db83293608e360d0ae6f2138c
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a554f35b6a17be128fce6b6dde18077ae472e710857689bc82a3a107a5064b70
3
  size 36730224
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b62d615ea1686f1839d2072736549e7fb167c1df54c146af3bd0e3c76f28b705
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7489e1764b16dc1b56ae149e381b65054da7550c9735ed9c2fdc8fc794efba0e
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:452f9d1449e365d91136837164af087a289325b1bf94b381d6eeb9d18ffa0513
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:12c22be93200b6ea29a5a43fdc896fd16b0e2c374bfcbb275fb907bb85c40cd3
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1f428dc4c0936653e8741f49c8b481e2cc2f0d2a1ce2182ef1df7e775a7d4db5
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6914f08d038d2fd5583fe9d1dfb74316a740319f66641c1d86781bc3fc4b8381
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b342bb583b996f802c3b66f99264c907dce5eb33ed829269a76eaa3a6c0e850c
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f20cc39b1ac70cfd962157cd5b033ae7e76e9f02816bf84f00fd90791d769ab
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": 5000,
3
  "best_metric": 4.414160251617432,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-5000",
5
- "epoch": 0.42304438120144605,
6
  "eval_steps": 250,
7
- "global_step": 5500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -954,6 +954,92 @@
954
  "eval_samples_per_second": 52.851,
955
  "eval_steps_per_second": 13.213,
956
  "step": 5500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
957
  }
958
  ],
959
  "logging_steps": 50,
 
2
  "best_global_step": 5000,
3
  "best_metric": 4.414160251617432,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-5000",
5
+ "epoch": 0.4615029613106684,
6
  "eval_steps": 250,
7
+ "global_step": 6000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
954
  "eval_samples_per_second": 52.851,
955
  "eval_steps_per_second": 13.213,
956
  "step": 5500
957
+ },
958
+ {
959
+ "epoch": 0.4268902392123683,
960
+ "grad_norm": 4.470078945159912,
961
+ "learning_rate": 0.00011199073563471931,
962
+ "loss": 4.543,
963
+ "step": 5550
964
+ },
965
+ {
966
+ "epoch": 0.43073609722329054,
967
+ "grad_norm": 5.099395751953125,
968
+ "learning_rate": 0.00011173339215470019,
969
+ "loss": 4.4515,
970
+ "step": 5600
971
+ },
972
+ {
973
+ "epoch": 0.4345819552342128,
974
+ "grad_norm": 3.210951805114746,
975
+ "learning_rate": 0.00011147604867468107,
976
+ "loss": 4.4605,
977
+ "step": 5650
978
+ },
979
+ {
980
+ "epoch": 0.43842781324513497,
981
+ "grad_norm": 4.092874050140381,
982
+ "learning_rate": 0.00011121870519466196,
983
+ "loss": 4.6267,
984
+ "step": 5700
985
+ },
986
+ {
987
+ "epoch": 0.4422736712560572,
988
+ "grad_norm": 2.756460666656494,
989
+ "learning_rate": 0.00011096136171464283,
990
+ "loss": 4.4338,
991
+ "step": 5750
992
+ },
993
+ {
994
+ "epoch": 0.4422736712560572,
995
+ "eval_loss": 4.457804203033447,
996
+ "eval_runtime": 18.7914,
997
+ "eval_samples_per_second": 53.216,
998
+ "eval_steps_per_second": 13.304,
999
+ "step": 5750
1000
+ },
1001
+ {
1002
+ "epoch": 0.44611952926697945,
1003
+ "grad_norm": 5.140827178955078,
1004
+ "learning_rate": 0.00011070401823462372,
1005
+ "loss": 4.5102,
1006
+ "step": 5800
1007
+ },
1008
+ {
1009
+ "epoch": 0.4499653872779017,
1010
+ "grad_norm": 6.364997863769531,
1011
+ "learning_rate": 0.0001104466747546046,
1012
+ "loss": 4.5594,
1013
+ "step": 5850
1014
+ },
1015
+ {
1016
+ "epoch": 0.45381124528882394,
1017
+ "grad_norm": 5.3479695320129395,
1018
+ "learning_rate": 0.00011018933127458548,
1019
+ "loss": 4.4067,
1020
+ "step": 5900
1021
+ },
1022
+ {
1023
+ "epoch": 0.4576571032997462,
1024
+ "grad_norm": 3.728893518447876,
1025
+ "learning_rate": 0.00010993198779456637,
1026
+ "loss": 4.4689,
1027
+ "step": 5950
1028
+ },
1029
+ {
1030
+ "epoch": 0.4615029613106684,
1031
+ "grad_norm": 6.3881611824035645,
1032
+ "learning_rate": 0.00010967464431454724,
1033
+ "loss": 4.5641,
1034
+ "step": 6000
1035
+ },
1036
+ {
1037
+ "epoch": 0.4615029613106684,
1038
+ "eval_loss": 4.457447052001953,
1039
+ "eval_runtime": 18.8382,
1040
+ "eval_samples_per_second": 53.084,
1041
+ "eval_steps_per_second": 13.271,
1042
+ "step": 6000
1043
  }
1044
  ],
1045
  "logging_steps": 50,