rootxhacker commited on
Commit
cb8921c
·
verified ·
1 Parent(s): ba81eb2

Training in progress, step 5500, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ccc114460bdc62c626ba50c920e272e7f575d8aeb0e2543e39fc6ee05fa17062
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4eb8910307bec61db122da2594e6b3f530d868dc95f89d2ca698471ef1bf7e40
3
  size 36730224
last-checkpoint/ar_diffusion_info.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:50d17eabb10066129b6714ff713807dcb69661d9ef6abc8937da238409918435
3
  size 1736
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f7c3bd2e2b53a0436de47612a1d9d71e57d2e478df8b52acc7aec65dc000186
3
  size 1736
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f41eb4686c9fcfd89bdeef4e240aa2cba7d92d7862f3e09543657417b8a31b1f
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42be51fbf276340c9331b95a10d361ad5f54ac5fddd8d3003ab5a2b3205b1bf3
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b3f90304776c4344263775ad85fa55a101daf353f68d13a5872b9fc01a14e2b3
3
- size 14180
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:edba3e2e11b49613ce2a377ec116d91655a77025fcd35863ca859fcc236283bd
3
+ size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:429536711957c0df9399d6c91b39d43aa4faf1c40a0c6bae8287d373fb9daa66
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee231d5124e1faa72bb59e0d3351580448c9875eebe2a58142e26f5eeab5354c
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:17a4cf673b793da03b6c2fcc104b7095156a3747182923bcd50f7b981d959c40
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1402f998a531b2d9e639df3a0d0bab3a51ee695db29df819edf5370824d49f4
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 5000,
3
- "best_metric": 1.4847265481948853,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-5000",
5
- "epoch": 0.3845858010922237,
6
  "eval_steps": 250,
7
- "global_step": 5000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -868,6 +868,92 @@
868
  "eval_samples_per_second": 59.163,
869
  "eval_steps_per_second": 14.791,
870
  "step": 5000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
871
  }
872
  ],
873
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 5250,
3
+ "best_metric": 1.4721767902374268,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-5000",
5
+ "epoch": 0.42304438120144605,
6
  "eval_steps": 250,
7
+ "global_step": 5500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
868
  "eval_samples_per_second": 59.163,
869
  "eval_steps_per_second": 14.791,
870
  "step": 5000
871
+ },
872
+ {
873
+ "epoch": 0.3884316591031459,
874
+ "grad_norm": 1.0815047025680542,
875
+ "learning_rate": 0.00017639145001688183,
876
+ "loss": 1.5059,
877
+ "step": 5050
878
+ },
879
+ {
880
+ "epoch": 0.39227751711406816,
881
+ "grad_norm": 0.8886498808860779,
882
+ "learning_rate": 0.00017613172999506534,
883
+ "loss": 1.5056,
884
+ "step": 5100
885
+ },
886
+ {
887
+ "epoch": 0.3961233751249904,
888
+ "grad_norm": 1.647912859916687,
889
+ "learning_rate": 0.00017587200997324883,
890
+ "loss": 1.4471,
891
+ "step": 5150
892
+ },
893
+ {
894
+ "epoch": 0.39996923313591265,
895
+ "grad_norm": 1.0519336462020874,
896
+ "learning_rate": 0.00017561228995143237,
897
+ "loss": 1.5272,
898
+ "step": 5200
899
+ },
900
+ {
901
+ "epoch": 0.40381509114683484,
902
+ "grad_norm": 0.7143021821975708,
903
+ "learning_rate": 0.00017535256992961588,
904
+ "loss": 1.5315,
905
+ "step": 5250
906
+ },
907
+ {
908
+ "epoch": 0.40381509114683484,
909
+ "eval_loss": 1.4721767902374268,
910
+ "eval_runtime": 16.9221,
911
+ "eval_samples_per_second": 59.094,
912
+ "eval_steps_per_second": 14.774,
913
+ "step": 5250
914
+ },
915
+ {
916
+ "epoch": 0.4076609491577571,
917
+ "grad_norm": 0.9163336157798767,
918
+ "learning_rate": 0.0001750928499077994,
919
+ "loss": 1.4824,
920
+ "step": 5300
921
+ },
922
+ {
923
+ "epoch": 0.4115068071686793,
924
+ "grad_norm": 1.6825785636901855,
925
+ "learning_rate": 0.00017483312988598293,
926
+ "loss": 1.5034,
927
+ "step": 5350
928
+ },
929
+ {
930
+ "epoch": 0.41535266517960157,
931
+ "grad_norm": 1.1994304656982422,
932
+ "learning_rate": 0.00017457340986416645,
933
+ "loss": 1.4712,
934
+ "step": 5400
935
+ },
936
+ {
937
+ "epoch": 0.4191985231905238,
938
+ "grad_norm": 2.0379066467285156,
939
+ "learning_rate": 0.00017431368984234996,
940
+ "loss": 1.409,
941
+ "step": 5450
942
+ },
943
+ {
944
+ "epoch": 0.42304438120144605,
945
+ "grad_norm": 1.5934576988220215,
946
+ "learning_rate": 0.00017405396982053347,
947
+ "loss": 1.4791,
948
+ "step": 5500
949
+ },
950
+ {
951
+ "epoch": 0.42304438120144605,
952
+ "eval_loss": 1.4770058393478394,
953
+ "eval_runtime": 16.8532,
954
+ "eval_samples_per_second": 59.336,
955
+ "eval_steps_per_second": 14.834,
956
+ "step": 5500
957
  }
958
  ],
959
  "logging_steps": 50,