rootxhacker commited on
Commit
9bcb30a
·
verified ·
1 Parent(s): 8408bc3

Training in progress, step 4500, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b445c42b37f9090e63ec9e650d67af120bdec1120aa2b99d075b10217d2f4041
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2da0677fa12d4198b9a2aefbc499f61536ce08274e2b77ec9742768adbc966b
3
  size 36730224
last-checkpoint/ar_diffusion_info.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4ae940e2a8fce62cfe8e7b2911818f4f043d9298216ee2524750b9c2bb4195bb
3
  size 1736
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d5cb22e083fb9b6ead9a6b9754983dfef6d4fd0258d21fc899979d4540b8ab1
3
  size 1736
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dacaa8970d573f7be27d82942d32e193ea6f961af3a7a55fdd87f418730a6ce8
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c840a4d894a9e893236ba3eb39bb9d761fb13f08794f23f4bb8cb0fa462367c
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1698c9cb87c2e75a65bbfdd5b5bac56ca042377422d15083ff3eeeb34095ca85
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:770b1304c731c8aa940b69b50f7998667dba8e089ec052ddb586fa531ecef79d
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1b33c54ce26acce0407557fdf2626c004064a12b74672d7c82d6602cfe737b3d
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcd3dca35f0088299a2a2e0ac814a4cd8b9d9681c16fef5ada156579adebbcb1
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cd43e4afdd546fe9e425b45803ea4071bedf22fd593f1e2dc9ea450f413acbe3
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d1c26a223e581bb15f1e95b6c820cdf95e07905cfe02588473944a7fc241ee1
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 4000,
3
- "best_metric": 1.5196877717971802,
4
- "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-4000",
5
- "epoch": 0.30766864087377893,
6
  "eval_steps": 250,
7
- "global_step": 4000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -696,6 +696,92 @@
696
  "eval_samples_per_second": 59.695,
697
  "eval_steps_per_second": 14.924,
698
  "step": 4000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
699
  }
700
  ],
701
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 4500,
3
+ "best_metric": 1.4980955123901367,
4
+ "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-4500",
5
+ "epoch": 0.3461272209830013,
6
  "eval_steps": 250,
7
+ "global_step": 4500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
696
  "eval_samples_per_second": 59.695,
697
  "eval_steps_per_second": 14.924,
698
  "step": 4000
699
+ },
700
+ {
701
+ "epoch": 0.3115144988847012,
702
+ "grad_norm": 0.9612170457839966,
703
+ "learning_rate": 0.00018158585045321144,
704
+ "loss": 1.5134,
705
+ "step": 4050
706
+ },
707
+ {
708
+ "epoch": 0.3153603568956234,
709
+ "grad_norm": 1.4119675159454346,
710
+ "learning_rate": 0.00018132613043139498,
711
+ "loss": 1.5019,
712
+ "step": 4100
713
+ },
714
+ {
715
+ "epoch": 0.31920621490654566,
716
+ "grad_norm": 1.5474581718444824,
717
+ "learning_rate": 0.0001810664104095785,
718
+ "loss": 1.5243,
719
+ "step": 4150
720
+ },
721
+ {
722
+ "epoch": 0.3230520729174679,
723
+ "grad_norm": 1.5670969486236572,
724
+ "learning_rate": 0.000180806690387762,
725
+ "loss": 1.5482,
726
+ "step": 4200
727
+ },
728
+ {
729
+ "epoch": 0.32689793092839015,
730
+ "grad_norm": 1.5624980926513672,
731
+ "learning_rate": 0.00018054697036594552,
732
+ "loss": 1.5429,
733
+ "step": 4250
734
+ },
735
+ {
736
+ "epoch": 0.32689793092839015,
737
+ "eval_loss": 1.5110443830490112,
738
+ "eval_runtime": 17.0856,
739
+ "eval_samples_per_second": 58.529,
740
+ "eval_steps_per_second": 14.632,
741
+ "step": 4250
742
+ },
743
+ {
744
+ "epoch": 0.33074378893931233,
745
+ "grad_norm": 1.0080711841583252,
746
+ "learning_rate": 0.00018028725034412903,
747
+ "loss": 1.5022,
748
+ "step": 4300
749
+ },
750
+ {
751
+ "epoch": 0.3345896469502346,
752
+ "grad_norm": 0.9390202760696411,
753
+ "learning_rate": 0.00018002753032231257,
754
+ "loss": 1.482,
755
+ "step": 4350
756
+ },
757
+ {
758
+ "epoch": 0.3384355049611568,
759
+ "grad_norm": 1.0671314001083374,
760
+ "learning_rate": 0.00017976781030049608,
761
+ "loss": 1.5271,
762
+ "step": 4400
763
+ },
764
+ {
765
+ "epoch": 0.34228136297207906,
766
+ "grad_norm": 1.290907382965088,
767
+ "learning_rate": 0.0001795080902786796,
768
+ "loss": 1.4667,
769
+ "step": 4450
770
+ },
771
+ {
772
+ "epoch": 0.3461272209830013,
773
+ "grad_norm": 0.6162307858467102,
774
+ "learning_rate": 0.0001792483702568631,
775
+ "loss": 1.4435,
776
+ "step": 4500
777
+ },
778
+ {
779
+ "epoch": 0.3461272209830013,
780
+ "eval_loss": 1.4980955123901367,
781
+ "eval_runtime": 16.8193,
782
+ "eval_samples_per_second": 59.456,
783
+ "eval_steps_per_second": 14.864,
784
+ "step": 4500
785
  }
786
  ],
787
  "logging_steps": 50,