rootxhacker commited on
Commit
303bcef
·
verified ·
1 Parent(s): ef73ff3

Training in progress, step 39000, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d9fccd6e8cf77b198a4f4b87f312750517824202a02166eb04e56577ecd2ba00
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aec93bf3af378bc69e279d3da1f22a2356cd072cc873629676a0de486c684213
3
  size 36730224
last-checkpoint/ar_diffusion_info.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:66d49ec8e5680b4ea0cac8d1930f19a010f9ae0bd4d5ca4b158ebc77b3d54acd
3
  size 1736
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1e17b4b57a9838b7215832af6318f536b3b6acee7c730845278b1a8f5d0cb17
3
  size 1736
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:38f23e9c9a2a443780a64b0cf18b93c2bd78a442ad08f1e37293ea92f7c88993
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e8f0bbc447de59f2033199d6477b7b4db91aeaa9c681dfd30f7ca1df276b9b1
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b1b4df89a65698a9335c43463fa8aa3cc49c1acdfae5f7a285af000ba78e95b1
3
- size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:60919181f100a38a6bc8dcdeb135786c172525306422517244d9c17e263e8748
3
+ size 14180
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1a1d56165e0890bb9432c217b765bdd79a8b4d93dcde80b4671874513824409b
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5cd25c48749969bd5bf972a3edb2b91874acf34f950f422e77399a6589ee1816
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f3dea87c40748698e240c93bc3eea5482b3f627ab3a1ed391f6cbe2fe5d7da89
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06d9f9352b3f9aa43ec304f953eb511fb6c42be93c910bdb749b28823ddd5892
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": 38250,
3
  "best_metric": -30.471132278442383,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-37500",
5
- "epoch": 2.961310668410122,
6
  "eval_steps": 250,
7
- "global_step": 38500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -6630,6 +6630,92 @@
6630
  "eval_samples_per_second": 59.287,
6631
  "eval_steps_per_second": 14.822,
6632
  "step": 38500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6633
  }
6634
  ],
6635
  "logging_steps": 50,
 
2
  "best_global_step": 38250,
3
  "best_metric": -30.471132278442383,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-37500",
5
+ "epoch": 2.999769248519345,
6
  "eval_steps": 250,
7
+ "global_step": 39000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
6630
  "eval_samples_per_second": 59.287,
6631
  "eval_steps_per_second": 14.822,
6632
  "step": 38500
6633
+ },
6634
+ {
6635
+ "epoch": 2.9651565264210444,
6636
+ "grad_norm": 0.5108075737953186,
6637
+ "learning_rate": 2.4361738046385997e-06,
6638
+ "loss": -30.467,
6639
+ "step": 38550
6640
+ },
6641
+ {
6642
+ "epoch": 2.969002384431967,
6643
+ "grad_norm": 0.608711838722229,
6644
+ "learning_rate": 2.1764537828221176e-06,
6645
+ "loss": -30.467,
6646
+ "step": 38600
6647
+ },
6648
+ {
6649
+ "epoch": 2.972848242442889,
6650
+ "grad_norm": 0.9723970890045166,
6651
+ "learning_rate": 1.916733761005636e-06,
6652
+ "loss": -30.4672,
6653
+ "step": 38650
6654
+ },
6655
+ {
6656
+ "epoch": 2.976694100453811,
6657
+ "grad_norm": 0.4361710846424103,
6658
+ "learning_rate": 1.6570137391891543e-06,
6659
+ "loss": -30.467,
6660
+ "step": 38700
6661
+ },
6662
+ {
6663
+ "epoch": 2.9805399584647336,
6664
+ "grad_norm": 1.1543878316879272,
6665
+ "learning_rate": 1.3972937173726722e-06,
6666
+ "loss": -30.4669,
6667
+ "step": 38750
6668
+ },
6669
+ {
6670
+ "epoch": 2.9805399584647336,
6671
+ "eval_loss": -30.47084617614746,
6672
+ "eval_runtime": 16.7842,
6673
+ "eval_samples_per_second": 59.58,
6674
+ "eval_steps_per_second": 14.895,
6675
+ "step": 38750
6676
+ },
6677
+ {
6678
+ "epoch": 2.9843858164756556,
6679
+ "grad_norm": 0.4299401640892029,
6680
+ "learning_rate": 1.1375736955561905e-06,
6681
+ "loss": -30.467,
6682
+ "step": 38800
6683
+ },
6684
+ {
6685
+ "epoch": 2.988231674486578,
6686
+ "grad_norm": 0.8476674556732178,
6687
+ "learning_rate": 8.778536737397087e-07,
6688
+ "loss": -30.467,
6689
+ "step": 38850
6690
+ },
6691
+ {
6692
+ "epoch": 2.9920775324975004,
6693
+ "grad_norm": 1.1194632053375244,
6694
+ "learning_rate": 6.181336519232268e-07,
6695
+ "loss": -30.4671,
6696
+ "step": 38900
6697
+ },
6698
+ {
6699
+ "epoch": 2.9959233905084224,
6700
+ "grad_norm": 0.7896573543548584,
6701
+ "learning_rate": 3.584136301067449e-07,
6702
+ "loss": -30.4671,
6703
+ "step": 38950
6704
+ },
6705
+ {
6706
+ "epoch": 2.999769248519345,
6707
+ "grad_norm": 0.32893452048301697,
6708
+ "learning_rate": 9.869360829026311e-08,
6709
+ "loss": -30.4671,
6710
+ "step": 39000
6711
+ },
6712
+ {
6713
+ "epoch": 2.999769248519345,
6714
+ "eval_loss": -30.471084594726562,
6715
+ "eval_runtime": 16.7188,
6716
+ "eval_samples_per_second": 59.813,
6717
+ "eval_steps_per_second": 14.953,
6718
+ "step": 39000
6719
  }
6720
  ],
6721
  "logging_steps": 50,