rootxhacker commited on
Commit
633a220
·
verified ·
1 Parent(s): 2fd3151

Training in progress, step 10500, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eb0e3c8eeaeaaeed92b7974a796b8555d972ab04a230ed8a0aa39040f919522e
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac66d1bed32cb77a6fe046d97ff60fe2cfaef95ddf73a9bf13a7ff56a55c11af
3
  size 36730224
last-checkpoint/ar_diffusion_info.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dbc9386aa433cf09ea700ba9554d99ebbf74421553fda8c6e9f15d579ffea58d
3
  size 1736
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:298a32675b40f08695b3fb4fa7bbd261866561afa1edbe21444c5e1018c7c6df
3
  size 1736
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:845a998395b0eed3947658b954a32b49a265bc7bcb21586d38b38bbd5b7b0ca3
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df829391bae99716149a25d0f76e0e75e0e3329d3b7ef7d8f960f81bf3520101
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:45788da03bae97e95c3537137dc4032cafd5cea57659195759c4661fda86a92e
3
- size 14180
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2286056e584e728d67ca2ee5e9e130931cc603f814dc38d9957d5e8ae618fb75
3
+ size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:59deba323ff65998dc2bdf22d9e342db9cffc30bd0e70e0e297b662f8184f414
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64e34184efeb9f5a5622c883abc37387f021455b8ebc365e06c304057c5bb8b6
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ab5aa9661d156b69bcd5976a8b5f6e5eaf3f91f473a83c47e1ba991c358f74ef
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83df1333fe8f59c1710a91e769544177d60bc5aa48c1018ddd6a061172e2a46b
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 10000,
3
- "best_metric": 1.36810302734375,
4
- "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-10000",
5
- "epoch": 0.7691716021844474,
6
  "eval_steps": 250,
7
- "global_step": 10000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -1728,6 +1728,92 @@
1728
  "eval_samples_per_second": 59.387,
1729
  "eval_steps_per_second": 14.847,
1730
  "step": 10000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1731
  }
1732
  ],
1733
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 10500,
3
+ "best_metric": 1.3613618612289429,
4
+ "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-10500",
5
+ "epoch": 0.8076301822936697,
6
  "eval_steps": 250,
7
+ "global_step": 10500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
1728
  "eval_samples_per_second": 59.387,
1729
  "eval_steps_per_second": 14.847,
1730
  "step": 10000
1731
+ },
1732
+ {
1733
+ "epoch": 0.7730174601953695,
1734
+ "grad_norm": 1.1201566457748413,
1735
+ "learning_rate": 0.00015042464223566997,
1736
+ "loss": 1.4327,
1737
+ "step": 10050
1738
+ },
1739
+ {
1740
+ "epoch": 0.7768633182062918,
1741
+ "grad_norm": 1.7418923377990723,
1742
+ "learning_rate": 0.00015016492221385348,
1743
+ "loss": 1.2776,
1744
+ "step": 10100
1745
+ },
1746
+ {
1747
+ "epoch": 0.780709176217214,
1748
+ "grad_norm": 1.3154345750808716,
1749
+ "learning_rate": 0.000149905202192037,
1750
+ "loss": 1.4374,
1751
+ "step": 10150
1752
+ },
1753
+ {
1754
+ "epoch": 0.7845550342281363,
1755
+ "grad_norm": 1.129528284072876,
1756
+ "learning_rate": 0.0001496454821702205,
1757
+ "loss": 1.3404,
1758
+ "step": 10200
1759
+ },
1760
+ {
1761
+ "epoch": 0.7884008922390585,
1762
+ "grad_norm": 1.3940690755844116,
1763
+ "learning_rate": 0.00014938576214840402,
1764
+ "loss": 1.3828,
1765
+ "step": 10250
1766
+ },
1767
+ {
1768
+ "epoch": 0.7884008922390585,
1769
+ "eval_loss": 1.3682047128677368,
1770
+ "eval_runtime": 17.0099,
1771
+ "eval_samples_per_second": 58.789,
1772
+ "eval_steps_per_second": 14.697,
1773
+ "step": 10250
1774
+ },
1775
+ {
1776
+ "epoch": 0.7922467502499808,
1777
+ "grad_norm": 1.3081247806549072,
1778
+ "learning_rate": 0.00014912604212658753,
1779
+ "loss": 1.4071,
1780
+ "step": 10300
1781
+ },
1782
+ {
1783
+ "epoch": 0.796092608260903,
1784
+ "grad_norm": 0.8190014362335205,
1785
+ "learning_rate": 0.00014886632210477107,
1786
+ "loss": 1.4668,
1787
+ "step": 10350
1788
+ },
1789
+ {
1790
+ "epoch": 0.7999384662718253,
1791
+ "grad_norm": 1.8602573871612549,
1792
+ "learning_rate": 0.00014860660208295459,
1793
+ "loss": 1.3456,
1794
+ "step": 10400
1795
+ },
1796
+ {
1797
+ "epoch": 0.8037843242827475,
1798
+ "grad_norm": 1.4184271097183228,
1799
+ "learning_rate": 0.0001483468820611381,
1800
+ "loss": 1.3851,
1801
+ "step": 10450
1802
+ },
1803
+ {
1804
+ "epoch": 0.8076301822936697,
1805
+ "grad_norm": 1.5128880739212036,
1806
+ "learning_rate": 0.00014808716203932164,
1807
+ "loss": 1.3869,
1808
+ "step": 10500
1809
+ },
1810
+ {
1811
+ "epoch": 0.8076301822936697,
1812
+ "eval_loss": 1.3613618612289429,
1813
+ "eval_runtime": 16.8001,
1814
+ "eval_samples_per_second": 59.523,
1815
+ "eval_steps_per_second": 14.881,
1816
+ "step": 10500
1817
  }
1818
  ],
1819
  "logging_steps": 50,