rootxhacker commited on
Commit
d6fa16f
·
verified ·
1 Parent(s): b414d5d

Training in progress, step 11000, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7bbfabbed60d5906524d5c89409a9d91e99af768406b59d0122cc4c00803b88b
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:932444defba9a5b8dd44f88b8bcb52ab659664a28676382e15a373cf36bf6034
3
  size 36730224
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c1f937a69fd1d3f87ca4cb29a7d82a362b45d23ba437525393801a63c53b09d6
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f80012b6836a364fdec2135af886ce664e83e4046832b07a32b6501d8f4ef656
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:371a8ee92df296caadab094d1243e6f6a8e33531055d940b9b2beb58c3296d37
3
- size 14180
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d7167878bddb7c0a88c66ad0b0be73885dac80e1980348d5e68d660c4a16ce9
3
+ size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5bd55c94e5f9a6b24aa6ffb24230950971ac714f0b527005b89121745edab7aa
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42adf2998e15e06307e725ba7dcb6c52fc0a94a297e1345e89da412c057a53a6
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e3263d434a0173d1bf88ad28980a5eff0fc4c115a390bc663e18dd956338ac6e
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f034f97f7b7aead49a2646d657b2abae10ec15daef82d04e9386fb7172dbeddc
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 9500,
3
- "best_metric": 4.328299045562744,
4
- "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-9500",
5
- "epoch": 0.8076301822936697,
6
  "eval_steps": 250,
7
- "global_step": 10500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -1814,6 +1814,92 @@
1814
  "eval_samples_per_second": 52.877,
1815
  "eval_steps_per_second": 13.219,
1816
  "step": 10500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1817
  }
1818
  ],
1819
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 11000,
3
+ "best_metric": 4.307990074157715,
4
+ "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-11000",
5
+ "epoch": 0.8460887624028921,
6
  "eval_steps": 250,
7
+ "global_step": 11000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
1814
  "eval_samples_per_second": 52.877,
1815
  "eval_steps_per_second": 13.219,
1816
  "step": 10500
1817
+ },
1818
+ {
1819
+ "epoch": 0.811476040304592,
1820
+ "grad_norm": 4.561670303344727,
1821
+ "learning_rate": 8.626668137200838e-05,
1822
+ "loss": 4.6428,
1823
+ "step": 10550
1824
+ },
1825
+ {
1826
+ "epoch": 0.8153218983155142,
1827
+ "grad_norm": 3.155518054962158,
1828
+ "learning_rate": 8.600933789198925e-05,
1829
+ "loss": 4.4605,
1830
+ "step": 10600
1831
+ },
1832
+ {
1833
+ "epoch": 0.8191677563264365,
1834
+ "grad_norm": 4.021768093109131,
1835
+ "learning_rate": 8.575199441197014e-05,
1836
+ "loss": 4.2982,
1837
+ "step": 10650
1838
+ },
1839
+ {
1840
+ "epoch": 0.8230136143373586,
1841
+ "grad_norm": 4.348796844482422,
1842
+ "learning_rate": 8.549465093195102e-05,
1843
+ "loss": 4.649,
1844
+ "step": 10700
1845
+ },
1846
+ {
1847
+ "epoch": 0.8268594723482809,
1848
+ "grad_norm": 4.647562503814697,
1849
+ "learning_rate": 8.523730745193191e-05,
1850
+ "loss": 4.2873,
1851
+ "step": 10750
1852
+ },
1853
+ {
1854
+ "epoch": 0.8268594723482809,
1855
+ "eval_loss": 4.3483662605285645,
1856
+ "eval_runtime": 18.9227,
1857
+ "eval_samples_per_second": 52.847,
1858
+ "eval_steps_per_second": 13.212,
1859
+ "step": 10750
1860
+ },
1861
+ {
1862
+ "epoch": 0.8307053303592031,
1863
+ "grad_norm": 3.9260427951812744,
1864
+ "learning_rate": 8.497996397191279e-05,
1865
+ "loss": 4.3823,
1866
+ "step": 10800
1867
+ },
1868
+ {
1869
+ "epoch": 0.8345511883701254,
1870
+ "grad_norm": 3.7108564376831055,
1871
+ "learning_rate": 8.472262049189368e-05,
1872
+ "loss": 4.42,
1873
+ "step": 10850
1874
+ },
1875
+ {
1876
+ "epoch": 0.8383970463810476,
1877
+ "grad_norm": 4.9123663902282715,
1878
+ "learning_rate": 8.446527701187455e-05,
1879
+ "loss": 4.5828,
1880
+ "step": 10900
1881
+ },
1882
+ {
1883
+ "epoch": 0.8422429043919698,
1884
+ "grad_norm": 3.7289183139801025,
1885
+ "learning_rate": 8.420793353185543e-05,
1886
+ "loss": 4.3134,
1887
+ "step": 10950
1888
+ },
1889
+ {
1890
+ "epoch": 0.8460887624028921,
1891
+ "grad_norm": 4.0350542068481445,
1892
+ "learning_rate": 8.395059005183632e-05,
1893
+ "loss": 4.3768,
1894
+ "step": 11000
1895
+ },
1896
+ {
1897
+ "epoch": 0.8460887624028921,
1898
+ "eval_loss": 4.307990074157715,
1899
+ "eval_runtime": 18.7713,
1900
+ "eval_samples_per_second": 53.273,
1901
+ "eval_steps_per_second": 13.318,
1902
+ "step": 11000
1903
  }
1904
  ],
1905
  "logging_steps": 50,