rootxhacker commited on
Commit
610d128
·
verified ·
1 Parent(s): 28486e6

Training in progress, step 11000, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ac66d1bed32cb77a6fe046d97ff60fe2cfaef95ddf73a9bf13a7ff56a55c11af
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78684fe66423fcc71792a533cdf4a615150a5d7719cd88fbc583e2cf7207b1f2
3
  size 36730224
last-checkpoint/ar_diffusion_info.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:298a32675b40f08695b3fb4fa7bbd261866561afa1edbe21444c5e1018c7c6df
3
  size 1736
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d827d0c5de1fe1d4d42261146337ddfecdb8784573ad919f9058a14e73074cb
3
  size 1736
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:df829391bae99716149a25d0f76e0e75e0e3329d3b7ef7d8f960f81bf3520101
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b118dfa8e1001390c89e0eac0d61d9f1d84255e6c68db2d2cc7fd7107e67ae24
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2286056e584e728d67ca2ee5e9e130931cc603f814dc38d9957d5e8ae618fb75
3
- size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b1422ca832a4908b348152649eb6c40e62866db2c96015a75fe3a4ed0effd72
3
+ size 14180
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:64e34184efeb9f5a5622c883abc37387f021455b8ebc365e06c304057c5bb8b6
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b71aebdeeb32cab8c73c6ce65908475546106933ecd5ef49363bd4d3f77b4ba2
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:83df1333fe8f59c1710a91e769544177d60bc5aa48c1018ddd6a061172e2a46b
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1d6a033c9c4c6cee23163acb32007f6f866e5b846f8755100cbc84bc6fe976e
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 10500,
3
- "best_metric": 1.3613618612289429,
4
- "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-10500",
5
- "epoch": 0.8076301822936697,
6
  "eval_steps": 250,
7
- "global_step": 10500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -1814,6 +1814,92 @@
1814
  "eval_samples_per_second": 59.523,
1815
  "eval_steps_per_second": 14.881,
1816
  "step": 10500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1817
  }
1818
  ],
1819
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 11000,
3
+ "best_metric": 1.3456926345825195,
4
+ "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-11000",
5
+ "epoch": 0.8460887624028921,
6
  "eval_steps": 250,
7
+ "global_step": 11000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
1814
  "eval_samples_per_second": 59.523,
1815
  "eval_steps_per_second": 14.881,
1816
  "step": 10500
1817
+ },
1818
+ {
1819
+ "epoch": 0.811476040304592,
1820
+ "grad_norm": 1.2509782314300537,
1821
+ "learning_rate": 0.00014782744201750512,
1822
+ "loss": 1.4795,
1823
+ "step": 10550
1824
+ },
1825
+ {
1826
+ "epoch": 0.8153218983155142,
1827
+ "grad_norm": 1.7617859840393066,
1828
+ "learning_rate": 0.00014756772199568864,
1829
+ "loss": 1.4244,
1830
+ "step": 10600
1831
+ },
1832
+ {
1833
+ "epoch": 0.8191677563264365,
1834
+ "grad_norm": 1.4050170183181763,
1835
+ "learning_rate": 0.00014730800197387218,
1836
+ "loss": 1.3426,
1837
+ "step": 10650
1838
+ },
1839
+ {
1840
+ "epoch": 0.8230136143373586,
1841
+ "grad_norm": 1.3139458894729614,
1842
+ "learning_rate": 0.0001470482819520557,
1843
+ "loss": 1.4864,
1844
+ "step": 10700
1845
+ },
1846
+ {
1847
+ "epoch": 0.8268594723482809,
1848
+ "grad_norm": 1.7595279216766357,
1849
+ "learning_rate": 0.0001467885619302392,
1850
+ "loss": 1.3767,
1851
+ "step": 10750
1852
+ },
1853
+ {
1854
+ "epoch": 0.8268594723482809,
1855
+ "eval_loss": 1.35465407371521,
1856
+ "eval_runtime": 16.8836,
1857
+ "eval_samples_per_second": 59.229,
1858
+ "eval_steps_per_second": 14.807,
1859
+ "step": 10750
1860
+ },
1861
+ {
1862
+ "epoch": 0.8307053303592031,
1863
+ "grad_norm": 1.3413208723068237,
1864
+ "learning_rate": 0.00014652884190842274,
1865
+ "loss": 1.3273,
1866
+ "step": 10800
1867
+ },
1868
+ {
1869
+ "epoch": 0.8345511883701254,
1870
+ "grad_norm": 1.6046358346939087,
1871
+ "learning_rate": 0.00014626912188660625,
1872
+ "loss": 1.3967,
1873
+ "step": 10850
1874
+ },
1875
+ {
1876
+ "epoch": 0.8383970463810476,
1877
+ "grad_norm": 1.2793117761611938,
1878
+ "learning_rate": 0.00014600940186478977,
1879
+ "loss": 1.4168,
1880
+ "step": 10900
1881
+ },
1882
+ {
1883
+ "epoch": 0.8422429043919698,
1884
+ "grad_norm": 1.1738851070404053,
1885
+ "learning_rate": 0.00014574968184297328,
1886
+ "loss": 1.3577,
1887
+ "step": 10950
1888
+ },
1889
+ {
1890
+ "epoch": 0.8460887624028921,
1891
+ "grad_norm": 1.3740490674972534,
1892
+ "learning_rate": 0.0001454899618211568,
1893
+ "loss": 1.3609,
1894
+ "step": 11000
1895
+ },
1896
+ {
1897
+ "epoch": 0.8460887624028921,
1898
+ "eval_loss": 1.3456926345825195,
1899
+ "eval_runtime": 16.7867,
1900
+ "eval_samples_per_second": 59.571,
1901
+ "eval_steps_per_second": 14.893,
1902
+ "step": 11000
1903
  }
1904
  ],
1905
  "logging_steps": 50,