rootxhacker commited on
Commit
f96435a
·
verified ·
1 Parent(s): 2b0535e

Training in progress, step 11000, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2355e3e82295bb14e617f48ba9a53dcaecdce82329ed609cc4929ae9d83b250b
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:81cf0e3dd4d992ca19a0a903b87889cb1dcb0561835bd1a05ad25d4805a11c3a
3
  size 36730224
last-checkpoint/ar_diffusion_info.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2e45f55b2e0437c4eb0d6bdb520b7b1c7edbedaf9253b60e08be085f388b8113
3
  size 1736
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24383c9348cb96ba012340aa8a76698ba299b93dfbd362d850260ceb6b5bff92
3
  size 1736
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:90e2bba69ad980b5be837dce2dab49768cfe98d25981ae562b11cdccf65223e4
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:714a435b99357bfd46a0b087ff628bd01321e6a7b7b84db498111b7fbf583143
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0766963e075955373626125e0506ee76918124a59001be5cd204b364faad380e
3
- size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43f28196579619a7a1a9fd513d6b01be77d05b370f16ae231d5c0e099a688b29
3
+ size 14180
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:864c2fbfca6a229661e71bfe02bc47d1f2746dcbece8e0e2d9bd754183c9462a
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ffb6ab86afd8a1b76a16f7a0e015b06dbb79dc500dae39323cc69f06851bdbd
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a599609b8ba5b4e05421477366bf16e78592175247e304a0d39fba83577db549
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:894bd6c57ea4e17b26aada8f86d81df3acee0b45c80103268804ff49bb2ddbbc
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 10500,
3
- "best_metric": 1.8220328092575073,
4
- "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-10500",
5
- "epoch": 0.8076301822936697,
6
  "eval_steps": 250,
7
- "global_step": 10500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -1814,6 +1814,92 @@
1814
  "eval_samples_per_second": 58.606,
1815
  "eval_steps_per_second": 14.652,
1816
  "step": 10500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1817
  }
1818
  ],
1819
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 11000,
3
+ "best_metric": 1.807216763496399,
4
+ "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-11000",
5
+ "epoch": 0.8460887624028921,
6
  "eval_steps": 250,
7
+ "global_step": 11000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
1814
  "eval_samples_per_second": 58.606,
1815
  "eval_steps_per_second": 14.652,
1816
  "step": 10500
1817
+ },
1818
+ {
1819
+ "epoch": 0.811476040304592,
1820
+ "grad_norm": 1.8630917072296143,
1821
+ "learning_rate": 0.00014783263641794146,
1822
+ "loss": 1.9692,
1823
+ "step": 10550
1824
+ },
1825
+ {
1826
+ "epoch": 0.8153218983155142,
1827
+ "grad_norm": 1.744325876235962,
1828
+ "learning_rate": 0.00014757291639612497,
1829
+ "loss": 1.8893,
1830
+ "step": 10600
1831
+ },
1832
+ {
1833
+ "epoch": 0.8191677563264365,
1834
+ "grad_norm": 1.8451564311981201,
1835
+ "learning_rate": 0.0001473131963743085,
1836
+ "loss": 1.7862,
1837
+ "step": 10650
1838
+ },
1839
+ {
1840
+ "epoch": 0.8230136143373586,
1841
+ "grad_norm": 1.9895585775375366,
1842
+ "learning_rate": 0.00014705347635249203,
1843
+ "loss": 1.9978,
1844
+ "step": 10700
1845
+ },
1846
+ {
1847
+ "epoch": 0.8268594723482809,
1848
+ "grad_norm": 2.574155330657959,
1849
+ "learning_rate": 0.00014679375633067554,
1850
+ "loss": 1.8705,
1851
+ "step": 10750
1852
+ },
1853
+ {
1854
+ "epoch": 0.8268594723482809,
1855
+ "eval_loss": 1.8234485387802124,
1856
+ "eval_runtime": 17.061,
1857
+ "eval_samples_per_second": 58.613,
1858
+ "eval_steps_per_second": 14.653,
1859
+ "step": 10750
1860
+ },
1861
+ {
1862
+ "epoch": 0.8307053303592031,
1863
+ "grad_norm": 1.347955584526062,
1864
+ "learning_rate": 0.00014653403630885905,
1865
+ "loss": 1.7948,
1866
+ "step": 10800
1867
+ },
1868
+ {
1869
+ "epoch": 0.8345511883701254,
1870
+ "grad_norm": 2.8867287635803223,
1871
+ "learning_rate": 0.00014627431628704257,
1872
+ "loss": 1.8986,
1873
+ "step": 10850
1874
+ },
1875
+ {
1876
+ "epoch": 0.8383970463810476,
1877
+ "grad_norm": 2.692473888397217,
1878
+ "learning_rate": 0.00014601459626522608,
1879
+ "loss": 1.9209,
1880
+ "step": 10900
1881
+ },
1882
+ {
1883
+ "epoch": 0.8422429043919698,
1884
+ "grad_norm": 1.7835667133331299,
1885
+ "learning_rate": 0.00014575487624340962,
1886
+ "loss": 1.8406,
1887
+ "step": 10950
1888
+ },
1889
+ {
1890
+ "epoch": 0.8460887624028921,
1891
+ "grad_norm": 1.8864330053329468,
1892
+ "learning_rate": 0.00014549515622159313,
1893
+ "loss": 1.8001,
1894
+ "step": 11000
1895
+ },
1896
+ {
1897
+ "epoch": 0.8460887624028921,
1898
+ "eval_loss": 1.807216763496399,
1899
+ "eval_runtime": 17.2188,
1900
+ "eval_samples_per_second": 58.076,
1901
+ "eval_steps_per_second": 14.519,
1902
+ "step": 11000
1903
  }
1904
  ],
1905
  "logging_steps": 50,