rootxhacker commited on
Commit
e2be31a
·
verified ·
1 Parent(s): 96f1399

Training in progress, step 10500, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:42f1ca9dac6b6ebcdce4b00359a28210be4ba525244b591572f5b2287e25b0f9
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2355e3e82295bb14e617f48ba9a53dcaecdce82329ed609cc4929ae9d83b250b
3
  size 36730224
last-checkpoint/ar_diffusion_info.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:93ef9d4a1c4c97b75d44e8a9d7ebd61f014e4cbe01f8111a09cb70d30d2060b0
3
  size 1736
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e45f55b2e0437c4eb0d6bdb520b7b1c7edbedaf9253b60e08be085f388b8113
3
  size 1736
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6c0943241d3c12af607553b4e6abe354fb336ff905d2c1fb46e1651cbdb980d6
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90e2bba69ad980b5be837dce2dab49768cfe98d25981ae562b11cdccf65223e4
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:233660262675d93c5fec35f1803b0486ea65b8e55d6ffa01a964eeb9acc8c92f
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0766963e075955373626125e0506ee76918124a59001be5cd204b364faad380e
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:250db0f613d9618dc37866ea93a85951fa852bff7fd15eb41a9bb89d17f93c5f
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:864c2fbfca6a229661e71bfe02bc47d1f2746dcbece8e0e2d9bd754183c9462a
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cd971f0e9f9fab99e52eb560d68aa87eb93fac7faa2dc0c68c4d5d10aa759271
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a599609b8ba5b4e05421477366bf16e78592175247e304a0d39fba83577db549
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 10000,
3
- "best_metric": 1.8548085689544678,
4
- "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-10000",
5
- "epoch": 0.7691716021844474,
6
  "eval_steps": 250,
7
- "global_step": 10000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -1728,6 +1728,92 @@
1728
  "eval_samples_per_second": 58.746,
1729
  "eval_steps_per_second": 14.687,
1730
  "step": 10000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1731
  }
1732
  ],
1733
  "logging_steps": 50,
 
1
  {
2
+ "best_global_step": 10500,
3
+ "best_metric": 1.8220328092575073,
4
+ "best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-10500",
5
+ "epoch": 0.8076301822936697,
6
  "eval_steps": 250,
7
+ "global_step": 10500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
1728
  "eval_samples_per_second": 58.746,
1729
  "eval_steps_per_second": 14.687,
1730
  "step": 10000
1731
+ },
1732
+ {
1733
+ "epoch": 0.7730174601953695,
1734
+ "grad_norm": 1.4667410850524902,
1735
+ "learning_rate": 0.00015042983663610628,
1736
+ "loss": 1.9318,
1737
+ "step": 10050
1738
+ },
1739
+ {
1740
+ "epoch": 0.7768633182062918,
1741
+ "grad_norm": 2.0754499435424805,
1742
+ "learning_rate": 0.00015017011661428982,
1743
+ "loss": 1.753,
1744
+ "step": 10100
1745
+ },
1746
+ {
1747
+ "epoch": 0.780709176217214,
1748
+ "grad_norm": 1.634293794631958,
1749
+ "learning_rate": 0.00014991039659247333,
1750
+ "loss": 1.9328,
1751
+ "step": 10150
1752
+ },
1753
+ {
1754
+ "epoch": 0.7845550342281363,
1755
+ "grad_norm": 1.5003916025161743,
1756
+ "learning_rate": 0.00014965067657065684,
1757
+ "loss": 1.8237,
1758
+ "step": 10200
1759
+ },
1760
+ {
1761
+ "epoch": 0.7884008922390585,
1762
+ "grad_norm": 1.7433470487594604,
1763
+ "learning_rate": 0.00014939095654884036,
1764
+ "loss": 1.859,
1765
+ "step": 10250
1766
+ },
1767
+ {
1768
+ "epoch": 0.7884008922390585,
1769
+ "eval_loss": 1.842873215675354,
1770
+ "eval_runtime": 16.9242,
1771
+ "eval_samples_per_second": 59.087,
1772
+ "eval_steps_per_second": 14.772,
1773
+ "step": 10250
1774
+ },
1775
+ {
1776
+ "epoch": 0.7922467502499808,
1777
+ "grad_norm": 1.8703033924102783,
1778
+ "learning_rate": 0.00014913123652702387,
1779
+ "loss": 1.9019,
1780
+ "step": 10300
1781
+ },
1782
+ {
1783
+ "epoch": 0.796092608260903,
1784
+ "grad_norm": 1.1216288805007935,
1785
+ "learning_rate": 0.00014887151650520738,
1786
+ "loss": 1.9812,
1787
+ "step": 10350
1788
+ },
1789
+ {
1790
+ "epoch": 0.7999384662718253,
1791
+ "grad_norm": 3.2229816913604736,
1792
+ "learning_rate": 0.00014861179648339092,
1793
+ "loss": 1.7986,
1794
+ "step": 10400
1795
+ },
1796
+ {
1797
+ "epoch": 0.8037843242827475,
1798
+ "grad_norm": 2.366506814956665,
1799
+ "learning_rate": 0.00014835207646157444,
1800
+ "loss": 1.8995,
1801
+ "step": 10450
1802
+ },
1803
+ {
1804
+ "epoch": 0.8076301822936697,
1805
+ "grad_norm": 2.0333054065704346,
1806
+ "learning_rate": 0.00014809235643975795,
1807
+ "loss": 1.916,
1808
+ "step": 10500
1809
+ },
1810
+ {
1811
+ "epoch": 0.8076301822936697,
1812
+ "eval_loss": 1.8220328092575073,
1813
+ "eval_runtime": 17.063,
1814
+ "eval_samples_per_second": 58.606,
1815
+ "eval_steps_per_second": 14.652,
1816
+ "step": 10500
1817
  }
1818
  ],
1819
  "logging_steps": 50,