rootxhacker commited on
Commit
ba0ca46
·
verified ·
1 Parent(s): 4d39fe7

Training in progress, step 10500, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fbda29e3fc4574d26f4c2d88945024a17cf0a9095cfd76acea4713a70561f4b9
3
  size 36730224
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7bbfabbed60d5906524d5c89409a9d91e99af768406b59d0122cc4c00803b88b
3
  size 36730224
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f807c16f8bec038522bfe252b213fc1087e8a39fb1f495399cdf7ee1d92e00a6
3
  size 73588346
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1f937a69fd1d3f87ca4cb29a7d82a362b45d23ba437525393801a63c53b09d6
3
  size 73588346
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2b4e6f78506ab09c8a3ff311a2dfeb7a6190585c49701c5bd2fcc763c2b50448
3
  size 14180
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:371a8ee92df296caadab094d1243e6f6a8e33531055d940b9b2beb58c3296d37
3
  size 14180
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c5057516ff1b0b207608fc5bf21c504e3b16c8f39cb674a438cf642593270922
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5bd55c94e5f9a6b24aa6ffb24230950971ac714f0b527005b89121745edab7aa
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cf5349680f22d749c8735deb8f593d381f787d0e5d89e99661139aee18144bbf
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3263d434a0173d1bf88ad28980a5eff0fc4c115a390bc663e18dd956338ac6e
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": 9500,
3
  "best_metric": 4.328299045562744,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-9500",
5
- "epoch": 0.7691716021844474,
6
  "eval_steps": 250,
7
- "global_step": 10000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -1728,6 +1728,92 @@
1728
  "eval_samples_per_second": 53.413,
1729
  "eval_steps_per_second": 13.353,
1730
  "step": 10000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1731
  }
1732
  ],
1733
  "logging_steps": 50,
 
2
  "best_global_step": 9500,
3
  "best_metric": 4.328299045562744,
4
  "best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-9500",
5
+ "epoch": 0.8076301822936697,
6
  "eval_steps": 250,
7
+ "global_step": 10500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
1728
  "eval_samples_per_second": 53.413,
1729
  "eval_steps_per_second": 13.353,
1730
  "step": 10000
1731
+ },
1732
+ {
1733
+ "epoch": 0.7730174601953695,
1734
+ "grad_norm": 3.490699529647827,
1735
+ "learning_rate": 8.884011617219954e-05,
1736
+ "loss": 4.5294,
1737
+ "step": 10050
1738
+ },
1739
+ {
1740
+ "epoch": 0.7768633182062918,
1741
+ "grad_norm": 4.614148139953613,
1742
+ "learning_rate": 8.858277269218042e-05,
1743
+ "loss": 4.2371,
1744
+ "step": 10100
1745
+ },
1746
+ {
1747
+ "epoch": 0.780709176217214,
1748
+ "grad_norm": 5.6906962394714355,
1749
+ "learning_rate": 8.832542921216132e-05,
1750
+ "loss": 4.5472,
1751
+ "step": 10150
1752
+ },
1753
+ {
1754
+ "epoch": 0.7845550342281363,
1755
+ "grad_norm": 4.382456302642822,
1756
+ "learning_rate": 8.80680857321422e-05,
1757
+ "loss": 4.4282,
1758
+ "step": 10200
1759
+ },
1760
+ {
1761
+ "epoch": 0.7884008922390585,
1762
+ "grad_norm": 4.546772003173828,
1763
+ "learning_rate": 8.781074225212309e-05,
1764
+ "loss": 4.4004,
1765
+ "step": 10250
1766
+ },
1767
+ {
1768
+ "epoch": 0.7884008922390585,
1769
+ "eval_loss": 4.373971462249756,
1770
+ "eval_runtime": 18.9303,
1771
+ "eval_samples_per_second": 52.825,
1772
+ "eval_steps_per_second": 13.206,
1773
+ "step": 10250
1774
+ },
1775
+ {
1776
+ "epoch": 0.7922467502499808,
1777
+ "grad_norm": 3.784317970275879,
1778
+ "learning_rate": 8.755339877210395e-05,
1779
+ "loss": 4.4422,
1780
+ "step": 10300
1781
+ },
1782
+ {
1783
+ "epoch": 0.796092608260903,
1784
+ "grad_norm": 3.11979341506958,
1785
+ "learning_rate": 8.729605529208483e-05,
1786
+ "loss": 4.4909,
1787
+ "step": 10350
1788
+ },
1789
+ {
1790
+ "epoch": 0.7999384662718253,
1791
+ "grad_norm": 4.9711012840271,
1792
+ "learning_rate": 8.703871181206573e-05,
1793
+ "loss": 4.2955,
1794
+ "step": 10400
1795
+ },
1796
+ {
1797
+ "epoch": 0.8037843242827475,
1798
+ "grad_norm": 3.7663426399230957,
1799
+ "learning_rate": 8.678136833204661e-05,
1800
+ "loss": 4.5105,
1801
+ "step": 10450
1802
+ },
1803
+ {
1804
+ "epoch": 0.8076301822936697,
1805
+ "grad_norm": 4.679628372192383,
1806
+ "learning_rate": 8.65240248520275e-05,
1807
+ "loss": 4.5038,
1808
+ "step": 10500
1809
+ },
1810
+ {
1811
+ "epoch": 0.8076301822936697,
1812
+ "eval_loss": 4.3565592765808105,
1813
+ "eval_runtime": 18.9119,
1814
+ "eval_samples_per_second": 52.877,
1815
+ "eval_steps_per_second": 13.219,
1816
+ "step": 10500
1817
  }
1818
  ],
1819
  "logging_steps": 50,