Training in progress, step 10500, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 36730224
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ac66d1bed32cb77a6fe046d97ff60fe2cfaef95ddf73a9bf13a7ff56a55c11af
|
| 3 |
size 36730224
|
last-checkpoint/ar_diffusion_info.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1736
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:298a32675b40f08695b3fb4fa7bbd261866561afa1edbe21444c5e1018c7c6df
|
| 3 |
size 1736
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 73588346
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:df829391bae99716149a25d0f76e0e75e0e3329d3b7ef7d8f960f81bf3520101
|
| 3 |
size 73588346
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2286056e584e728d67ca2ee5e9e130931cc603f814dc38d9957d5e8ae618fb75
|
| 3 |
+
size 14244
|
last-checkpoint/scaler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 988
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:64e34184efeb9f5a5622c883abc37387f021455b8ebc365e06c304057c5bb8b6
|
| 3 |
size 988
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:83df1333fe8f59c1710a91e769544177d60bc5aa48c1018ddd6a061172e2a46b
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -1,10 +1,10 @@
|
|
| 1 |
{
|
| 2 |
-
"best_global_step":
|
| 3 |
-
"best_metric": 1.
|
| 4 |
-
"best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-
|
| 5 |
-
"epoch": 0.
|
| 6 |
"eval_steps": 250,
|
| 7 |
-
"global_step":
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
@@ -1728,6 +1728,92 @@
|
|
| 1728 |
"eval_samples_per_second": 59.387,
|
| 1729 |
"eval_steps_per_second": 14.847,
|
| 1730 |
"step": 10000
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1731 |
}
|
| 1732 |
],
|
| 1733 |
"logging_steps": 50,
|
|
|
|
| 1 |
{
|
| 2 |
+
"best_global_step": 10500,
|
| 3 |
+
"best_metric": 1.3613618612289429,
|
| 4 |
+
"best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-10500",
|
| 5 |
+
"epoch": 0.8076301822936697,
|
| 6 |
"eval_steps": 250,
|
| 7 |
+
"global_step": 10500,
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
|
|
| 1728 |
"eval_samples_per_second": 59.387,
|
| 1729 |
"eval_steps_per_second": 14.847,
|
| 1730 |
"step": 10000
|
| 1731 |
+
},
|
| 1732 |
+
{
|
| 1733 |
+
"epoch": 0.7730174601953695,
|
| 1734 |
+
"grad_norm": 1.1201566457748413,
|
| 1735 |
+
"learning_rate": 0.00015042464223566997,
|
| 1736 |
+
"loss": 1.4327,
|
| 1737 |
+
"step": 10050
|
| 1738 |
+
},
|
| 1739 |
+
{
|
| 1740 |
+
"epoch": 0.7768633182062918,
|
| 1741 |
+
"grad_norm": 1.7418923377990723,
|
| 1742 |
+
"learning_rate": 0.00015016492221385348,
|
| 1743 |
+
"loss": 1.2776,
|
| 1744 |
+
"step": 10100
|
| 1745 |
+
},
|
| 1746 |
+
{
|
| 1747 |
+
"epoch": 0.780709176217214,
|
| 1748 |
+
"grad_norm": 1.3154345750808716,
|
| 1749 |
+
"learning_rate": 0.000149905202192037,
|
| 1750 |
+
"loss": 1.4374,
|
| 1751 |
+
"step": 10150
|
| 1752 |
+
},
|
| 1753 |
+
{
|
| 1754 |
+
"epoch": 0.7845550342281363,
|
| 1755 |
+
"grad_norm": 1.129528284072876,
|
| 1756 |
+
"learning_rate": 0.0001496454821702205,
|
| 1757 |
+
"loss": 1.3404,
|
| 1758 |
+
"step": 10200
|
| 1759 |
+
},
|
| 1760 |
+
{
|
| 1761 |
+
"epoch": 0.7884008922390585,
|
| 1762 |
+
"grad_norm": 1.3940690755844116,
|
| 1763 |
+
"learning_rate": 0.00014938576214840402,
|
| 1764 |
+
"loss": 1.3828,
|
| 1765 |
+
"step": 10250
|
| 1766 |
+
},
|
| 1767 |
+
{
|
| 1768 |
+
"epoch": 0.7884008922390585,
|
| 1769 |
+
"eval_loss": 1.3682047128677368,
|
| 1770 |
+
"eval_runtime": 17.0099,
|
| 1771 |
+
"eval_samples_per_second": 58.789,
|
| 1772 |
+
"eval_steps_per_second": 14.697,
|
| 1773 |
+
"step": 10250
|
| 1774 |
+
},
|
| 1775 |
+
{
|
| 1776 |
+
"epoch": 0.7922467502499808,
|
| 1777 |
+
"grad_norm": 1.3081247806549072,
|
| 1778 |
+
"learning_rate": 0.00014912604212658753,
|
| 1779 |
+
"loss": 1.4071,
|
| 1780 |
+
"step": 10300
|
| 1781 |
+
},
|
| 1782 |
+
{
|
| 1783 |
+
"epoch": 0.796092608260903,
|
| 1784 |
+
"grad_norm": 0.8190014362335205,
|
| 1785 |
+
"learning_rate": 0.00014886632210477107,
|
| 1786 |
+
"loss": 1.4668,
|
| 1787 |
+
"step": 10350
|
| 1788 |
+
},
|
| 1789 |
+
{
|
| 1790 |
+
"epoch": 0.7999384662718253,
|
| 1791 |
+
"grad_norm": 1.8602573871612549,
|
| 1792 |
+
"learning_rate": 0.00014860660208295459,
|
| 1793 |
+
"loss": 1.3456,
|
| 1794 |
+
"step": 10400
|
| 1795 |
+
},
|
| 1796 |
+
{
|
| 1797 |
+
"epoch": 0.8037843242827475,
|
| 1798 |
+
"grad_norm": 1.4184271097183228,
|
| 1799 |
+
"learning_rate": 0.0001483468820611381,
|
| 1800 |
+
"loss": 1.3851,
|
| 1801 |
+
"step": 10450
|
| 1802 |
+
},
|
| 1803 |
+
{
|
| 1804 |
+
"epoch": 0.8076301822936697,
|
| 1805 |
+
"grad_norm": 1.5128880739212036,
|
| 1806 |
+
"learning_rate": 0.00014808716203932164,
|
| 1807 |
+
"loss": 1.3869,
|
| 1808 |
+
"step": 10500
|
| 1809 |
+
},
|
| 1810 |
+
{
|
| 1811 |
+
"epoch": 0.8076301822936697,
|
| 1812 |
+
"eval_loss": 1.3613618612289429,
|
| 1813 |
+
"eval_runtime": 16.8001,
|
| 1814 |
+
"eval_samples_per_second": 59.523,
|
| 1815 |
+
"eval_steps_per_second": 14.881,
|
| 1816 |
+
"step": 10500
|
| 1817 |
}
|
| 1818 |
],
|
| 1819 |
"logging_steps": 50,
|