Training in progress, step 22500, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 36730224
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:48890795503039966e7c3f8690ea5e69f170d6d775e60a7c19814730a3ff98f0
|
| 3 |
size 36730224
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 73588346
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:323d035573abc68ea1fb28a0dbd169b42c919244458b9d48c24a8196ef4f1fbd
|
| 3 |
size 73588346
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0edeec917b285b0e88139f206de61b3e301e9b677aa51d87d86f36aba6dd8e7e
|
| 3 |
size 14244
|
last-checkpoint/scaler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 988
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b66cb5271b1546b480a6c7621e1addbb2863a83f6bfe9ced4a812a79ec729e68
|
| 3 |
size 988
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ee960841273de5fa43863b36a2f8d901df0b6ff85b61e6b4992f108a85a45d9c
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -1,10 +1,10 @@
|
|
| 1 |
{
|
| 2 |
-
"best_global_step":
|
| 3 |
-
"best_metric": 1.
|
| 4 |
"best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-22000",
|
| 5 |
-
"epoch": 1.
|
| 6 |
"eval_steps": 250,
|
| 7 |
-
"global_step":
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
@@ -3792,6 +3792,92 @@
|
|
| 3792 |
"eval_samples_per_second": 55.454,
|
| 3793 |
"eval_steps_per_second": 13.863,
|
| 3794 |
"step": 22000
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 3795 |
}
|
| 3796 |
],
|
| 3797 |
"logging_steps": 50,
|
|
|
|
| 1 |
{
|
| 2 |
+
"best_global_step": 22250,
|
| 3 |
+
"best_metric": 1.457463264465332,
|
| 4 |
"best_model_checkpoint": "./ar-diffusion-checkpoints/checkpoint-22000",
|
| 5 |
+
"epoch": 1.7306361049150065,
|
| 6 |
"eval_steps": 250,
|
| 7 |
+
"global_step": 22500,
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
|
|
| 3792 |
"eval_samples_per_second": 55.454,
|
| 3793 |
"eval_steps_per_second": 13.863,
|
| 3794 |
"step": 22000
|
| 3795 |
+
},
|
| 3796 |
+
{
|
| 3797 |
+
"epoch": 1.6960233828167064,
|
| 3798 |
+
"grad_norm": 0.8262001276016235,
|
| 3799 |
+
"learning_rate": 2.7087974706812248e-05,
|
| 3800 |
+
"loss": 1.4304,
|
| 3801 |
+
"step": 22050
|
| 3802 |
+
},
|
| 3803 |
+
{
|
| 3804 |
+
"epoch": 1.6998692408276286,
|
| 3805 |
+
"grad_norm": 1.6224443912506104,
|
| 3806 |
+
"learning_rate": 2.6830631226793128e-05,
|
| 3807 |
+
"loss": 1.4127,
|
| 3808 |
+
"step": 22100
|
| 3809 |
+
},
|
| 3810 |
+
{
|
| 3811 |
+
"epoch": 1.703715098838551,
|
| 3812 |
+
"grad_norm": 1.3338160514831543,
|
| 3813 |
+
"learning_rate": 2.6573287746774014e-05,
|
| 3814 |
+
"loss": 1.4842,
|
| 3815 |
+
"step": 22150
|
| 3816 |
+
},
|
| 3817 |
+
{
|
| 3818 |
+
"epoch": 1.707560956849473,
|
| 3819 |
+
"grad_norm": 1.940238356590271,
|
| 3820 |
+
"learning_rate": 2.6315944266754897e-05,
|
| 3821 |
+
"loss": 1.4279,
|
| 3822 |
+
"step": 22200
|
| 3823 |
+
},
|
| 3824 |
+
{
|
| 3825 |
+
"epoch": 1.7114068148603954,
|
| 3826 |
+
"grad_norm": 2.091132164001465,
|
| 3827 |
+
"learning_rate": 2.6058600786735777e-05,
|
| 3828 |
+
"loss": 1.3779,
|
| 3829 |
+
"step": 22250
|
| 3830 |
+
},
|
| 3831 |
+
{
|
| 3832 |
+
"epoch": 1.7114068148603954,
|
| 3833 |
+
"eval_loss": 1.457463264465332,
|
| 3834 |
+
"eval_runtime": 18.1835,
|
| 3835 |
+
"eval_samples_per_second": 54.995,
|
| 3836 |
+
"eval_steps_per_second": 13.749,
|
| 3837 |
+
"step": 22250
|
| 3838 |
+
},
|
| 3839 |
+
{
|
| 3840 |
+
"epoch": 1.7152526728713176,
|
| 3841 |
+
"grad_norm": 1.4367913007736206,
|
| 3842 |
+
"learning_rate": 2.5801257306716663e-05,
|
| 3843 |
+
"loss": 1.4821,
|
| 3844 |
+
"step": 22300
|
| 3845 |
+
},
|
| 3846 |
+
{
|
| 3847 |
+
"epoch": 1.7190985308822397,
|
| 3848 |
+
"grad_norm": 1.9735435247421265,
|
| 3849 |
+
"learning_rate": 2.5543913826697546e-05,
|
| 3850 |
+
"loss": 1.3754,
|
| 3851 |
+
"step": 22350
|
| 3852 |
+
},
|
| 3853 |
+
{
|
| 3854 |
+
"epoch": 1.7229443888931621,
|
| 3855 |
+
"grad_norm": 1.4968055486679077,
|
| 3856 |
+
"learning_rate": 2.5286570346678426e-05,
|
| 3857 |
+
"loss": 1.4045,
|
| 3858 |
+
"step": 22400
|
| 3859 |
+
},
|
| 3860 |
+
{
|
| 3861 |
+
"epoch": 1.7267902469040843,
|
| 3862 |
+
"grad_norm": 1.0449949502944946,
|
| 3863 |
+
"learning_rate": 2.5029226866659312e-05,
|
| 3864 |
+
"loss": 1.4458,
|
| 3865 |
+
"step": 22450
|
| 3866 |
+
},
|
| 3867 |
+
{
|
| 3868 |
+
"epoch": 1.7306361049150065,
|
| 3869 |
+
"grad_norm": 1.164890170097351,
|
| 3870 |
+
"learning_rate": 2.4771883386640196e-05,
|
| 3871 |
+
"loss": 1.4407,
|
| 3872 |
+
"step": 22500
|
| 3873 |
+
},
|
| 3874 |
+
{
|
| 3875 |
+
"epoch": 1.7306361049150065,
|
| 3876 |
+
"eval_loss": 1.4607012271881104,
|
| 3877 |
+
"eval_runtime": 18.2079,
|
| 3878 |
+
"eval_samples_per_second": 54.921,
|
| 3879 |
+
"eval_steps_per_second": 13.73,
|
| 3880 |
+
"step": 22500
|
| 3881 |
}
|
| 3882 |
],
|
| 3883 |
"logging_steps": 50,
|