Training in progress, step 12000, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 36730224
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:da26990df234d94448e406f402bbaab14595948f6b2bb6aaf6233e11502b39c8
|
| 3 |
size 36730224
|
last-checkpoint/ar_diffusion_info.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1736
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4fd8f2ea0f8f3f2b6317bf012fe6c554f0ab9b95b3c3bc864dd5cae3b6eef992
|
| 3 |
size 1736
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 73588346
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:70a89db0f278efb20b8c741665f32f39c512b7877cd595dd7f15cfb9e97916f3
|
| 3 |
size 73588346
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1696a7b978a5af1f9de4c1c4ce7f56376b2325426ded966dd6d734962f026c10
|
| 3 |
size 14244
|
last-checkpoint/scaler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 988
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7f076f7a63b00dbdf4d7a2eb7609297124339e4b64abf631b5119fa735b47e6a
|
| 3 |
size 988
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b7a1d25be56e05cb4cbafc0981ade0df33b24e8c1de440cdbae85f1fea41cdfa
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -1,10 +1,10 @@
|
|
| 1 |
{
|
| 2 |
-
"best_global_step":
|
| 3 |
-
"best_metric":
|
| 4 |
-
"best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-
|
| 5 |
-
"epoch": 0.
|
| 6 |
"eval_steps": 250,
|
| 7 |
-
"global_step":
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
@@ -1986,6 +1986,92 @@
|
|
| 1986 |
"eval_samples_per_second": 59.282,
|
| 1987 |
"eval_steps_per_second": 14.821,
|
| 1988 |
"step": 11500
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1989 |
}
|
| 1990 |
],
|
| 1991 |
"logging_steps": 50,
|
|
|
|
| 1 |
{
|
| 2 |
+
"best_global_step": 12000,
|
| 3 |
+
"best_metric": -30.337785720825195,
|
| 4 |
+
"best_model_checkpoint": "./ar-diffusion-checkpoints-progressive-attention/checkpoint-12000",
|
| 5 |
+
"epoch": 0.9230059226213368,
|
| 6 |
"eval_steps": 250,
|
| 7 |
+
"global_step": 12000,
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
|
|
| 1986 |
"eval_samples_per_second": 59.282,
|
| 1987 |
"eval_steps_per_second": 14.821,
|
| 1988 |
"step": 11500
|
| 1989 |
+
},
|
| 1990 |
+
{
|
| 1991 |
+
"epoch": 0.8883932005230367,
|
| 1992 |
+
"grad_norm": 2.3821170330047607,
|
| 1993 |
+
"learning_rate": 0.00014264343038204814,
|
| 1994 |
+
"loss": 1.4094,
|
| 1995 |
+
"step": 11550
|
| 1996 |
+
},
|
| 1997 |
+
{
|
| 1998 |
+
"epoch": 0.8922390585339589,
|
| 1999 |
+
"grad_norm": 1.0703688859939575,
|
| 2000 |
+
"learning_rate": 0.00014238371036023168,
|
| 2001 |
+
"loss": 1.3967,
|
| 2002 |
+
"step": 11600
|
| 2003 |
+
},
|
| 2004 |
+
{
|
| 2005 |
+
"epoch": 0.8960849165448812,
|
| 2006 |
+
"grad_norm": 0.9679076075553894,
|
| 2007 |
+
"learning_rate": 0.0001421239903384152,
|
| 2008 |
+
"loss": 1.3855,
|
| 2009 |
+
"step": 11650
|
| 2010 |
+
},
|
| 2011 |
+
{
|
| 2012 |
+
"epoch": 0.8999307745558034,
|
| 2013 |
+
"grad_norm": 1.2558847665786743,
|
| 2014 |
+
"learning_rate": 0.00014186427031659873,
|
| 2015 |
+
"loss": 1.3501,
|
| 2016 |
+
"step": 11700
|
| 2017 |
+
},
|
| 2018 |
+
{
|
| 2019 |
+
"epoch": 0.9037766325667257,
|
| 2020 |
+
"grad_norm": 2.6319808959960938,
|
| 2021 |
+
"learning_rate": 0.00014160455029478224,
|
| 2022 |
+
"loss": -15.5666,
|
| 2023 |
+
"step": 11750
|
| 2024 |
+
},
|
| 2025 |
+
{
|
| 2026 |
+
"epoch": 0.9037766325667257,
|
| 2027 |
+
"eval_loss": -26.468090057373047,
|
| 2028 |
+
"eval_runtime": 16.5143,
|
| 2029 |
+
"eval_samples_per_second": 60.554,
|
| 2030 |
+
"eval_steps_per_second": 15.138,
|
| 2031 |
+
"step": 11750
|
| 2032 |
+
},
|
| 2033 |
+
{
|
| 2034 |
+
"epoch": 0.9076224905776479,
|
| 2035 |
+
"grad_norm": 2.3091208934783936,
|
| 2036 |
+
"learning_rate": 0.00014134483027296576,
|
| 2037 |
+
"loss": -28.8843,
|
| 2038 |
+
"step": 11800
|
| 2039 |
+
},
|
| 2040 |
+
{
|
| 2041 |
+
"epoch": 0.9114683485885701,
|
| 2042 |
+
"grad_norm": 1.527166724205017,
|
| 2043 |
+
"learning_rate": 0.00014108511025114927,
|
| 2044 |
+
"loss": -30.1099,
|
| 2045 |
+
"step": 11850
|
| 2046 |
+
},
|
| 2047 |
+
{
|
| 2048 |
+
"epoch": 0.9153142065994924,
|
| 2049 |
+
"grad_norm": 3.9258713722229004,
|
| 2050 |
+
"learning_rate": 0.00014082539022933278,
|
| 2051 |
+
"loss": -30.2295,
|
| 2052 |
+
"step": 11900
|
| 2053 |
+
},
|
| 2054 |
+
{
|
| 2055 |
+
"epoch": 0.9191600646104146,
|
| 2056 |
+
"grad_norm": 1.618696928024292,
|
| 2057 |
+
"learning_rate": 0.0001405656702075163,
|
| 2058 |
+
"loss": -30.2778,
|
| 2059 |
+
"step": 11950
|
| 2060 |
+
},
|
| 2061 |
+
{
|
| 2062 |
+
"epoch": 0.9230059226213368,
|
| 2063 |
+
"grad_norm": 1.5081305503845215,
|
| 2064 |
+
"learning_rate": 0.00014030595018569984,
|
| 2065 |
+
"loss": -30.3095,
|
| 2066 |
+
"step": 12000
|
| 2067 |
+
},
|
| 2068 |
+
{
|
| 2069 |
+
"epoch": 0.9230059226213368,
|
| 2070 |
+
"eval_loss": -30.337785720825195,
|
| 2071 |
+
"eval_runtime": 16.3273,
|
| 2072 |
+
"eval_samples_per_second": 61.247,
|
| 2073 |
+
"eval_steps_per_second": 15.312,
|
| 2074 |
+
"step": 12000
|
| 2075 |
}
|
| 2076 |
],
|
| 2077 |
"logging_steps": 50,
|