Training in progress, step 12000, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 36730224
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f905e1780ead207018fcc63d3eee314969b429ab6c3fb1b4b7d55b82dd5e705f
|
| 3 |
size 36730224
|
last-checkpoint/ar_diffusion_info.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1544
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0543052e8eb8d6818c70e093e941fc4b27e91541fab33a19886d162002c431a7
|
| 3 |
size 1544
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 73588346
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7a12af399801c91cf8dd422d90b7d653b67f6015e364538d005d0a7d95d3b977
|
| 3 |
size 73588346
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:205a7550a4c055924e0126c3501b4b656ca55ae3243c46218f86d29fc2fb9758
|
| 3 |
size 14244
|
last-checkpoint/scaler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 988
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a647bb95d822173d4ee9be8d10c936b89d3eece1d84247711563ab2d41513f54
|
| 3 |
size 988
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5752b9d4c614021fa8b8f77fd50a0a39abf6d6bd5fececfd9db71cacad56f038
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -1,10 +1,10 @@
|
|
| 1 |
{
|
| 2 |
-
"best_global_step":
|
| 3 |
-
"best_metric":
|
| 4 |
-
"best_model_checkpoint": "./ar-diffusion-checkpoints-fixed/checkpoint-
|
| 5 |
-
"epoch": 0.
|
| 6 |
"eval_steps": 250,
|
| 7 |
-
"global_step":
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
@@ -1986,6 +1986,92 @@
|
|
| 1986 |
"eval_samples_per_second": 53.887,
|
| 1987 |
"eval_steps_per_second": 13.472,
|
| 1988 |
"step": 11500
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1989 |
}
|
| 1990 |
],
|
| 1991 |
"logging_steps": 50,
|
|
|
|
| 1 |
{
|
| 2 |
+
"best_global_step": 12000,
|
| 3 |
+
"best_metric": 1.5273067951202393,
|
| 4 |
+
"best_model_checkpoint": "./ar-diffusion-checkpoints-fixed/checkpoint-12000",
|
| 5 |
+
"epoch": 0.9230059226213368,
|
| 6 |
"eval_steps": 250,
|
| 7 |
+
"global_step": 12000,
|
| 8 |
"is_hyper_param_search": false,
|
| 9 |
"is_local_process_zero": true,
|
| 10 |
"is_world_process_zero": true,
|
|
|
|
| 1986 |
"eval_samples_per_second": 53.887,
|
| 1987 |
"eval_steps_per_second": 13.472,
|
| 1988 |
"step": 11500
|
| 1989 |
+
},
|
| 1990 |
+
{
|
| 1991 |
+
"epoch": 0.8883932005230367,
|
| 1992 |
+
"grad_norm": 4.711415767669678,
|
| 1993 |
+
"learning_rate": 0.00014265381918292082,
|
| 1994 |
+
"loss": 4.5701,
|
| 1995 |
+
"step": 11550
|
| 1996 |
+
},
|
| 1997 |
+
{
|
| 1998 |
+
"epoch": 0.8922390585339589,
|
| 1999 |
+
"grad_norm": 3.1737523078918457,
|
| 2000 |
+
"learning_rate": 0.00014239409916110433,
|
| 2001 |
+
"loss": 4.4522,
|
| 2002 |
+
"step": 11600
|
| 2003 |
+
},
|
| 2004 |
+
{
|
| 2005 |
+
"epoch": 0.8960849165448812,
|
| 2006 |
+
"grad_norm": 4.876018047332764,
|
| 2007 |
+
"learning_rate": 0.00014213437913928787,
|
| 2008 |
+
"loss": 4.3937,
|
| 2009 |
+
"step": 11650
|
| 2010 |
+
},
|
| 2011 |
+
{
|
| 2012 |
+
"epoch": 0.8999307745558034,
|
| 2013 |
+
"grad_norm": 7.117967128753662,
|
| 2014 |
+
"learning_rate": 0.00014187465911747138,
|
| 2015 |
+
"loss": 4.3585,
|
| 2016 |
+
"step": 11700
|
| 2017 |
+
},
|
| 2018 |
+
{
|
| 2019 |
+
"epoch": 0.9037766325667257,
|
| 2020 |
+
"grad_norm": 2.587160587310791,
|
| 2021 |
+
"learning_rate": 0.0001416149390956549,
|
| 2022 |
+
"loss": 1.5054,
|
| 2023 |
+
"step": 11750
|
| 2024 |
+
},
|
| 2025 |
+
{
|
| 2026 |
+
"epoch": 0.9037766325667257,
|
| 2027 |
+
"eval_loss": 1.5468424558639526,
|
| 2028 |
+
"eval_runtime": 17.9563,
|
| 2029 |
+
"eval_samples_per_second": 55.691,
|
| 2030 |
+
"eval_steps_per_second": 13.923,
|
| 2031 |
+
"step": 11750
|
| 2032 |
+
},
|
| 2033 |
+
{
|
| 2034 |
+
"epoch": 0.9076224905776479,
|
| 2035 |
+
"grad_norm": 1.4935526847839355,
|
| 2036 |
+
"learning_rate": 0.0001413552190738384,
|
| 2037 |
+
"loss": 1.5719,
|
| 2038 |
+
"step": 11800
|
| 2039 |
+
},
|
| 2040 |
+
{
|
| 2041 |
+
"epoch": 0.9114683485885701,
|
| 2042 |
+
"grad_norm": 1.4879201650619507,
|
| 2043 |
+
"learning_rate": 0.00014109549905202192,
|
| 2044 |
+
"loss": 1.4862,
|
| 2045 |
+
"step": 11850
|
| 2046 |
+
},
|
| 2047 |
+
{
|
| 2048 |
+
"epoch": 0.9153142065994924,
|
| 2049 |
+
"grad_norm": 1.7936193943023682,
|
| 2050 |
+
"learning_rate": 0.00014083577903020543,
|
| 2051 |
+
"loss": 1.488,
|
| 2052 |
+
"step": 11900
|
| 2053 |
+
},
|
| 2054 |
+
{
|
| 2055 |
+
"epoch": 0.9191600646104146,
|
| 2056 |
+
"grad_norm": 2.14953875541687,
|
| 2057 |
+
"learning_rate": 0.00014057605900838897,
|
| 2058 |
+
"loss": 1.5383,
|
| 2059 |
+
"step": 11950
|
| 2060 |
+
},
|
| 2061 |
+
{
|
| 2062 |
+
"epoch": 0.9230059226213368,
|
| 2063 |
+
"grad_norm": 1.5575013160705566,
|
| 2064 |
+
"learning_rate": 0.00014031633898657249,
|
| 2065 |
+
"loss": 1.4814,
|
| 2066 |
+
"step": 12000
|
| 2067 |
+
},
|
| 2068 |
+
{
|
| 2069 |
+
"epoch": 0.9230059226213368,
|
| 2070 |
+
"eval_loss": 1.5273067951202393,
|
| 2071 |
+
"eval_runtime": 17.7259,
|
| 2072 |
+
"eval_samples_per_second": 56.414,
|
| 2073 |
+
"eval_steps_per_second": 14.104,
|
| 2074 |
+
"step": 12000
|
| 2075 |
}
|
| 2076 |
],
|
| 2077 |
"logging_steps": 50,
|