Training in progress, epoch 21, checkpoint
Browse files
last-checkpoint/model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1227009528
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:759e11911622505bdb5a77d511b901f277ce0bbc06c416da623a53f9c0f0b663
|
| 3 |
size 1227009528
|
last-checkpoint/optimizer.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 2454133690
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c251d2bf032c0284e58db70391eefabbbff2c2fb3bc7e50e9d6ba42d7c0345d5
|
| 3 |
size 2454133690
|
last-checkpoint/rng_state.pth
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14244
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4db8bac039de345d3f184975bc589d15c0b93c070db1287ed5147dba9f5ba405
|
| 3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1064
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e7adf3363afe258d5bb46d306ca21d84331e969bff032715bb85737194722fbc
|
| 3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
|
@@ -1,9 +1,9 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 34.54485321044922,
|
| 3 |
"best_model_checkpoint": "/kaggle/working/output/checkpoint-20880",
|
| 4 |
-
"epoch":
|
| 5 |
"eval_steps": 500,
|
| 6 |
-
"global_step":
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
@@ -1994,6 +1994,105 @@
|
|
| 1994 |
"eval_samples_per_second": 26.478,
|
| 1995 |
"eval_steps_per_second": 3.328,
|
| 1996 |
"step": 26100
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1997 |
}
|
| 1998 |
],
|
| 1999 |
"logging_steps": 100,
|
|
@@ -2008,7 +2107,7 @@
|
|
| 2008 |
"early_stopping_threshold": 0.0
|
| 2009 |
},
|
| 2010 |
"attributes": {
|
| 2011 |
-
"early_stopping_patience_counter":
|
| 2012 |
}
|
| 2013 |
},
|
| 2014 |
"TrainerControl": {
|
|
@@ -2022,7 +2121,7 @@
|
|
| 2022 |
"attributes": {}
|
| 2023 |
}
|
| 2024 |
},
|
| 2025 |
-
"total_flos": 2.
|
| 2026 |
"train_batch_size": 8,
|
| 2027 |
"trial_name": null,
|
| 2028 |
"trial_params": null
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 34.54485321044922,
|
| 3 |
"best_model_checkpoint": "/kaggle/working/output/checkpoint-20880",
|
| 4 |
+
"epoch": 21.0,
|
| 5 |
"eval_steps": 500,
|
| 6 |
+
"global_step": 27405,
|
| 7 |
"is_hyper_param_search": false,
|
| 8 |
"is_local_process_zero": true,
|
| 9 |
"is_world_process_zero": true,
|
|
|
|
| 1994 |
"eval_samples_per_second": 26.478,
|
| 1995 |
"eval_steps_per_second": 3.328,
|
| 1996 |
"step": 26100
|
| 1997 |
+
},
|
| 1998 |
+
{
|
| 1999 |
+
"epoch": 20.07662835249042,
|
| 2000 |
+
"grad_norm": 4.07724142074585,
|
| 2001 |
+
"learning_rate": 3.745785440613027e-05,
|
| 2002 |
+
"loss": 33.562,
|
| 2003 |
+
"step": 26200
|
| 2004 |
+
},
|
| 2005 |
+
{
|
| 2006 |
+
"epoch": 20.153256704980844,
|
| 2007 |
+
"grad_norm": 4.335379600524902,
|
| 2008 |
+
"learning_rate": 3.7409961685823756e-05,
|
| 2009 |
+
"loss": 33.166,
|
| 2010 |
+
"step": 26300
|
| 2011 |
+
},
|
| 2012 |
+
{
|
| 2013 |
+
"epoch": 20.229885057471265,
|
| 2014 |
+
"grad_norm": 5.472820281982422,
|
| 2015 |
+
"learning_rate": 3.736206896551724e-05,
|
| 2016 |
+
"loss": 33.8918,
|
| 2017 |
+
"step": 26400
|
| 2018 |
+
},
|
| 2019 |
+
{
|
| 2020 |
+
"epoch": 20.306513409961685,
|
| 2021 |
+
"grad_norm": 3.011789321899414,
|
| 2022 |
+
"learning_rate": 3.731417624521073e-05,
|
| 2023 |
+
"loss": 33.395,
|
| 2024 |
+
"step": 26500
|
| 2025 |
+
},
|
| 2026 |
+
{
|
| 2027 |
+
"epoch": 20.38314176245211,
|
| 2028 |
+
"grad_norm": 3.251089096069336,
|
| 2029 |
+
"learning_rate": 3.7266283524904216e-05,
|
| 2030 |
+
"loss": 32.9072,
|
| 2031 |
+
"step": 26600
|
| 2032 |
+
},
|
| 2033 |
+
{
|
| 2034 |
+
"epoch": 20.45977011494253,
|
| 2035 |
+
"grad_norm": 2.7508978843688965,
|
| 2036 |
+
"learning_rate": 3.72183908045977e-05,
|
| 2037 |
+
"loss": 33.92,
|
| 2038 |
+
"step": 26700
|
| 2039 |
+
},
|
| 2040 |
+
{
|
| 2041 |
+
"epoch": 20.53639846743295,
|
| 2042 |
+
"grad_norm": 2.8051536083221436,
|
| 2043 |
+
"learning_rate": 3.717049808429119e-05,
|
| 2044 |
+
"loss": 33.9392,
|
| 2045 |
+
"step": 26800
|
| 2046 |
+
},
|
| 2047 |
+
{
|
| 2048 |
+
"epoch": 20.613026819923373,
|
| 2049 |
+
"grad_norm": 7.377379417419434,
|
| 2050 |
+
"learning_rate": 3.712260536398468e-05,
|
| 2051 |
+
"loss": 33.0382,
|
| 2052 |
+
"step": 26900
|
| 2053 |
+
},
|
| 2054 |
+
{
|
| 2055 |
+
"epoch": 20.689655172413794,
|
| 2056 |
+
"grad_norm": 3.7770464420318604,
|
| 2057 |
+
"learning_rate": 3.7074712643678164e-05,
|
| 2058 |
+
"loss": 32.6836,
|
| 2059 |
+
"step": 27000
|
| 2060 |
+
},
|
| 2061 |
+
{
|
| 2062 |
+
"epoch": 20.766283524904214,
|
| 2063 |
+
"grad_norm": 4.923346996307373,
|
| 2064 |
+
"learning_rate": 3.702681992337165e-05,
|
| 2065 |
+
"loss": 33.2129,
|
| 2066 |
+
"step": 27100
|
| 2067 |
+
},
|
| 2068 |
+
{
|
| 2069 |
+
"epoch": 20.842911877394634,
|
| 2070 |
+
"grad_norm": 4.790703773498535,
|
| 2071 |
+
"learning_rate": 3.697892720306513e-05,
|
| 2072 |
+
"loss": 33.5413,
|
| 2073 |
+
"step": 27200
|
| 2074 |
+
},
|
| 2075 |
+
{
|
| 2076 |
+
"epoch": 20.919540229885058,
|
| 2077 |
+
"grad_norm": 4.592926025390625,
|
| 2078 |
+
"learning_rate": 3.6931034482758624e-05,
|
| 2079 |
+
"loss": 33.2436,
|
| 2080 |
+
"step": 27300
|
| 2081 |
+
},
|
| 2082 |
+
{
|
| 2083 |
+
"epoch": 20.99616858237548,
|
| 2084 |
+
"grad_norm": 3.0529520511627197,
|
| 2085 |
+
"learning_rate": 3.688314176245211e-05,
|
| 2086 |
+
"loss": 33.2415,
|
| 2087 |
+
"step": 27400
|
| 2088 |
+
},
|
| 2089 |
+
{
|
| 2090 |
+
"epoch": 21.0,
|
| 2091 |
+
"eval_loss": 34.59661865234375,
|
| 2092 |
+
"eval_runtime": 49.3345,
|
| 2093 |
+
"eval_samples_per_second": 26.452,
|
| 2094 |
+
"eval_steps_per_second": 3.324,
|
| 2095 |
+
"step": 27405
|
| 2096 |
}
|
| 2097 |
],
|
| 2098 |
"logging_steps": 100,
|
|
|
|
| 2107 |
"early_stopping_threshold": 0.0
|
| 2108 |
},
|
| 2109 |
"attributes": {
|
| 2110 |
+
"early_stopping_patience_counter": 5
|
| 2111 |
}
|
| 2112 |
},
|
| 2113 |
"TrainerControl": {
|
|
|
|
| 2121 |
"attributes": {}
|
| 2122 |
}
|
| 2123 |
},
|
| 2124 |
+
"total_flos": 2.955352896399053e+16,
|
| 2125 |
"train_batch_size": 8,
|
| 2126 |
"trial_name": null,
|
| 2127 |
"trial_params": null
|