"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-53261 β checkpoint-53883}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-53261 β checkpoint-53883}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-53261 β checkpoint-53883}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-53261 β checkpoint-53883}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-53261 β checkpoint-53883}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-53261 β checkpoint-53883}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-53261 β checkpoint-53883}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-53261 β checkpoint-53883}/trainer_state.json +793 -4
- model-bin/finetune/base/{checkpoint-53261 β checkpoint-53883}/training_args.bin +0 -0
- model-bin/finetune/base/log/1629758991.8110738/events.out.tfevents.1629758991.74272264b15c.932.243 +3 -0
- model-bin/finetune/base/log/1629759638.1886127/events.out.tfevents.1629759638.74272264b15c.932.245 +3 -0
- model-bin/finetune/base/log/1629760286.003077/events.out.tfevents.1629760286.74272264b15c.932.247 +3 -0
- model-bin/finetune/base/log/1629760955.239782/events.out.tfevents.1629760955.74272264b15c.932.249 +3 -0
- model-bin/finetune/base/log/1629761608.7262728/events.out.tfevents.1629761608.74272264b15c.932.251 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629758991.74272264b15c.932.242 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629759638.74272264b15c.932.244 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629760285.74272264b15c.932.246 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629760955.74272264b15c.932.248 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629761608.74272264b15c.932.250 +3 -0
model-bin/finetune/base/{checkpoint-53261 β checkpoint-53883}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-53261 β checkpoint-53883}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165009
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:54045d565c66c3e5b8018fc3afcc2c41061b9918603d0883df91a655d33fee22
|
| 3 |
size 722165009
|
model-bin/finetune/base/{checkpoint-53261 β checkpoint-53883}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-53261 β checkpoint-53883}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1d25122eb873e0266a53d0c13a8c157bfbcf84c44543a60b0d7a792d97aeb34d
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-53261 β checkpoint-53883}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14439
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:303bcbfcafc202ca744ae4ed8003b741b08ac41373010724ddcd8ab1d77d651a
|
| 3 |
size 14439
|
model-bin/finetune/base/{checkpoint-53261 β checkpoint-53883}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:517b8e9913beb30b2ae33d6974fd6fee1e6f87a69ee2028115af6866d0917bf4
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-53261 β checkpoint-53883}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:41f48a461a4300556f9a200eddd4d70732e267633a4abe25cef3f92554b7b8c0
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-53261 β checkpoint-53883}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.18992848189928482,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-52515",
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -167130,11 +167130,800 @@
|
|
| 167130 |
"eval_steps_per_second": 0.69,
|
| 167131 |
"eval_wer": 0.19746457867263237,
|
| 167132 |
"step": 53261
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 167133 |
}
|
| 167134 |
],
|
| 167135 |
-
"max_steps":
|
| 167136 |
"num_train_epochs": 5000,
|
| 167137 |
-
"total_flos": 1.
|
| 167138 |
"trial_name": null,
|
| 167139 |
"trial_params": null
|
| 167140 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.18992848189928482,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-52515",
|
| 4 |
+
"epoch": 431.0,
|
| 5 |
+
"global_step": 53883,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 167130 |
"eval_steps_per_second": 0.69,
|
| 167131 |
"eval_wer": 0.19746457867263237,
|
| 167132 |
"step": 53261
|
| 167133 |
+
},
|
| 167134 |
+
{
|
| 167135 |
+
"epoch": 429.03,
|
| 167136 |
+
"learning_rate": 9.162724358974359e-06,
|
| 167137 |
+
"loss": 0.4907,
|
| 167138 |
+
"step": 53265
|
| 167139 |
+
},
|
| 167140 |
+
{
|
| 167141 |
+
"epoch": 429.07,
|
| 167142 |
+
"learning_rate": 9.162644230769232e-06,
|
| 167143 |
+
"loss": 0.413,
|
| 167144 |
+
"step": 53270
|
| 167145 |
+
},
|
| 167146 |
+
{
|
| 167147 |
+
"epoch": 429.11,
|
| 167148 |
+
"learning_rate": 9.162564102564103e-06,
|
| 167149 |
+
"loss": 0.3893,
|
| 167150 |
+
"step": 53275
|
| 167151 |
+
},
|
| 167152 |
+
{
|
| 167153 |
+
"epoch": 429.15,
|
| 167154 |
+
"learning_rate": 9.162483974358975e-06,
|
| 167155 |
+
"loss": 0.4385,
|
| 167156 |
+
"step": 53280
|
| 167157 |
+
},
|
| 167158 |
+
{
|
| 167159 |
+
"epoch": 429.19,
|
| 167160 |
+
"learning_rate": 9.162403846153846e-06,
|
| 167161 |
+
"loss": 1.2454,
|
| 167162 |
+
"step": 53285
|
| 167163 |
+
},
|
| 167164 |
+
{
|
| 167165 |
+
"epoch": 429.23,
|
| 167166 |
+
"learning_rate": 9.162323717948719e-06,
|
| 167167 |
+
"loss": 0.7755,
|
| 167168 |
+
"step": 53290
|
| 167169 |
+
},
|
| 167170 |
+
{
|
| 167171 |
+
"epoch": 429.27,
|
| 167172 |
+
"learning_rate": 9.16224358974359e-06,
|
| 167173 |
+
"loss": 0.3312,
|
| 167174 |
+
"step": 53295
|
| 167175 |
+
},
|
| 167176 |
+
{
|
| 167177 |
+
"epoch": 429.31,
|
| 167178 |
+
"learning_rate": 9.162163461538462e-06,
|
| 167179 |
+
"loss": 0.3771,
|
| 167180 |
+
"step": 53300
|
| 167181 |
+
},
|
| 167182 |
+
{
|
| 167183 |
+
"epoch": 429.35,
|
| 167184 |
+
"learning_rate": 9.162083333333333e-06,
|
| 167185 |
+
"loss": 0.4456,
|
| 167186 |
+
"step": 53305
|
| 167187 |
+
},
|
| 167188 |
+
{
|
| 167189 |
+
"epoch": 429.39,
|
| 167190 |
+
"learning_rate": 9.162003205128206e-06,
|
| 167191 |
+
"loss": 0.9052,
|
| 167192 |
+
"step": 53310
|
| 167193 |
+
},
|
| 167194 |
+
{
|
| 167195 |
+
"epoch": 429.43,
|
| 167196 |
+
"learning_rate": 9.161923076923078e-06,
|
| 167197 |
+
"loss": 0.7603,
|
| 167198 |
+
"step": 53315
|
| 167199 |
+
},
|
| 167200 |
+
{
|
| 167201 |
+
"epoch": 429.47,
|
| 167202 |
+
"learning_rate": 9.161842948717949e-06,
|
| 167203 |
+
"loss": 0.4158,
|
| 167204 |
+
"step": 53320
|
| 167205 |
+
},
|
| 167206 |
+
{
|
| 167207 |
+
"epoch": 429.51,
|
| 167208 |
+
"learning_rate": 9.161762820512822e-06,
|
| 167209 |
+
"loss": 0.3428,
|
| 167210 |
+
"step": 53325
|
| 167211 |
+
},
|
| 167212 |
+
{
|
| 167213 |
+
"epoch": 429.55,
|
| 167214 |
+
"learning_rate": 9.161682692307693e-06,
|
| 167215 |
+
"loss": 0.475,
|
| 167216 |
+
"step": 53330
|
| 167217 |
+
},
|
| 167218 |
+
{
|
| 167219 |
+
"epoch": 429.59,
|
| 167220 |
+
"learning_rate": 9.161602564102565e-06,
|
| 167221 |
+
"loss": 1.054,
|
| 167222 |
+
"step": 53335
|
| 167223 |
+
},
|
| 167224 |
+
{
|
| 167225 |
+
"epoch": 429.63,
|
| 167226 |
+
"learning_rate": 9.161522435897436e-06,
|
| 167227 |
+
"loss": 0.6962,
|
| 167228 |
+
"step": 53340
|
| 167229 |
+
},
|
| 167230 |
+
{
|
| 167231 |
+
"epoch": 429.67,
|
| 167232 |
+
"learning_rate": 9.16144230769231e-06,
|
| 167233 |
+
"loss": 0.3644,
|
| 167234 |
+
"step": 53345
|
| 167235 |
+
},
|
| 167236 |
+
{
|
| 167237 |
+
"epoch": 429.71,
|
| 167238 |
+
"learning_rate": 9.16136217948718e-06,
|
| 167239 |
+
"loss": 0.4176,
|
| 167240 |
+
"step": 53350
|
| 167241 |
+
},
|
| 167242 |
+
{
|
| 167243 |
+
"epoch": 429.76,
|
| 167244 |
+
"learning_rate": 9.161282051282052e-06,
|
| 167245 |
+
"loss": 0.48,
|
| 167246 |
+
"step": 53355
|
| 167247 |
+
},
|
| 167248 |
+
{
|
| 167249 |
+
"epoch": 429.8,
|
| 167250 |
+
"learning_rate": 9.161201923076923e-06,
|
| 167251 |
+
"loss": 0.9505,
|
| 167252 |
+
"step": 53360
|
| 167253 |
+
},
|
| 167254 |
+
{
|
| 167255 |
+
"epoch": 429.84,
|
| 167256 |
+
"learning_rate": 9.161121794871796e-06,
|
| 167257 |
+
"loss": 0.81,
|
| 167258 |
+
"step": 53365
|
| 167259 |
+
},
|
| 167260 |
+
{
|
| 167261 |
+
"epoch": 429.88,
|
| 167262 |
+
"learning_rate": 9.161041666666668e-06,
|
| 167263 |
+
"loss": 0.3829,
|
| 167264 |
+
"step": 53370
|
| 167265 |
+
},
|
| 167266 |
+
{
|
| 167267 |
+
"epoch": 429.92,
|
| 167268 |
+
"learning_rate": 9.160961538461539e-06,
|
| 167269 |
+
"loss": 0.3667,
|
| 167270 |
+
"step": 53375
|
| 167271 |
+
},
|
| 167272 |
+
{
|
| 167273 |
+
"epoch": 429.96,
|
| 167274 |
+
"learning_rate": 9.160881410256412e-06,
|
| 167275 |
+
"loss": 0.4709,
|
| 167276 |
+
"step": 53380
|
| 167277 |
+
},
|
| 167278 |
+
{
|
| 167279 |
+
"epoch": 430.0,
|
| 167280 |
+
"learning_rate": 9.160801282051283e-06,
|
| 167281 |
+
"loss": 1.0048,
|
| 167282 |
+
"step": 53385
|
| 167283 |
+
},
|
| 167284 |
+
{
|
| 167285 |
+
"epoch": 430.0,
|
| 167286 |
+
"eval_loss": 0.4243323802947998,
|
| 167287 |
+
"eval_runtime": 39.8421,
|
| 167288 |
+
"eval_samples_per_second": 21.083,
|
| 167289 |
+
"eval_steps_per_second": 0.678,
|
| 167290 |
+
"eval_wer": 0.20694374414751854,
|
| 167291 |
+
"step": 53385
|
| 167292 |
+
},
|
| 167293 |
+
{
|
| 167294 |
+
"epoch": 430.04,
|
| 167295 |
+
"learning_rate": 9.160721153846155e-06,
|
| 167296 |
+
"loss": 0.3929,
|
| 167297 |
+
"step": 53390
|
| 167298 |
+
},
|
| 167299 |
+
{
|
| 167300 |
+
"epoch": 430.08,
|
| 167301 |
+
"learning_rate": 9.160641025641026e-06,
|
| 167302 |
+
"loss": 0.2955,
|
| 167303 |
+
"step": 53395
|
| 167304 |
+
},
|
| 167305 |
+
{
|
| 167306 |
+
"epoch": 430.12,
|
| 167307 |
+
"learning_rate": 9.1605608974359e-06,
|
| 167308 |
+
"loss": 0.3902,
|
| 167309 |
+
"step": 53400
|
| 167310 |
+
},
|
| 167311 |
+
{
|
| 167312 |
+
"epoch": 430.16,
|
| 167313 |
+
"learning_rate": 9.160480769230769e-06,
|
| 167314 |
+
"loss": 0.4729,
|
| 167315 |
+
"step": 53405
|
| 167316 |
+
},
|
| 167317 |
+
{
|
| 167318 |
+
"epoch": 430.2,
|
| 167319 |
+
"learning_rate": 9.160400641025642e-06,
|
| 167320 |
+
"loss": 1.3877,
|
| 167321 |
+
"step": 53410
|
| 167322 |
+
},
|
| 167323 |
+
{
|
| 167324 |
+
"epoch": 430.24,
|
| 167325 |
+
"learning_rate": 9.160320512820513e-06,
|
| 167326 |
+
"loss": 0.4524,
|
| 167327 |
+
"step": 53415
|
| 167328 |
+
},
|
| 167329 |
+
{
|
| 167330 |
+
"epoch": 430.28,
|
| 167331 |
+
"learning_rate": 9.160240384615385e-06,
|
| 167332 |
+
"loss": 0.3233,
|
| 167333 |
+
"step": 53420
|
| 167334 |
+
},
|
| 167335 |
+
{
|
| 167336 |
+
"epoch": 430.32,
|
| 167337 |
+
"learning_rate": 9.160160256410258e-06,
|
| 167338 |
+
"loss": 0.4162,
|
| 167339 |
+
"step": 53425
|
| 167340 |
+
},
|
| 167341 |
+
{
|
| 167342 |
+
"epoch": 430.36,
|
| 167343 |
+
"learning_rate": 9.160080128205129e-06,
|
| 167344 |
+
"loss": 0.4888,
|
| 167345 |
+
"step": 53430
|
| 167346 |
+
},
|
| 167347 |
+
{
|
| 167348 |
+
"epoch": 430.4,
|
| 167349 |
+
"learning_rate": 9.16e-06,
|
| 167350 |
+
"loss": 1.219,
|
| 167351 |
+
"step": 53435
|
| 167352 |
+
},
|
| 167353 |
+
{
|
| 167354 |
+
"epoch": 430.44,
|
| 167355 |
+
"learning_rate": 9.159919871794872e-06,
|
| 167356 |
+
"loss": 0.3253,
|
| 167357 |
+
"step": 53440
|
| 167358 |
+
},
|
| 167359 |
+
{
|
| 167360 |
+
"epoch": 430.48,
|
| 167361 |
+
"learning_rate": 9.159839743589745e-06,
|
| 167362 |
+
"loss": 0.3146,
|
| 167363 |
+
"step": 53445
|
| 167364 |
+
},
|
| 167365 |
+
{
|
| 167366 |
+
"epoch": 430.52,
|
| 167367 |
+
"learning_rate": 9.159759615384616e-06,
|
| 167368 |
+
"loss": 0.3463,
|
| 167369 |
+
"step": 53450
|
| 167370 |
+
},
|
| 167371 |
+
{
|
| 167372 |
+
"epoch": 430.56,
|
| 167373 |
+
"learning_rate": 9.159679487179488e-06,
|
| 167374 |
+
"loss": 0.4703,
|
| 167375 |
+
"step": 53455
|
| 167376 |
+
},
|
| 167377 |
+
{
|
| 167378 |
+
"epoch": 430.6,
|
| 167379 |
+
"learning_rate": 9.159599358974359e-06,
|
| 167380 |
+
"loss": 1.2574,
|
| 167381 |
+
"step": 53460
|
| 167382 |
+
},
|
| 167383 |
+
{
|
| 167384 |
+
"epoch": 430.64,
|
| 167385 |
+
"learning_rate": 9.159519230769232e-06,
|
| 167386 |
+
"loss": 0.3584,
|
| 167387 |
+
"step": 53465
|
| 167388 |
+
},
|
| 167389 |
+
{
|
| 167390 |
+
"epoch": 430.68,
|
| 167391 |
+
"learning_rate": 9.159439102564103e-06,
|
| 167392 |
+
"loss": 0.3258,
|
| 167393 |
+
"step": 53470
|
| 167394 |
+
},
|
| 167395 |
+
{
|
| 167396 |
+
"epoch": 430.72,
|
| 167397 |
+
"learning_rate": 9.159358974358975e-06,
|
| 167398 |
+
"loss": 0.3546,
|
| 167399 |
+
"step": 53475
|
| 167400 |
+
},
|
| 167401 |
+
{
|
| 167402 |
+
"epoch": 430.76,
|
| 167403 |
+
"learning_rate": 9.159278846153848e-06,
|
| 167404 |
+
"loss": 0.4754,
|
| 167405 |
+
"step": 53480
|
| 167406 |
+
},
|
| 167407 |
+
{
|
| 167408 |
+
"epoch": 430.8,
|
| 167409 |
+
"learning_rate": 9.159198717948719e-06,
|
| 167410 |
+
"loss": 1.2657,
|
| 167411 |
+
"step": 53485
|
| 167412 |
+
},
|
| 167413 |
+
{
|
| 167414 |
+
"epoch": 430.84,
|
| 167415 |
+
"learning_rate": 9.15911858974359e-06,
|
| 167416 |
+
"loss": 0.3547,
|
| 167417 |
+
"step": 53490
|
| 167418 |
+
},
|
| 167419 |
+
{
|
| 167420 |
+
"epoch": 430.88,
|
| 167421 |
+
"learning_rate": 9.159038461538462e-06,
|
| 167422 |
+
"loss": 0.3136,
|
| 167423 |
+
"step": 53495
|
| 167424 |
+
},
|
| 167425 |
+
{
|
| 167426 |
+
"epoch": 430.92,
|
| 167427 |
+
"learning_rate": 9.158958333333335e-06,
|
| 167428 |
+
"loss": 0.397,
|
| 167429 |
+
"step": 53500
|
| 167430 |
+
},
|
| 167431 |
+
{
|
| 167432 |
+
"epoch": 430.96,
|
| 167433 |
+
"learning_rate": 9.158878205128206e-06,
|
| 167434 |
+
"loss": 0.5967,
|
| 167435 |
+
"step": 53505
|
| 167436 |
+
},
|
| 167437 |
+
{
|
| 167438 |
+
"epoch": 431.0,
|
| 167439 |
+
"eval_loss": 0.3771889805793762,
|
| 167440 |
+
"eval_runtime": 39.5555,
|
| 167441 |
+
"eval_samples_per_second": 21.236,
|
| 167442 |
+
"eval_steps_per_second": 0.683,
|
| 167443 |
+
"eval_wer": 0.19051832190518322,
|
| 167444 |
+
"step": 53509
|
| 167445 |
+
},
|
| 167446 |
+
{
|
| 167447 |
+
"epoch": 428.01,
|
| 167448 |
+
"learning_rate": 9.158798076923078e-06,
|
| 167449 |
+
"loss": 0.3878,
|
| 167450 |
+
"step": 53510
|
| 167451 |
+
},
|
| 167452 |
+
{
|
| 167453 |
+
"epoch": 428.05,
|
| 167454 |
+
"learning_rate": 9.158717948717949e-06,
|
| 167455 |
+
"loss": 0.3302,
|
| 167456 |
+
"step": 53515
|
| 167457 |
+
},
|
| 167458 |
+
{
|
| 167459 |
+
"epoch": 428.09,
|
| 167460 |
+
"learning_rate": 9.158637820512822e-06,
|
| 167461 |
+
"loss": 0.3114,
|
| 167462 |
+
"step": 53520
|
| 167463 |
+
},
|
| 167464 |
+
{
|
| 167465 |
+
"epoch": 428.13,
|
| 167466 |
+
"learning_rate": 9.158557692307693e-06,
|
| 167467 |
+
"loss": 0.3538,
|
| 167468 |
+
"step": 53525
|
| 167469 |
+
},
|
| 167470 |
+
{
|
| 167471 |
+
"epoch": 428.17,
|
| 167472 |
+
"learning_rate": 9.158477564102565e-06,
|
| 167473 |
+
"loss": 0.6061,
|
| 167474 |
+
"step": 53530
|
| 167475 |
+
},
|
| 167476 |
+
{
|
| 167477 |
+
"epoch": 428.21,
|
| 167478 |
+
"learning_rate": 9.158397435897438e-06,
|
| 167479 |
+
"loss": 1.3689,
|
| 167480 |
+
"step": 53535
|
| 167481 |
+
},
|
| 167482 |
+
{
|
| 167483 |
+
"epoch": 428.25,
|
| 167484 |
+
"learning_rate": 9.158317307692307e-06,
|
| 167485 |
+
"loss": 0.5001,
|
| 167486 |
+
"step": 53540
|
| 167487 |
+
},
|
| 167488 |
+
{
|
| 167489 |
+
"epoch": 428.29,
|
| 167490 |
+
"learning_rate": 9.15823717948718e-06,
|
| 167491 |
+
"loss": 0.3581,
|
| 167492 |
+
"step": 53545
|
| 167493 |
+
},
|
| 167494 |
+
{
|
| 167495 |
+
"epoch": 428.33,
|
| 167496 |
+
"learning_rate": 9.158157051282052e-06,
|
| 167497 |
+
"loss": 0.335,
|
| 167498 |
+
"step": 53550
|
| 167499 |
+
},
|
| 167500 |
+
{
|
| 167501 |
+
"epoch": 428.37,
|
| 167502 |
+
"learning_rate": 9.158076923076923e-06,
|
| 167503 |
+
"loss": 0.6611,
|
| 167504 |
+
"step": 53555
|
| 167505 |
+
},
|
| 167506 |
+
{
|
| 167507 |
+
"epoch": 428.41,
|
| 167508 |
+
"learning_rate": 9.157996794871795e-06,
|
| 167509 |
+
"loss": 1.2383,
|
| 167510 |
+
"step": 53560
|
| 167511 |
+
},
|
| 167512 |
+
{
|
| 167513 |
+
"epoch": 428.45,
|
| 167514 |
+
"learning_rate": 9.157916666666668e-06,
|
| 167515 |
+
"loss": 0.3176,
|
| 167516 |
+
"step": 53565
|
| 167517 |
+
},
|
| 167518 |
+
{
|
| 167519 |
+
"epoch": 428.49,
|
| 167520 |
+
"learning_rate": 9.157836538461539e-06,
|
| 167521 |
+
"loss": 0.2605,
|
| 167522 |
+
"step": 53570
|
| 167523 |
+
},
|
| 167524 |
+
{
|
| 167525 |
+
"epoch": 428.53,
|
| 167526 |
+
"learning_rate": 9.15775641025641e-06,
|
| 167527 |
+
"loss": 0.3196,
|
| 167528 |
+
"step": 53575
|
| 167529 |
+
},
|
| 167530 |
+
{
|
| 167531 |
+
"epoch": 428.57,
|
| 167532 |
+
"learning_rate": 9.157676282051283e-06,
|
| 167533 |
+
"loss": 0.6539,
|
| 167534 |
+
"step": 53580
|
| 167535 |
+
},
|
| 167536 |
+
{
|
| 167537 |
+
"epoch": 428.61,
|
| 167538 |
+
"learning_rate": 9.157596153846155e-06,
|
| 167539 |
+
"loss": 1.1682,
|
| 167540 |
+
"step": 53585
|
| 167541 |
+
},
|
| 167542 |
+
{
|
| 167543 |
+
"epoch": 428.65,
|
| 167544 |
+
"learning_rate": 9.157516025641026e-06,
|
| 167545 |
+
"loss": 0.338,
|
| 167546 |
+
"step": 53590
|
| 167547 |
+
},
|
| 167548 |
+
{
|
| 167549 |
+
"epoch": 428.69,
|
| 167550 |
+
"learning_rate": 9.157435897435897e-06,
|
| 167551 |
+
"loss": 0.3081,
|
| 167552 |
+
"step": 53595
|
| 167553 |
+
},
|
| 167554 |
+
{
|
| 167555 |
+
"epoch": 428.73,
|
| 167556 |
+
"learning_rate": 9.15735576923077e-06,
|
| 167557 |
+
"loss": 0.3158,
|
| 167558 |
+
"step": 53600
|
| 167559 |
+
},
|
| 167560 |
+
{
|
| 167561 |
+
"epoch": 428.77,
|
| 167562 |
+
"learning_rate": 9.157275641025642e-06,
|
| 167563 |
+
"loss": 0.5656,
|
| 167564 |
+
"step": 53605
|
| 167565 |
+
},
|
| 167566 |
+
{
|
| 167567 |
+
"epoch": 428.81,
|
| 167568 |
+
"learning_rate": 9.157195512820513e-06,
|
| 167569 |
+
"loss": 1.0353,
|
| 167570 |
+
"step": 53610
|
| 167571 |
+
},
|
| 167572 |
+
{
|
| 167573 |
+
"epoch": 428.85,
|
| 167574 |
+
"learning_rate": 9.157115384615385e-06,
|
| 167575 |
+
"loss": 0.3622,
|
| 167576 |
+
"step": 53615
|
| 167577 |
+
},
|
| 167578 |
+
{
|
| 167579 |
+
"epoch": 428.89,
|
| 167580 |
+
"learning_rate": 9.157035256410258e-06,
|
| 167581 |
+
"loss": 0.3066,
|
| 167582 |
+
"step": 53620
|
| 167583 |
+
},
|
| 167584 |
+
{
|
| 167585 |
+
"epoch": 428.93,
|
| 167586 |
+
"learning_rate": 9.156955128205129e-06,
|
| 167587 |
+
"loss": 0.4568,
|
| 167588 |
+
"step": 53625
|
| 167589 |
+
},
|
| 167590 |
+
{
|
| 167591 |
+
"epoch": 428.97,
|
| 167592 |
+
"learning_rate": 9.156875e-06,
|
| 167593 |
+
"loss": 0.5429,
|
| 167594 |
+
"step": 53630
|
| 167595 |
+
},
|
| 167596 |
+
{
|
| 167597 |
+
"epoch": 429.0,
|
| 167598 |
+
"eval_loss": 0.4089108407497406,
|
| 167599 |
+
"eval_runtime": 39.7496,
|
| 167600 |
+
"eval_samples_per_second": 21.132,
|
| 167601 |
+
"eval_steps_per_second": 0.679,
|
| 167602 |
+
"eval_wer": 0.2008918780612618,
|
| 167603 |
+
"step": 53634
|
| 167604 |
+
},
|
| 167605 |
+
{
|
| 167606 |
+
"epoch": 432.01,
|
| 167607 |
+
"learning_rate": 9.156794871794873e-06,
|
| 167608 |
+
"loss": 0.462,
|
| 167609 |
+
"step": 53635
|
| 167610 |
+
},
|
| 167611 |
+
{
|
| 167612 |
+
"epoch": 432.05,
|
| 167613 |
+
"learning_rate": 9.156714743589745e-06,
|
| 167614 |
+
"loss": 0.3044,
|
| 167615 |
+
"step": 53640
|
| 167616 |
+
},
|
| 167617 |
+
{
|
| 167618 |
+
"epoch": 432.09,
|
| 167619 |
+
"learning_rate": 9.156634615384616e-06,
|
| 167620 |
+
"loss": 0.3536,
|
| 167621 |
+
"step": 53645
|
| 167622 |
+
},
|
| 167623 |
+
{
|
| 167624 |
+
"epoch": 432.13,
|
| 167625 |
+
"learning_rate": 9.156554487179487e-06,
|
| 167626 |
+
"loss": 0.3502,
|
| 167627 |
+
"step": 53650
|
| 167628 |
+
},
|
| 167629 |
+
{
|
| 167630 |
+
"epoch": 432.17,
|
| 167631 |
+
"learning_rate": 9.15647435897436e-06,
|
| 167632 |
+
"loss": 0.5653,
|
| 167633 |
+
"step": 53655
|
| 167634 |
+
},
|
| 167635 |
+
{
|
| 167636 |
+
"epoch": 432.21,
|
| 167637 |
+
"learning_rate": 9.156394230769232e-06,
|
| 167638 |
+
"loss": 1.2789,
|
| 167639 |
+
"step": 53660
|
| 167640 |
+
},
|
| 167641 |
+
{
|
| 167642 |
+
"epoch": 432.25,
|
| 167643 |
+
"learning_rate": 9.156314102564103e-06,
|
| 167644 |
+
"loss": 0.3512,
|
| 167645 |
+
"step": 53665
|
| 167646 |
+
},
|
| 167647 |
+
{
|
| 167648 |
+
"epoch": 432.29,
|
| 167649 |
+
"learning_rate": 9.156233974358976e-06,
|
| 167650 |
+
"loss": 0.3391,
|
| 167651 |
+
"step": 53670
|
| 167652 |
+
},
|
| 167653 |
+
{
|
| 167654 |
+
"epoch": 432.33,
|
| 167655 |
+
"learning_rate": 9.156153846153848e-06,
|
| 167656 |
+
"loss": 0.3438,
|
| 167657 |
+
"step": 53675
|
| 167658 |
+
},
|
| 167659 |
+
{
|
| 167660 |
+
"epoch": 432.37,
|
| 167661 |
+
"learning_rate": 9.156073717948719e-06,
|
| 167662 |
+
"loss": 0.5578,
|
| 167663 |
+
"step": 53680
|
| 167664 |
+
},
|
| 167665 |
+
{
|
| 167666 |
+
"epoch": 432.41,
|
| 167667 |
+
"learning_rate": 9.15599358974359e-06,
|
| 167668 |
+
"loss": 1.1817,
|
| 167669 |
+
"step": 53685
|
| 167670 |
+
},
|
| 167671 |
+
{
|
| 167672 |
+
"epoch": 432.45,
|
| 167673 |
+
"learning_rate": 9.155913461538463e-06,
|
| 167674 |
+
"loss": 0.3402,
|
| 167675 |
+
"step": 53690
|
| 167676 |
+
},
|
| 167677 |
+
{
|
| 167678 |
+
"epoch": 432.49,
|
| 167679 |
+
"learning_rate": 9.155833333333333e-06,
|
| 167680 |
+
"loss": 0.2861,
|
| 167681 |
+
"step": 53695
|
| 167682 |
+
},
|
| 167683 |
+
{
|
| 167684 |
+
"epoch": 432.53,
|
| 167685 |
+
"learning_rate": 9.155753205128206e-06,
|
| 167686 |
+
"loss": 0.2979,
|
| 167687 |
+
"step": 53700
|
| 167688 |
+
},
|
| 167689 |
+
{
|
| 167690 |
+
"epoch": 432.57,
|
| 167691 |
+
"learning_rate": 9.155673076923077e-06,
|
| 167692 |
+
"loss": 0.5803,
|
| 167693 |
+
"step": 53705
|
| 167694 |
+
},
|
| 167695 |
+
{
|
| 167696 |
+
"epoch": 432.61,
|
| 167697 |
+
"learning_rate": 9.155592948717949e-06,
|
| 167698 |
+
"loss": 1.1293,
|
| 167699 |
+
"step": 53710
|
| 167700 |
+
},
|
| 167701 |
+
{
|
| 167702 |
+
"epoch": 432.65,
|
| 167703 |
+
"learning_rate": 9.15551282051282e-06,
|
| 167704 |
+
"loss": 0.4187,
|
| 167705 |
+
"step": 53715
|
| 167706 |
+
},
|
| 167707 |
+
{
|
| 167708 |
+
"epoch": 432.69,
|
| 167709 |
+
"learning_rate": 9.155432692307693e-06,
|
| 167710 |
+
"loss": 0.3486,
|
| 167711 |
+
"step": 53720
|
| 167712 |
+
},
|
| 167713 |
+
{
|
| 167714 |
+
"epoch": 432.73,
|
| 167715 |
+
"learning_rate": 9.155352564102565e-06,
|
| 167716 |
+
"loss": 0.4529,
|
| 167717 |
+
"step": 53725
|
| 167718 |
+
},
|
| 167719 |
+
{
|
| 167720 |
+
"epoch": 432.77,
|
| 167721 |
+
"learning_rate": 9.155272435897436e-06,
|
| 167722 |
+
"loss": 0.6183,
|
| 167723 |
+
"step": 53730
|
| 167724 |
+
},
|
| 167725 |
+
{
|
| 167726 |
+
"epoch": 432.81,
|
| 167727 |
+
"learning_rate": 9.155192307692309e-06,
|
| 167728 |
+
"loss": 1.154,
|
| 167729 |
+
"step": 53735
|
| 167730 |
+
},
|
| 167731 |
+
{
|
| 167732 |
+
"epoch": 432.85,
|
| 167733 |
+
"learning_rate": 9.15511217948718e-06,
|
| 167734 |
+
"loss": 0.3461,
|
| 167735 |
+
"step": 53740
|
| 167736 |
+
},
|
| 167737 |
+
{
|
| 167738 |
+
"epoch": 432.9,
|
| 167739 |
+
"learning_rate": 9.155032051282052e-06,
|
| 167740 |
+
"loss": 0.3242,
|
| 167741 |
+
"step": 53745
|
| 167742 |
+
},
|
| 167743 |
+
{
|
| 167744 |
+
"epoch": 432.94,
|
| 167745 |
+
"learning_rate": 9.154951923076923e-06,
|
| 167746 |
+
"loss": 0.4448,
|
| 167747 |
+
"step": 53750
|
| 167748 |
+
},
|
| 167749 |
+
{
|
| 167750 |
+
"epoch": 432.98,
|
| 167751 |
+
"learning_rate": 9.154871794871796e-06,
|
| 167752 |
+
"loss": 0.6214,
|
| 167753 |
+
"step": 53755
|
| 167754 |
+
},
|
| 167755 |
+
{
|
| 167756 |
+
"epoch": 433.0,
|
| 167757 |
+
"eval_loss": 0.45085570216178894,
|
| 167758 |
+
"eval_runtime": 39.3642,
|
| 167759 |
+
"eval_samples_per_second": 21.339,
|
| 167760 |
+
"eval_steps_per_second": 0.686,
|
| 167761 |
+
"eval_wer": 0.2042432392051397,
|
| 167762 |
+
"step": 53758
|
| 167763 |
+
},
|
| 167764 |
+
{
|
| 167765 |
+
"epoch": 430.02,
|
| 167766 |
+
"learning_rate": 9.154791666666667e-06,
|
| 167767 |
+
"loss": 0.3722,
|
| 167768 |
+
"step": 53760
|
| 167769 |
+
},
|
| 167770 |
+
{
|
| 167771 |
+
"epoch": 430.06,
|
| 167772 |
+
"learning_rate": 9.154711538461539e-06,
|
| 167773 |
+
"loss": 0.2811,
|
| 167774 |
+
"step": 53765
|
| 167775 |
+
},
|
| 167776 |
+
{
|
| 167777 |
+
"epoch": 430.1,
|
| 167778 |
+
"learning_rate": 9.154631410256412e-06,
|
| 167779 |
+
"loss": 0.3527,
|
| 167780 |
+
"step": 53770
|
| 167781 |
+
},
|
| 167782 |
+
{
|
| 167783 |
+
"epoch": 430.14,
|
| 167784 |
+
"learning_rate": 9.154551282051283e-06,
|
| 167785 |
+
"loss": 0.4463,
|
| 167786 |
+
"step": 53775
|
| 167787 |
+
},
|
| 167788 |
+
{
|
| 167789 |
+
"epoch": 430.18,
|
| 167790 |
+
"learning_rate": 9.154471153846155e-06,
|
| 167791 |
+
"loss": 0.7272,
|
| 167792 |
+
"step": 53780
|
| 167793 |
+
},
|
| 167794 |
+
{
|
| 167795 |
+
"epoch": 430.22,
|
| 167796 |
+
"learning_rate": 9.154391025641026e-06,
|
| 167797 |
+
"loss": 1.0655,
|
| 167798 |
+
"step": 53785
|
| 167799 |
+
},
|
| 167800 |
+
{
|
| 167801 |
+
"epoch": 430.26,
|
| 167802 |
+
"learning_rate": 9.154310897435899e-06,
|
| 167803 |
+
"loss": 0.3984,
|
| 167804 |
+
"step": 53790
|
| 167805 |
+
},
|
| 167806 |
+
{
|
| 167807 |
+
"epoch": 430.3,
|
| 167808 |
+
"learning_rate": 9.15423076923077e-06,
|
| 167809 |
+
"loss": 0.3087,
|
| 167810 |
+
"step": 53795
|
| 167811 |
+
},
|
| 167812 |
+
{
|
| 167813 |
+
"epoch": 430.34,
|
| 167814 |
+
"learning_rate": 9.154150641025642e-06,
|
| 167815 |
+
"loss": 0.3548,
|
| 167816 |
+
"step": 53800
|
| 167817 |
+
},
|
| 167818 |
+
{
|
| 167819 |
+
"epoch": 430.38,
|
| 167820 |
+
"learning_rate": 9.154070512820513e-06,
|
| 167821 |
+
"loss": 0.6702,
|
| 167822 |
+
"step": 53805
|
| 167823 |
+
},
|
| 167824 |
+
{
|
| 167825 |
+
"epoch": 430.42,
|
| 167826 |
+
"learning_rate": 9.153990384615386e-06,
|
| 167827 |
+
"loss": 1.0987,
|
| 167828 |
+
"step": 53810
|
| 167829 |
+
},
|
| 167830 |
+
{
|
| 167831 |
+
"epoch": 430.46,
|
| 167832 |
+
"learning_rate": 9.153910256410256e-06,
|
| 167833 |
+
"loss": 0.2952,
|
| 167834 |
+
"step": 53815
|
| 167835 |
+
},
|
| 167836 |
+
{
|
| 167837 |
+
"epoch": 430.5,
|
| 167838 |
+
"learning_rate": 9.153830128205129e-06,
|
| 167839 |
+
"loss": 0.3559,
|
| 167840 |
+
"step": 53820
|
| 167841 |
+
},
|
| 167842 |
+
{
|
| 167843 |
+
"epoch": 430.54,
|
| 167844 |
+
"learning_rate": 9.153750000000002e-06,
|
| 167845 |
+
"loss": 0.4175,
|
| 167846 |
+
"step": 53825
|
| 167847 |
+
},
|
| 167848 |
+
{
|
| 167849 |
+
"epoch": 430.58,
|
| 167850 |
+
"learning_rate": 9.153669871794872e-06,
|
| 167851 |
+
"loss": 0.6887,
|
| 167852 |
+
"step": 53830
|
| 167853 |
+
},
|
| 167854 |
+
{
|
| 167855 |
+
"epoch": 430.62,
|
| 167856 |
+
"learning_rate": 9.153589743589745e-06,
|
| 167857 |
+
"loss": 1.087,
|
| 167858 |
+
"step": 53835
|
| 167859 |
+
},
|
| 167860 |
+
{
|
| 167861 |
+
"epoch": 430.66,
|
| 167862 |
+
"learning_rate": 9.153509615384616e-06,
|
| 167863 |
+
"loss": 0.3551,
|
| 167864 |
+
"step": 53840
|
| 167865 |
+
},
|
| 167866 |
+
{
|
| 167867 |
+
"epoch": 430.7,
|
| 167868 |
+
"learning_rate": 9.153429487179487e-06,
|
| 167869 |
+
"loss": 0.4347,
|
| 167870 |
+
"step": 53845
|
| 167871 |
+
},
|
| 167872 |
+
{
|
| 167873 |
+
"epoch": 430.74,
|
| 167874 |
+
"learning_rate": 9.153349358974359e-06,
|
| 167875 |
+
"loss": 0.3289,
|
| 167876 |
+
"step": 53850
|
| 167877 |
+
},
|
| 167878 |
+
{
|
| 167879 |
+
"epoch": 430.78,
|
| 167880 |
+
"learning_rate": 9.153269230769232e-06,
|
| 167881 |
+
"loss": 0.7435,
|
| 167882 |
+
"step": 53855
|
| 167883 |
+
},
|
| 167884 |
+
{
|
| 167885 |
+
"epoch": 430.82,
|
| 167886 |
+
"learning_rate": 9.153189102564103e-06,
|
| 167887 |
+
"loss": 1.0853,
|
| 167888 |
+
"step": 53860
|
| 167889 |
+
},
|
| 167890 |
+
{
|
| 167891 |
+
"epoch": 430.86,
|
| 167892 |
+
"learning_rate": 9.153108974358974e-06,
|
| 167893 |
+
"loss": 0.3564,
|
| 167894 |
+
"step": 53865
|
| 167895 |
+
},
|
| 167896 |
+
{
|
| 167897 |
+
"epoch": 430.9,
|
| 167898 |
+
"learning_rate": 9.153028846153846e-06,
|
| 167899 |
+
"loss": 0.3129,
|
| 167900 |
+
"step": 53870
|
| 167901 |
+
},
|
| 167902 |
+
{
|
| 167903 |
+
"epoch": 430.94,
|
| 167904 |
+
"learning_rate": 9.152948717948719e-06,
|
| 167905 |
+
"loss": 0.392,
|
| 167906 |
+
"step": 53875
|
| 167907 |
+
},
|
| 167908 |
+
{
|
| 167909 |
+
"epoch": 430.98,
|
| 167910 |
+
"learning_rate": 9.15286858974359e-06,
|
| 167911 |
+
"loss": 0.6987,
|
| 167912 |
+
"step": 53880
|
| 167913 |
+
},
|
| 167914 |
+
{
|
| 167915 |
+
"epoch": 431.0,
|
| 167916 |
+
"eval_loss": 0.4261249005794525,
|
| 167917 |
+
"eval_runtime": 40.6123,
|
| 167918 |
+
"eval_samples_per_second": 20.659,
|
| 167919 |
+
"eval_steps_per_second": 0.665,
|
| 167920 |
+
"eval_wer": 0.20484629294755877,
|
| 167921 |
+
"step": 53883
|
| 167922 |
}
|
| 167923 |
],
|
| 167924 |
+
"max_steps": 625000,
|
| 167925 |
"num_train_epochs": 5000,
|
| 167926 |
+
"total_flos": 1.5162302896695742e+20,
|
| 167927 |
"trial_name": null,
|
| 167928 |
"trial_params": null
|
| 167929 |
}
|
model-bin/finetune/base/{checkpoint-53261 β checkpoint-53883}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1629758991.8110738/events.out.tfevents.1629758991.74272264b15c.932.243
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8f0fc3ca94b527b56044fb11636857cb312846428b7d1c25b83e74e109074f54
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629759638.1886127/events.out.tfevents.1629759638.74272264b15c.932.245
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0f6359b4ca72e068cd143972392b8861fb25bb26fff76fad9f39e9ac10e40a02
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629760286.003077/events.out.tfevents.1629760286.74272264b15c.932.247
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ef32d9aba34e34b5e0f7cc8a3c906b2dc71fcd8341702a6b588aeccab8106973
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629760955.239782/events.out.tfevents.1629760955.74272264b15c.932.249
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9f1da4b68b9d4aefb7159d273cd7ec0e3b69f54b12f5804637195205ee331f63
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629761608.7262728/events.out.tfevents.1629761608.74272264b15c.932.251
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1ea57a4f3dc82d4c084a699c376f70748c4c1bbe875f50b1cbaa6c55cf6647e2
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1629758991.74272264b15c.932.242
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:178ab82b89560feef777c069a0f7b3726e6f12a3b5599864d43de710700cc8da
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629759638.74272264b15c.932.244
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b8a51922dc61333f94c718cad295656d8b7c3f913e9cf0b82b99642cb70c2dcc
|
| 3 |
+
size 8462
|
model-bin/finetune/base/log/events.out.tfevents.1629760285.74272264b15c.932.246
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:18426d2d0c2f1b79173d109598f1c9a8059024fc5e4fadf82d5733417fafa301
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629760955.74272264b15c.932.248
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:156683756fdd361565d13aba4dd180679aa27b81624d9bd9a5645718f956c3ec
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629761608.74272264b15c.932.250
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ee1f37541d338cead8ce83ac5eeed746f1b61263f11f9e905fdfcbbf1a96750f
|
| 3 |
+
size 8622
|