"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-142119 β checkpoint-142742}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-142119 β checkpoint-142742}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-142119 β checkpoint-142742}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-142119 β checkpoint-142742}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-142119 β checkpoint-142742}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-142119 β checkpoint-142742}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-142119 β checkpoint-142742}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-142119 β checkpoint-142742}/trainer_state.json +799 -4
- model-bin/finetune/base/{checkpoint-142119 β checkpoint-142742}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630214833.9430623/events.out.tfevents.1630214833.cc93b136ebf5.1086.49 +3 -0
- model-bin/finetune/base/log/1630215268.1180196/events.out.tfevents.1630215268.cc93b136ebf5.1086.51 +3 -0
- model-bin/finetune/base/log/1630215709.9436204/events.out.tfevents.1630215709.cc93b136ebf5.1086.53 +3 -0
- model-bin/finetune/base/log/1630216147.602746/events.out.tfevents.1630216147.cc93b136ebf5.1086.55 +3 -0
- model-bin/finetune/base/log/1630216585.8185782/events.out.tfevents.1630216585.cc93b136ebf5.1086.57 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630214833.cc93b136ebf5.1086.48 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630215268.cc93b136ebf5.1086.50 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630215709.cc93b136ebf5.1086.52 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630216147.cc93b136ebf5.1086.54 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630216585.cc93b136ebf5.1086.56 +3 -0
model-bin/finetune/base/{checkpoint-142119 β checkpoint-142742}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-142119 β checkpoint-142742}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:13943fb05087ba1925fc85cf6d8ccdcd436e67bce7b73387491c31d71a0d9a4d
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-142119 β checkpoint-142742}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-142119 β checkpoint-142742}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:60daade2e90914f426e13eda15e69959b6dfbcf964cfa92071ea99fff71da02f
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-142119 β checkpoint-142742}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e42e08f28ef8b1308862c7a353c1fb5d4b225eeb6cba470f033516a5bfc09a8a
|
| 3 |
size 14503
|
model-bin/finetune/base/{checkpoint-142119 β checkpoint-142742}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:08811403bda30859be9404fa2f69eb552c512d466566eb146d6719d79f4b49f8
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-142119 β checkpoint-142742}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4ef285b8f55ac571b7d259714b492a3bca4226307e36590c40ca749e5758e1e7
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-142119 β checkpoint-142742}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1689111747851003,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-132910",
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -280182,11 +280182,806 @@
|
|
| 280182 |
"eval_steps_per_second": 0.652,
|
| 280183 |
"eval_wer": 0.18064516129032257,
|
| 280184 |
"step": 142119
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 280185 |
}
|
| 280186 |
],
|
| 280187 |
-
"max_steps":
|
| 280188 |
"num_train_epochs": 5000,
|
| 280189 |
-
"total_flos":
|
| 280190 |
"trial_name": null,
|
| 280191 |
"trial_params": null
|
| 280192 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1689111747851003,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-132910",
|
| 4 |
+
"epoch": 1141.0,
|
| 5 |
+
"global_step": 142742,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 280182 |
"eval_steps_per_second": 0.652,
|
| 280183 |
"eval_wer": 0.18064516129032257,
|
| 280184 |
"step": 142119
|
| 280185 |
+
},
|
| 280186 |
+
{
|
| 280187 |
+
"epoch": 1136.01,
|
| 280188 |
+
"learning_rate": 7.721082390953151e-06,
|
| 280189 |
+
"loss": 0.3707,
|
| 280190 |
+
"step": 142120
|
| 280191 |
+
},
|
| 280192 |
+
{
|
| 280193 |
+
"epoch": 1136.05,
|
| 280194 |
+
"learning_rate": 7.721001615508887e-06,
|
| 280195 |
+
"loss": 0.2748,
|
| 280196 |
+
"step": 142125
|
| 280197 |
+
},
|
| 280198 |
+
{
|
| 280199 |
+
"epoch": 1136.09,
|
| 280200 |
+
"learning_rate": 7.720920840064621e-06,
|
| 280201 |
+
"loss": 0.2521,
|
| 280202 |
+
"step": 142130
|
| 280203 |
+
},
|
| 280204 |
+
{
|
| 280205 |
+
"epoch": 1136.13,
|
| 280206 |
+
"learning_rate": 7.720840064620357e-06,
|
| 280207 |
+
"loss": 0.2975,
|
| 280208 |
+
"step": 142135
|
| 280209 |
+
},
|
| 280210 |
+
{
|
| 280211 |
+
"epoch": 1136.17,
|
| 280212 |
+
"learning_rate": 7.720759289176091e-06,
|
| 280213 |
+
"loss": 0.5674,
|
| 280214 |
+
"step": 142140
|
| 280215 |
+
},
|
| 280216 |
+
{
|
| 280217 |
+
"epoch": 1136.21,
|
| 280218 |
+
"learning_rate": 7.720678513731827e-06,
|
| 280219 |
+
"loss": 1.1433,
|
| 280220 |
+
"step": 142145
|
| 280221 |
+
},
|
| 280222 |
+
{
|
| 280223 |
+
"epoch": 1136.25,
|
| 280224 |
+
"learning_rate": 7.720597738287561e-06,
|
| 280225 |
+
"loss": 0.3099,
|
| 280226 |
+
"step": 142150
|
| 280227 |
+
},
|
| 280228 |
+
{
|
| 280229 |
+
"epoch": 1136.29,
|
| 280230 |
+
"learning_rate": 7.720516962843297e-06,
|
| 280231 |
+
"loss": 0.2866,
|
| 280232 |
+
"step": 142155
|
| 280233 |
+
},
|
| 280234 |
+
{
|
| 280235 |
+
"epoch": 1136.33,
|
| 280236 |
+
"learning_rate": 7.720436187399031e-06,
|
| 280237 |
+
"loss": 0.3084,
|
| 280238 |
+
"step": 142160
|
| 280239 |
+
},
|
| 280240 |
+
{
|
| 280241 |
+
"epoch": 1136.37,
|
| 280242 |
+
"learning_rate": 7.720355411954767e-06,
|
| 280243 |
+
"loss": 0.5635,
|
| 280244 |
+
"step": 142165
|
| 280245 |
+
},
|
| 280246 |
+
{
|
| 280247 |
+
"epoch": 1136.41,
|
| 280248 |
+
"learning_rate": 7.720274636510501e-06,
|
| 280249 |
+
"loss": 1.0665,
|
| 280250 |
+
"step": 142170
|
| 280251 |
+
},
|
| 280252 |
+
{
|
| 280253 |
+
"epoch": 1136.45,
|
| 280254 |
+
"learning_rate": 7.720193861066237e-06,
|
| 280255 |
+
"loss": 0.2342,
|
| 280256 |
+
"step": 142175
|
| 280257 |
+
},
|
| 280258 |
+
{
|
| 280259 |
+
"epoch": 1136.49,
|
| 280260 |
+
"learning_rate": 7.720113085621971e-06,
|
| 280261 |
+
"loss": 0.3193,
|
| 280262 |
+
"step": 142180
|
| 280263 |
+
},
|
| 280264 |
+
{
|
| 280265 |
+
"epoch": 1136.53,
|
| 280266 |
+
"learning_rate": 7.720032310177707e-06,
|
| 280267 |
+
"loss": 0.333,
|
| 280268 |
+
"step": 142185
|
| 280269 |
+
},
|
| 280270 |
+
{
|
| 280271 |
+
"epoch": 1136.57,
|
| 280272 |
+
"learning_rate": 7.719951534733441e-06,
|
| 280273 |
+
"loss": 0.4402,
|
| 280274 |
+
"step": 142190
|
| 280275 |
+
},
|
| 280276 |
+
{
|
| 280277 |
+
"epoch": 1136.61,
|
| 280278 |
+
"learning_rate": 7.719870759289177e-06,
|
| 280279 |
+
"loss": 1.1355,
|
| 280280 |
+
"step": 142195
|
| 280281 |
+
},
|
| 280282 |
+
{
|
| 280283 |
+
"epoch": 1136.65,
|
| 280284 |
+
"learning_rate": 7.719789983844913e-06,
|
| 280285 |
+
"loss": 0.2931,
|
| 280286 |
+
"step": 142200
|
| 280287 |
+
},
|
| 280288 |
+
{
|
| 280289 |
+
"epoch": 1136.69,
|
| 280290 |
+
"learning_rate": 7.719709208400647e-06,
|
| 280291 |
+
"loss": 0.2583,
|
| 280292 |
+
"step": 142205
|
| 280293 |
+
},
|
| 280294 |
+
{
|
| 280295 |
+
"epoch": 1136.73,
|
| 280296 |
+
"learning_rate": 7.719628432956383e-06,
|
| 280297 |
+
"loss": 0.3204,
|
| 280298 |
+
"step": 142210
|
| 280299 |
+
},
|
| 280300 |
+
{
|
| 280301 |
+
"epoch": 1136.77,
|
| 280302 |
+
"learning_rate": 7.719547657512117e-06,
|
| 280303 |
+
"loss": 0.4336,
|
| 280304 |
+
"step": 142215
|
| 280305 |
+
},
|
| 280306 |
+
{
|
| 280307 |
+
"epoch": 1136.81,
|
| 280308 |
+
"learning_rate": 7.719466882067853e-06,
|
| 280309 |
+
"loss": 1.0034,
|
| 280310 |
+
"step": 142220
|
| 280311 |
+
},
|
| 280312 |
+
{
|
| 280313 |
+
"epoch": 1136.85,
|
| 280314 |
+
"learning_rate": 7.719386106623587e-06,
|
| 280315 |
+
"loss": 0.2522,
|
| 280316 |
+
"step": 142225
|
| 280317 |
+
},
|
| 280318 |
+
{
|
| 280319 |
+
"epoch": 1136.89,
|
| 280320 |
+
"learning_rate": 7.719305331179323e-06,
|
| 280321 |
+
"loss": 0.2879,
|
| 280322 |
+
"step": 142230
|
| 280323 |
+
},
|
| 280324 |
+
{
|
| 280325 |
+
"epoch": 1136.93,
|
| 280326 |
+
"learning_rate": 7.719224555735057e-06,
|
| 280327 |
+
"loss": 0.317,
|
| 280328 |
+
"step": 142235
|
| 280329 |
+
},
|
| 280330 |
+
{
|
| 280331 |
+
"epoch": 1136.97,
|
| 280332 |
+
"learning_rate": 7.719143780290793e-06,
|
| 280333 |
+
"loss": 0.5506,
|
| 280334 |
+
"step": 142240
|
| 280335 |
+
},
|
| 280336 |
+
{
|
| 280337 |
+
"epoch": 1137.0,
|
| 280338 |
+
"eval_loss": 0.3600503206253052,
|
| 280339 |
+
"eval_runtime": 41.6622,
|
| 280340 |
+
"eval_samples_per_second": 20.114,
|
| 280341 |
+
"eval_steps_per_second": 0.648,
|
| 280342 |
+
"eval_wer": 0.1694229797067957,
|
| 280343 |
+
"step": 142244
|
| 280344 |
+
},
|
| 280345 |
+
{
|
| 280346 |
+
"epoch": 1147.01,
|
| 280347 |
+
"learning_rate": 7.71907915993538e-06,
|
| 280348 |
+
"loss": 0.325,
|
| 280349 |
+
"step": 142245
|
| 280350 |
+
},
|
| 280351 |
+
{
|
| 280352 |
+
"epoch": 1147.05,
|
| 280353 |
+
"learning_rate": 7.718998384491116e-06,
|
| 280354 |
+
"loss": 0.3236,
|
| 280355 |
+
"step": 142250
|
| 280356 |
+
},
|
| 280357 |
+
{
|
| 280358 |
+
"epoch": 1147.09,
|
| 280359 |
+
"learning_rate": 7.71891760904685e-06,
|
| 280360 |
+
"loss": 0.2434,
|
| 280361 |
+
"step": 142255
|
| 280362 |
+
},
|
| 280363 |
+
{
|
| 280364 |
+
"epoch": 1147.13,
|
| 280365 |
+
"learning_rate": 7.718836833602586e-06,
|
| 280366 |
+
"loss": 0.3158,
|
| 280367 |
+
"step": 142260
|
| 280368 |
+
},
|
| 280369 |
+
{
|
| 280370 |
+
"epoch": 1147.17,
|
| 280371 |
+
"learning_rate": 7.718756058158322e-06,
|
| 280372 |
+
"loss": 0.5659,
|
| 280373 |
+
"step": 142265
|
| 280374 |
+
},
|
| 280375 |
+
{
|
| 280376 |
+
"epoch": 1147.21,
|
| 280377 |
+
"learning_rate": 7.718675282714056e-06,
|
| 280378 |
+
"loss": 0.9828,
|
| 280379 |
+
"step": 142270
|
| 280380 |
+
},
|
| 280381 |
+
{
|
| 280382 |
+
"epoch": 1147.25,
|
| 280383 |
+
"learning_rate": 7.718594507269792e-06,
|
| 280384 |
+
"loss": 0.3256,
|
| 280385 |
+
"step": 142275
|
| 280386 |
+
},
|
| 280387 |
+
{
|
| 280388 |
+
"epoch": 1147.29,
|
| 280389 |
+
"learning_rate": 7.718513731825526e-06,
|
| 280390 |
+
"loss": 0.2851,
|
| 280391 |
+
"step": 142280
|
| 280392 |
+
},
|
| 280393 |
+
{
|
| 280394 |
+
"epoch": 1147.33,
|
| 280395 |
+
"learning_rate": 7.718432956381262e-06,
|
| 280396 |
+
"loss": 0.3086,
|
| 280397 |
+
"step": 142285
|
| 280398 |
+
},
|
| 280399 |
+
{
|
| 280400 |
+
"epoch": 1147.37,
|
| 280401 |
+
"learning_rate": 7.718352180936996e-06,
|
| 280402 |
+
"loss": 0.4722,
|
| 280403 |
+
"step": 142290
|
| 280404 |
+
},
|
| 280405 |
+
{
|
| 280406 |
+
"epoch": 1147.41,
|
| 280407 |
+
"learning_rate": 7.718271405492732e-06,
|
| 280408 |
+
"loss": 1.1545,
|
| 280409 |
+
"step": 142295
|
| 280410 |
+
},
|
| 280411 |
+
{
|
| 280412 |
+
"epoch": 1147.45,
|
| 280413 |
+
"learning_rate": 7.718190630048466e-06,
|
| 280414 |
+
"loss": 0.2882,
|
| 280415 |
+
"step": 142300
|
| 280416 |
+
},
|
| 280417 |
+
{
|
| 280418 |
+
"epoch": 1147.49,
|
| 280419 |
+
"learning_rate": 7.718109854604202e-06,
|
| 280420 |
+
"loss": 0.4264,
|
| 280421 |
+
"step": 142305
|
| 280422 |
+
},
|
| 280423 |
+
{
|
| 280424 |
+
"epoch": 1147.53,
|
| 280425 |
+
"learning_rate": 7.718029079159936e-06,
|
| 280426 |
+
"loss": 0.4184,
|
| 280427 |
+
"step": 142310
|
| 280428 |
+
},
|
| 280429 |
+
{
|
| 280430 |
+
"epoch": 1147.57,
|
| 280431 |
+
"learning_rate": 7.717948303715672e-06,
|
| 280432 |
+
"loss": 0.5507,
|
| 280433 |
+
"step": 142315
|
| 280434 |
+
},
|
| 280435 |
+
{
|
| 280436 |
+
"epoch": 1147.61,
|
| 280437 |
+
"learning_rate": 7.717867528271407e-06,
|
| 280438 |
+
"loss": 0.9842,
|
| 280439 |
+
"step": 142320
|
| 280440 |
+
},
|
| 280441 |
+
{
|
| 280442 |
+
"epoch": 1147.65,
|
| 280443 |
+
"learning_rate": 7.717786752827141e-06,
|
| 280444 |
+
"loss": 0.2279,
|
| 280445 |
+
"step": 142325
|
| 280446 |
+
},
|
| 280447 |
+
{
|
| 280448 |
+
"epoch": 1147.69,
|
| 280449 |
+
"learning_rate": 7.717705977382877e-06,
|
| 280450 |
+
"loss": 0.2926,
|
| 280451 |
+
"step": 142330
|
| 280452 |
+
},
|
| 280453 |
+
{
|
| 280454 |
+
"epoch": 1147.73,
|
| 280455 |
+
"learning_rate": 7.717625201938611e-06,
|
| 280456 |
+
"loss": 0.381,
|
| 280457 |
+
"step": 142335
|
| 280458 |
+
},
|
| 280459 |
+
{
|
| 280460 |
+
"epoch": 1147.77,
|
| 280461 |
+
"learning_rate": 7.717544426494347e-06,
|
| 280462 |
+
"loss": 0.4725,
|
| 280463 |
+
"step": 142340
|
| 280464 |
+
},
|
| 280465 |
+
{
|
| 280466 |
+
"epoch": 1147.81,
|
| 280467 |
+
"learning_rate": 7.717463651050081e-06,
|
| 280468 |
+
"loss": 1.0339,
|
| 280469 |
+
"step": 142345
|
| 280470 |
+
},
|
| 280471 |
+
{
|
| 280472 |
+
"epoch": 1147.85,
|
| 280473 |
+
"learning_rate": 7.717382875605817e-06,
|
| 280474 |
+
"loss": 0.3305,
|
| 280475 |
+
"step": 142350
|
| 280476 |
+
},
|
| 280477 |
+
{
|
| 280478 |
+
"epoch": 1147.89,
|
| 280479 |
+
"learning_rate": 7.717302100161551e-06,
|
| 280480 |
+
"loss": 0.2422,
|
| 280481 |
+
"step": 142355
|
| 280482 |
+
},
|
| 280483 |
+
{
|
| 280484 |
+
"epoch": 1147.93,
|
| 280485 |
+
"learning_rate": 7.717221324717287e-06,
|
| 280486 |
+
"loss": 0.3134,
|
| 280487 |
+
"step": 142360
|
| 280488 |
+
},
|
| 280489 |
+
{
|
| 280490 |
+
"epoch": 1147.97,
|
| 280491 |
+
"learning_rate": 7.717140549273021e-06,
|
| 280492 |
+
"loss": 0.5877,
|
| 280493 |
+
"step": 142365
|
| 280494 |
+
},
|
| 280495 |
+
{
|
| 280496 |
+
"epoch": 1148.0,
|
| 280497 |
+
"eval_loss": 0.3255111873149872,
|
| 280498 |
+
"eval_runtime": 41.3411,
|
| 280499 |
+
"eval_samples_per_second": 20.27,
|
| 280500 |
+
"eval_steps_per_second": 0.653,
|
| 280501 |
+
"eval_wer": 0.18185745140388768,
|
| 280502 |
+
"step": 142368
|
| 280503 |
+
},
|
| 280504 |
+
{
|
| 280505 |
+
"epoch": 1148.02,
|
| 280506 |
+
"learning_rate": 7.717059773828757e-06,
|
| 280507 |
+
"loss": 0.3388,
|
| 280508 |
+
"step": 142370
|
| 280509 |
+
},
|
| 280510 |
+
{
|
| 280511 |
+
"epoch": 1148.06,
|
| 280512 |
+
"learning_rate": 7.716978998384493e-06,
|
| 280513 |
+
"loss": 0.2702,
|
| 280514 |
+
"step": 142375
|
| 280515 |
+
},
|
| 280516 |
+
{
|
| 280517 |
+
"epoch": 1148.1,
|
| 280518 |
+
"learning_rate": 7.716898222940227e-06,
|
| 280519 |
+
"loss": 0.2971,
|
| 280520 |
+
"step": 142380
|
| 280521 |
+
},
|
| 280522 |
+
{
|
| 280523 |
+
"epoch": 1148.14,
|
| 280524 |
+
"learning_rate": 7.716817447495963e-06,
|
| 280525 |
+
"loss": 0.328,
|
| 280526 |
+
"step": 142385
|
| 280527 |
+
},
|
| 280528 |
+
{
|
| 280529 |
+
"epoch": 1148.18,
|
| 280530 |
+
"learning_rate": 7.716736672051697e-06,
|
| 280531 |
+
"loss": 0.6229,
|
| 280532 |
+
"step": 142390
|
| 280533 |
+
},
|
| 280534 |
+
{
|
| 280535 |
+
"epoch": 1148.22,
|
| 280536 |
+
"learning_rate": 7.716655896607433e-06,
|
| 280537 |
+
"loss": 1.0692,
|
| 280538 |
+
"step": 142395
|
| 280539 |
+
},
|
| 280540 |
+
{
|
| 280541 |
+
"epoch": 1148.26,
|
| 280542 |
+
"learning_rate": 7.716575121163167e-06,
|
| 280543 |
+
"loss": 0.3869,
|
| 280544 |
+
"step": 142400
|
| 280545 |
+
},
|
| 280546 |
+
{
|
| 280547 |
+
"epoch": 1148.3,
|
| 280548 |
+
"learning_rate": 7.716494345718903e-06,
|
| 280549 |
+
"loss": 0.2883,
|
| 280550 |
+
"step": 142405
|
| 280551 |
+
},
|
| 280552 |
+
{
|
| 280553 |
+
"epoch": 1148.34,
|
| 280554 |
+
"learning_rate": 7.716413570274637e-06,
|
| 280555 |
+
"loss": 0.3579,
|
| 280556 |
+
"step": 142410
|
| 280557 |
+
},
|
| 280558 |
+
{
|
| 280559 |
+
"epoch": 1148.38,
|
| 280560 |
+
"learning_rate": 7.716332794830373e-06,
|
| 280561 |
+
"loss": 0.643,
|
| 280562 |
+
"step": 142415
|
| 280563 |
+
},
|
| 280564 |
+
{
|
| 280565 |
+
"epoch": 1148.42,
|
| 280566 |
+
"learning_rate": 7.716252019386107e-06,
|
| 280567 |
+
"loss": 1.0174,
|
| 280568 |
+
"step": 142420
|
| 280569 |
+
},
|
| 280570 |
+
{
|
| 280571 |
+
"epoch": 1148.46,
|
| 280572 |
+
"learning_rate": 7.716171243941843e-06,
|
| 280573 |
+
"loss": 0.3438,
|
| 280574 |
+
"step": 142425
|
| 280575 |
+
},
|
| 280576 |
+
{
|
| 280577 |
+
"epoch": 1148.5,
|
| 280578 |
+
"learning_rate": 7.716090468497577e-06,
|
| 280579 |
+
"loss": 0.2701,
|
| 280580 |
+
"step": 142430
|
| 280581 |
+
},
|
| 280582 |
+
{
|
| 280583 |
+
"epoch": 1148.54,
|
| 280584 |
+
"learning_rate": 7.716009693053313e-06,
|
| 280585 |
+
"loss": 0.3322,
|
| 280586 |
+
"step": 142435
|
| 280587 |
+
},
|
| 280588 |
+
{
|
| 280589 |
+
"epoch": 1148.58,
|
| 280590 |
+
"learning_rate": 7.715928917609049e-06,
|
| 280591 |
+
"loss": 0.5148,
|
| 280592 |
+
"step": 142440
|
| 280593 |
+
},
|
| 280594 |
+
{
|
| 280595 |
+
"epoch": 1148.62,
|
| 280596 |
+
"learning_rate": 7.715848142164783e-06,
|
| 280597 |
+
"loss": 0.9305,
|
| 280598 |
+
"step": 142445
|
| 280599 |
+
},
|
| 280600 |
+
{
|
| 280601 |
+
"epoch": 1148.66,
|
| 280602 |
+
"learning_rate": 7.715767366720519e-06,
|
| 280603 |
+
"loss": 0.2918,
|
| 280604 |
+
"step": 142450
|
| 280605 |
+
},
|
| 280606 |
+
{
|
| 280607 |
+
"epoch": 1148.7,
|
| 280608 |
+
"learning_rate": 7.715686591276253e-06,
|
| 280609 |
+
"loss": 0.2845,
|
| 280610 |
+
"step": 142455
|
| 280611 |
+
},
|
| 280612 |
+
{
|
| 280613 |
+
"epoch": 1148.74,
|
| 280614 |
+
"learning_rate": 7.715605815831989e-06,
|
| 280615 |
+
"loss": 0.368,
|
| 280616 |
+
"step": 142460
|
| 280617 |
+
},
|
| 280618 |
+
{
|
| 280619 |
+
"epoch": 1148.78,
|
| 280620 |
+
"learning_rate": 7.715525040387723e-06,
|
| 280621 |
+
"loss": 0.6089,
|
| 280622 |
+
"step": 142465
|
| 280623 |
+
},
|
| 280624 |
+
{
|
| 280625 |
+
"epoch": 1148.82,
|
| 280626 |
+
"learning_rate": 7.715444264943459e-06,
|
| 280627 |
+
"loss": 0.993,
|
| 280628 |
+
"step": 142470
|
| 280629 |
+
},
|
| 280630 |
+
{
|
| 280631 |
+
"epoch": 1148.86,
|
| 280632 |
+
"learning_rate": 7.715363489499193e-06,
|
| 280633 |
+
"loss": 0.2292,
|
| 280634 |
+
"step": 142475
|
| 280635 |
+
},
|
| 280636 |
+
{
|
| 280637 |
+
"epoch": 1148.9,
|
| 280638 |
+
"learning_rate": 7.715282714054929e-06,
|
| 280639 |
+
"loss": 0.315,
|
| 280640 |
+
"step": 142480
|
| 280641 |
+
},
|
| 280642 |
+
{
|
| 280643 |
+
"epoch": 1148.94,
|
| 280644 |
+
"learning_rate": 7.715201938610663e-06,
|
| 280645 |
+
"loss": 0.3183,
|
| 280646 |
+
"step": 142485
|
| 280647 |
+
},
|
| 280648 |
+
{
|
| 280649 |
+
"epoch": 1148.98,
|
| 280650 |
+
"learning_rate": 7.715121163166399e-06,
|
| 280651 |
+
"loss": 0.6854,
|
| 280652 |
+
"step": 142490
|
| 280653 |
+
},
|
| 280654 |
+
{
|
| 280655 |
+
"epoch": 1149.0,
|
| 280656 |
+
"eval_loss": 0.35613495111465454,
|
| 280657 |
+
"eval_runtime": 42.22,
|
| 280658 |
+
"eval_samples_per_second": 19.825,
|
| 280659 |
+
"eval_steps_per_second": 0.64,
|
| 280660 |
+
"eval_wer": 0.17750628817822495,
|
| 280661 |
+
"step": 142492
|
| 280662 |
+
},
|
| 280663 |
+
{
|
| 280664 |
+
"epoch": 1139.02,
|
| 280665 |
+
"learning_rate": 7.715040387722133e-06,
|
| 280666 |
+
"loss": 0.3242,
|
| 280667 |
+
"step": 142495
|
| 280668 |
+
},
|
| 280669 |
+
{
|
| 280670 |
+
"epoch": 1139.06,
|
| 280671 |
+
"learning_rate": 7.714959612277869e-06,
|
| 280672 |
+
"loss": 0.3263,
|
| 280673 |
+
"step": 142500
|
| 280674 |
+
},
|
| 280675 |
+
{
|
| 280676 |
+
"epoch": 1139.1,
|
| 280677 |
+
"learning_rate": 7.714878836833604e-06,
|
| 280678 |
+
"loss": 0.2828,
|
| 280679 |
+
"step": 142505
|
| 280680 |
+
},
|
| 280681 |
+
{
|
| 280682 |
+
"epoch": 1139.14,
|
| 280683 |
+
"learning_rate": 7.714798061389339e-06,
|
| 280684 |
+
"loss": 0.343,
|
| 280685 |
+
"step": 142510
|
| 280686 |
+
},
|
| 280687 |
+
{
|
| 280688 |
+
"epoch": 1139.18,
|
| 280689 |
+
"learning_rate": 7.714717285945074e-06,
|
| 280690 |
+
"loss": 0.859,
|
| 280691 |
+
"step": 142515
|
| 280692 |
+
},
|
| 280693 |
+
{
|
| 280694 |
+
"epoch": 1139.22,
|
| 280695 |
+
"learning_rate": 7.714636510500809e-06,
|
| 280696 |
+
"loss": 0.8421,
|
| 280697 |
+
"step": 142520
|
| 280698 |
+
},
|
| 280699 |
+
{
|
| 280700 |
+
"epoch": 1139.26,
|
| 280701 |
+
"learning_rate": 7.714555735056544e-06,
|
| 280702 |
+
"loss": 0.3022,
|
| 280703 |
+
"step": 142525
|
| 280704 |
+
},
|
| 280705 |
+
{
|
| 280706 |
+
"epoch": 1139.3,
|
| 280707 |
+
"learning_rate": 7.714474959612279e-06,
|
| 280708 |
+
"loss": 0.2578,
|
| 280709 |
+
"step": 142530
|
| 280710 |
+
},
|
| 280711 |
+
{
|
| 280712 |
+
"epoch": 1139.34,
|
| 280713 |
+
"learning_rate": 7.714394184168014e-06,
|
| 280714 |
+
"loss": 0.3935,
|
| 280715 |
+
"step": 142535
|
| 280716 |
+
},
|
| 280717 |
+
{
|
| 280718 |
+
"epoch": 1139.38,
|
| 280719 |
+
"learning_rate": 7.714313408723748e-06,
|
| 280720 |
+
"loss": 0.6255,
|
| 280721 |
+
"step": 142540
|
| 280722 |
+
},
|
| 280723 |
+
{
|
| 280724 |
+
"epoch": 1139.42,
|
| 280725 |
+
"learning_rate": 7.714232633279484e-06,
|
| 280726 |
+
"loss": 0.7272,
|
| 280727 |
+
"step": 142545
|
| 280728 |
+
},
|
| 280729 |
+
{
|
| 280730 |
+
"epoch": 1139.46,
|
| 280731 |
+
"learning_rate": 7.714151857835218e-06,
|
| 280732 |
+
"loss": 0.2663,
|
| 280733 |
+
"step": 142550
|
| 280734 |
+
},
|
| 280735 |
+
{
|
| 280736 |
+
"epoch": 1139.5,
|
| 280737 |
+
"learning_rate": 7.714071082390954e-06,
|
| 280738 |
+
"loss": 0.2698,
|
| 280739 |
+
"step": 142555
|
| 280740 |
+
},
|
| 280741 |
+
{
|
| 280742 |
+
"epoch": 1139.54,
|
| 280743 |
+
"learning_rate": 7.713990306946688e-06,
|
| 280744 |
+
"loss": 0.3076,
|
| 280745 |
+
"step": 142560
|
| 280746 |
+
},
|
| 280747 |
+
{
|
| 280748 |
+
"epoch": 1139.58,
|
| 280749 |
+
"learning_rate": 7.713909531502424e-06,
|
| 280750 |
+
"loss": 0.735,
|
| 280751 |
+
"step": 142565
|
| 280752 |
+
},
|
| 280753 |
+
{
|
| 280754 |
+
"epoch": 1139.62,
|
| 280755 |
+
"learning_rate": 7.71382875605816e-06,
|
| 280756 |
+
"loss": 0.8629,
|
| 280757 |
+
"step": 142570
|
| 280758 |
+
},
|
| 280759 |
+
{
|
| 280760 |
+
"epoch": 1139.66,
|
| 280761 |
+
"learning_rate": 7.713747980613894e-06,
|
| 280762 |
+
"loss": 0.3157,
|
| 280763 |
+
"step": 142575
|
| 280764 |
+
},
|
| 280765 |
+
{
|
| 280766 |
+
"epoch": 1139.7,
|
| 280767 |
+
"learning_rate": 7.71366720516963e-06,
|
| 280768 |
+
"loss": 0.2521,
|
| 280769 |
+
"step": 142580
|
| 280770 |
+
},
|
| 280771 |
+
{
|
| 280772 |
+
"epoch": 1139.74,
|
| 280773 |
+
"learning_rate": 7.713586429725364e-06,
|
| 280774 |
+
"loss": 0.3982,
|
| 280775 |
+
"step": 142585
|
| 280776 |
+
},
|
| 280777 |
+
{
|
| 280778 |
+
"epoch": 1139.78,
|
| 280779 |
+
"learning_rate": 7.7135056542811e-06,
|
| 280780 |
+
"loss": 0.7916,
|
| 280781 |
+
"step": 142590
|
| 280782 |
+
},
|
| 280783 |
+
{
|
| 280784 |
+
"epoch": 1139.82,
|
| 280785 |
+
"learning_rate": 7.713424878836834e-06,
|
| 280786 |
+
"loss": 0.9487,
|
| 280787 |
+
"step": 142595
|
| 280788 |
+
},
|
| 280789 |
+
{
|
| 280790 |
+
"epoch": 1139.86,
|
| 280791 |
+
"learning_rate": 7.71334410339257e-06,
|
| 280792 |
+
"loss": 0.2883,
|
| 280793 |
+
"step": 142600
|
| 280794 |
+
},
|
| 280795 |
+
{
|
| 280796 |
+
"epoch": 1139.9,
|
| 280797 |
+
"learning_rate": 7.713263327948304e-06,
|
| 280798 |
+
"loss": 0.4633,
|
| 280799 |
+
"step": 142605
|
| 280800 |
+
},
|
| 280801 |
+
{
|
| 280802 |
+
"epoch": 1139.94,
|
| 280803 |
+
"learning_rate": 7.71318255250404e-06,
|
| 280804 |
+
"loss": 0.3633,
|
| 280805 |
+
"step": 142610
|
| 280806 |
+
},
|
| 280807 |
+
{
|
| 280808 |
+
"epoch": 1139.98,
|
| 280809 |
+
"learning_rate": 7.713101777059774e-06,
|
| 280810 |
+
"loss": 0.7164,
|
| 280811 |
+
"step": 142615
|
| 280812 |
+
},
|
| 280813 |
+
{
|
| 280814 |
+
"epoch": 1140.0,
|
| 280815 |
+
"eval_loss": 0.40156617760658264,
|
| 280816 |
+
"eval_runtime": 41.5385,
|
| 280817 |
+
"eval_samples_per_second": 20.15,
|
| 280818 |
+
"eval_steps_per_second": 0.65,
|
| 280819 |
+
"eval_wer": 0.1809362827956199,
|
| 280820 |
+
"step": 142617
|
| 280821 |
+
},
|
| 280822 |
+
{
|
| 280823 |
+
"epoch": 1140.02,
|
| 280824 |
+
"learning_rate": 7.71302100161551e-06,
|
| 280825 |
+
"loss": 0.3018,
|
| 280826 |
+
"step": 142620
|
| 280827 |
+
},
|
| 280828 |
+
{
|
| 280829 |
+
"epoch": 1140.06,
|
| 280830 |
+
"learning_rate": 7.712940226171244e-06,
|
| 280831 |
+
"loss": 0.2666,
|
| 280832 |
+
"step": 142625
|
| 280833 |
+
},
|
| 280834 |
+
{
|
| 280835 |
+
"epoch": 1140.1,
|
| 280836 |
+
"learning_rate": 7.71285945072698e-06,
|
| 280837 |
+
"loss": 0.2767,
|
| 280838 |
+
"step": 142630
|
| 280839 |
+
},
|
| 280840 |
+
{
|
| 280841 |
+
"epoch": 1140.14,
|
| 280842 |
+
"learning_rate": 7.712778675282716e-06,
|
| 280843 |
+
"loss": 0.3826,
|
| 280844 |
+
"step": 142635
|
| 280845 |
+
},
|
| 280846 |
+
{
|
| 280847 |
+
"epoch": 1140.18,
|
| 280848 |
+
"learning_rate": 7.71269789983845e-06,
|
| 280849 |
+
"loss": 0.6727,
|
| 280850 |
+
"step": 142640
|
| 280851 |
+
},
|
| 280852 |
+
{
|
| 280853 |
+
"epoch": 1140.22,
|
| 280854 |
+
"learning_rate": 7.712633279483039e-06,
|
| 280855 |
+
"loss": 0.8758,
|
| 280856 |
+
"step": 142645
|
| 280857 |
+
},
|
| 280858 |
+
{
|
| 280859 |
+
"epoch": 1140.26,
|
| 280860 |
+
"learning_rate": 7.712552504038773e-06,
|
| 280861 |
+
"loss": 0.3001,
|
| 280862 |
+
"step": 142650
|
| 280863 |
+
},
|
| 280864 |
+
{
|
| 280865 |
+
"epoch": 1140.3,
|
| 280866 |
+
"learning_rate": 7.712471728594509e-06,
|
| 280867 |
+
"loss": 0.31,
|
| 280868 |
+
"step": 142655
|
| 280869 |
+
},
|
| 280870 |
+
{
|
| 280871 |
+
"epoch": 1140.34,
|
| 280872 |
+
"learning_rate": 7.712390953150243e-06,
|
| 280873 |
+
"loss": 0.3155,
|
| 280874 |
+
"step": 142660
|
| 280875 |
+
},
|
| 280876 |
+
{
|
| 280877 |
+
"epoch": 1140.38,
|
| 280878 |
+
"learning_rate": 7.712310177705979e-06,
|
| 280879 |
+
"loss": 0.746,
|
| 280880 |
+
"step": 142665
|
| 280881 |
+
},
|
| 280882 |
+
{
|
| 280883 |
+
"epoch": 1140.42,
|
| 280884 |
+
"learning_rate": 7.712229402261713e-06,
|
| 280885 |
+
"loss": 0.867,
|
| 280886 |
+
"step": 142670
|
| 280887 |
+
},
|
| 280888 |
+
{
|
| 280889 |
+
"epoch": 1140.46,
|
| 280890 |
+
"learning_rate": 7.712148626817449e-06,
|
| 280891 |
+
"loss": 0.261,
|
| 280892 |
+
"step": 142675
|
| 280893 |
+
},
|
| 280894 |
+
{
|
| 280895 |
+
"epoch": 1140.5,
|
| 280896 |
+
"learning_rate": 7.712067851373183e-06,
|
| 280897 |
+
"loss": 0.309,
|
| 280898 |
+
"step": 142680
|
| 280899 |
+
},
|
| 280900 |
+
{
|
| 280901 |
+
"epoch": 1140.54,
|
| 280902 |
+
"learning_rate": 7.711987075928919e-06,
|
| 280903 |
+
"loss": 0.3376,
|
| 280904 |
+
"step": 142685
|
| 280905 |
+
},
|
| 280906 |
+
{
|
| 280907 |
+
"epoch": 1140.58,
|
| 280908 |
+
"learning_rate": 7.711906300484653e-06,
|
| 280909 |
+
"loss": 0.836,
|
| 280910 |
+
"step": 142690
|
| 280911 |
+
},
|
| 280912 |
+
{
|
| 280913 |
+
"epoch": 1140.62,
|
| 280914 |
+
"learning_rate": 7.711825525040389e-06,
|
| 280915 |
+
"loss": 0.8401,
|
| 280916 |
+
"step": 142695
|
| 280917 |
+
},
|
| 280918 |
+
{
|
| 280919 |
+
"epoch": 1140.66,
|
| 280920 |
+
"learning_rate": 7.711744749596125e-06,
|
| 280921 |
+
"loss": 0.2793,
|
| 280922 |
+
"step": 142700
|
| 280923 |
+
},
|
| 280924 |
+
{
|
| 280925 |
+
"epoch": 1140.7,
|
| 280926 |
+
"learning_rate": 7.711663974151859e-06,
|
| 280927 |
+
"loss": 0.3063,
|
| 280928 |
+
"step": 142705
|
| 280929 |
+
},
|
| 280930 |
+
{
|
| 280931 |
+
"epoch": 1140.74,
|
| 280932 |
+
"learning_rate": 7.711583198707595e-06,
|
| 280933 |
+
"loss": 0.3074,
|
| 280934 |
+
"step": 142710
|
| 280935 |
+
},
|
| 280936 |
+
{
|
| 280937 |
+
"epoch": 1140.78,
|
| 280938 |
+
"learning_rate": 7.711502423263329e-06,
|
| 280939 |
+
"loss": 0.7983,
|
| 280940 |
+
"step": 142715
|
| 280941 |
+
},
|
| 280942 |
+
{
|
| 280943 |
+
"epoch": 1140.82,
|
| 280944 |
+
"learning_rate": 7.711421647819065e-06,
|
| 280945 |
+
"loss": 0.8764,
|
| 280946 |
+
"step": 142720
|
| 280947 |
+
},
|
| 280948 |
+
{
|
| 280949 |
+
"epoch": 1140.86,
|
| 280950 |
+
"learning_rate": 7.711340872374799e-06,
|
| 280951 |
+
"loss": 0.2742,
|
| 280952 |
+
"step": 142725
|
| 280953 |
+
},
|
| 280954 |
+
{
|
| 280955 |
+
"epoch": 1140.9,
|
| 280956 |
+
"learning_rate": 7.711260096930535e-06,
|
| 280957 |
+
"loss": 0.3481,
|
| 280958 |
+
"step": 142730
|
| 280959 |
+
},
|
| 280960 |
+
{
|
| 280961 |
+
"epoch": 1140.94,
|
| 280962 |
+
"learning_rate": 7.711179321486269e-06,
|
| 280963 |
+
"loss": 0.3124,
|
| 280964 |
+
"step": 142735
|
| 280965 |
+
},
|
| 280966 |
+
{
|
| 280967 |
+
"epoch": 1140.98,
|
| 280968 |
+
"learning_rate": 7.711098546042005e-06,
|
| 280969 |
+
"loss": 0.8163,
|
| 280970 |
+
"step": 142740
|
| 280971 |
+
},
|
| 280972 |
+
{
|
| 280973 |
+
"epoch": 1141.0,
|
| 280974 |
+
"eval_loss": 0.37549889087677,
|
| 280975 |
+
"eval_runtime": 41.1914,
|
| 280976 |
+
"eval_samples_per_second": 20.32,
|
| 280977 |
+
"eval_steps_per_second": 0.655,
|
| 280978 |
+
"eval_wer": 0.18999632217727105,
|
| 280979 |
+
"step": 142742
|
| 280980 |
}
|
| 280981 |
],
|
| 280982 |
+
"max_steps": 625000,
|
| 280983 |
"num_train_epochs": 5000,
|
| 280984 |
+
"total_flos": 4.0167170527047575e+20,
|
| 280985 |
"trial_name": null,
|
| 280986 |
"trial_params": null
|
| 280987 |
}
|
model-bin/finetune/base/{checkpoint-142119 β checkpoint-142742}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1630214833.9430623/events.out.tfevents.1630214833.cc93b136ebf5.1086.49
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f143c1996764d184ea6e0142485429f172d965a5beee99c8d12635385968f4ba
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630215268.1180196/events.out.tfevents.1630215268.cc93b136ebf5.1086.51
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bfbe1ada41cc7d3cc8cba21ccfef201ccd24706f6294b62d6f1db3b299da27e1
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630215709.9436204/events.out.tfevents.1630215709.cc93b136ebf5.1086.53
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8991206a1452b6cf9c05c982a5f3a8d9c2138da33486b32f4c94a31bc711bcd7
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630216147.602746/events.out.tfevents.1630216147.cc93b136ebf5.1086.55
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:98e3c0e7d6e201e8688302d20444d2a50524d7fa3bcea16ecaf8f14bf5776fdc
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630216585.8185782/events.out.tfevents.1630216585.cc93b136ebf5.1086.57
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4c378405f199baeb6e346553baa7f54b6cb213ebc7c86aaa0482bfa189053278
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630214833.cc93b136ebf5.1086.48
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d86e40343bd585bcbd5d65c3e7326f3a5e490b9bf51dbe05a59f5cc1e3a10ff0
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630215268.cc93b136ebf5.1086.50
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5c6af5f14ffe58fc7865a60458f9244806462798491b9b175869cadbbe303b31
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630215709.cc93b136ebf5.1086.52
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:363d8446ccdaf00a932c78a1481e7c27f004b1c679fdf0ea960c2a0a6816d940
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630216147.cc93b136ebf5.1086.54
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1f0084b3c04544b2030f6f07fd8c6ef86f4137babca34e875305cb7761985d0b
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630216585.cc93b136ebf5.1086.56
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ec5a183ccbf22dba5982b3551c42c7c10552ac78bbc45c8030bd3152945d15f6
|
| 3 |
+
size 8622
|