"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-113370 β checkpoint-113992}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-113370 β checkpoint-113992}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-113370 β checkpoint-113992}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-113370 β checkpoint-113992}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-113370 β checkpoint-113992}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-113370 β checkpoint-113992}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-113370 β checkpoint-113992}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-113370 β checkpoint-113992}/trainer_state.json +793 -4
- model-bin/finetune/base/{checkpoint-113370 β checkpoint-113992}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630085091.0178604/events.out.tfevents.1630085091.dea8124df033.1391.17 +3 -0
- model-bin/finetune/base/log/1630085509.6253698/events.out.tfevents.1630085509.dea8124df033.1391.19 +3 -0
- model-bin/finetune/base/log/1630085938.0146723/events.out.tfevents.1630085938.dea8124df033.1391.21 +3 -0
- model-bin/finetune/base/log/1630086359.258517/events.out.tfevents.1630086359.dea8124df033.1391.23 +3 -0
- model-bin/finetune/base/log/1630086783.2838607/events.out.tfevents.1630086783.dea8124df033.1391.25 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630085091.dea8124df033.1391.16 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630085509.dea8124df033.1391.18 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630085938.dea8124df033.1391.20 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630086359.dea8124df033.1391.22 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630086783.dea8124df033.1391.24 +3 -0
model-bin/finetune/base/{checkpoint-113370 β checkpoint-113992}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-113370 β checkpoint-113992}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0f4f9e0ee9471c71831eaec61932733e15a611cf26ed1b7f271e6474f3f01c9f
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-113370 β checkpoint-113992}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-113370 β checkpoint-113992}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:24e4b1f38df5421de8e487702e1e4f0709dc5d00e9a56d78221675e881e8d441
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-113370 β checkpoint-113992}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:eb8c86e3a58cf24e873bdae44ba1346a463db062b7dde9639caebc5bcd4b49b7
|
| 3 |
size 14503
|
model-bin/finetune/base/{checkpoint-113370 β checkpoint-113992}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1c561868acf735ac6c9b5bb98ac6184be244d99c25305b71d752f027d4438d76
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-113370 β checkpoint-113992}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e9ec5eb205fb1186456b27df8a3a22735db56014ea3ec653c0f4bf9495c4e307
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-113370 β checkpoint-113992}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1743826049391605,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -243609,11 +243609,800 @@
|
|
| 243609 |
"eval_steps_per_second": 0.678,
|
| 243610 |
"eval_wer": 0.18508635386553998,
|
| 243611 |
"step": 113370
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 243612 |
}
|
| 243613 |
],
|
| 243614 |
-
"max_steps":
|
| 243615 |
"num_train_epochs": 5000,
|
| 243616 |
-
"total_flos": 3.
|
| 243617 |
"trial_name": null,
|
| 243618 |
"trial_params": null
|
| 243619 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1743826049391605,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
|
| 4 |
+
"epoch": 911.0,
|
| 5 |
+
"global_step": 113992,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 243609 |
"eval_steps_per_second": 0.678,
|
| 243610 |
"eval_wer": 0.18508635386553998,
|
| 243611 |
"step": 113370
|
| 243612 |
+
},
|
| 243613 |
+
{
|
| 243614 |
+
"epoch": 914.04,
|
| 243615 |
+
"learning_rate": 8.18529886914378e-06,
|
| 243616 |
+
"loss": 0.3386,
|
| 243617 |
+
"step": 113375
|
| 243618 |
+
},
|
| 243619 |
+
{
|
| 243620 |
+
"epoch": 914.08,
|
| 243621 |
+
"learning_rate": 8.185218093699517e-06,
|
| 243622 |
+
"loss": 0.3232,
|
| 243623 |
+
"step": 113380
|
| 243624 |
+
},
|
| 243625 |
+
{
|
| 243626 |
+
"epoch": 914.12,
|
| 243627 |
+
"learning_rate": 8.18513731825525e-06,
|
| 243628 |
+
"loss": 0.3823,
|
| 243629 |
+
"step": 113385
|
| 243630 |
+
},
|
| 243631 |
+
{
|
| 243632 |
+
"epoch": 914.16,
|
| 243633 |
+
"learning_rate": 8.185056542810987e-06,
|
| 243634 |
+
"loss": 0.4816,
|
| 243635 |
+
"step": 113390
|
| 243636 |
+
},
|
| 243637 |
+
{
|
| 243638 |
+
"epoch": 914.2,
|
| 243639 |
+
"learning_rate": 8.18497576736672e-06,
|
| 243640 |
+
"loss": 1.133,
|
| 243641 |
+
"step": 113395
|
| 243642 |
+
},
|
| 243643 |
+
{
|
| 243644 |
+
"epoch": 914.24,
|
| 243645 |
+
"learning_rate": 8.184894991922456e-06,
|
| 243646 |
+
"loss": 0.3591,
|
| 243647 |
+
"step": 113400
|
| 243648 |
+
},
|
| 243649 |
+
{
|
| 243650 |
+
"epoch": 914.28,
|
| 243651 |
+
"learning_rate": 8.184814216478192e-06,
|
| 243652 |
+
"loss": 0.298,
|
| 243653 |
+
"step": 113405
|
| 243654 |
+
},
|
| 243655 |
+
{
|
| 243656 |
+
"epoch": 914.32,
|
| 243657 |
+
"learning_rate": 8.184733441033926e-06,
|
| 243658 |
+
"loss": 0.3137,
|
| 243659 |
+
"step": 113410
|
| 243660 |
+
},
|
| 243661 |
+
{
|
| 243662 |
+
"epoch": 914.36,
|
| 243663 |
+
"learning_rate": 8.184652665589662e-06,
|
| 243664 |
+
"loss": 0.4768,
|
| 243665 |
+
"step": 113415
|
| 243666 |
+
},
|
| 243667 |
+
{
|
| 243668 |
+
"epoch": 914.4,
|
| 243669 |
+
"learning_rate": 8.184571890145396e-06,
|
| 243670 |
+
"loss": 1.1675,
|
| 243671 |
+
"step": 113420
|
| 243672 |
+
},
|
| 243673 |
+
{
|
| 243674 |
+
"epoch": 914.44,
|
| 243675 |
+
"learning_rate": 8.184491114701132e-06,
|
| 243676 |
+
"loss": 0.3729,
|
| 243677 |
+
"step": 113425
|
| 243678 |
+
},
|
| 243679 |
+
{
|
| 243680 |
+
"epoch": 914.48,
|
| 243681 |
+
"learning_rate": 8.184410339256866e-06,
|
| 243682 |
+
"loss": 0.285,
|
| 243683 |
+
"step": 113430
|
| 243684 |
+
},
|
| 243685 |
+
{
|
| 243686 |
+
"epoch": 914.52,
|
| 243687 |
+
"learning_rate": 8.184329563812602e-06,
|
| 243688 |
+
"loss": 0.2975,
|
| 243689 |
+
"step": 113435
|
| 243690 |
+
},
|
| 243691 |
+
{
|
| 243692 |
+
"epoch": 914.56,
|
| 243693 |
+
"learning_rate": 8.184248788368336e-06,
|
| 243694 |
+
"loss": 0.4192,
|
| 243695 |
+
"step": 113440
|
| 243696 |
+
},
|
| 243697 |
+
{
|
| 243698 |
+
"epoch": 914.6,
|
| 243699 |
+
"learning_rate": 8.184168012924072e-06,
|
| 243700 |
+
"loss": 1.1731,
|
| 243701 |
+
"step": 113445
|
| 243702 |
+
},
|
| 243703 |
+
{
|
| 243704 |
+
"epoch": 914.65,
|
| 243705 |
+
"learning_rate": 8.184087237479806e-06,
|
| 243706 |
+
"loss": 0.2925,
|
| 243707 |
+
"step": 113450
|
| 243708 |
+
},
|
| 243709 |
+
{
|
| 243710 |
+
"epoch": 914.69,
|
| 243711 |
+
"learning_rate": 8.184006462035542e-06,
|
| 243712 |
+
"loss": 0.2705,
|
| 243713 |
+
"step": 113455
|
| 243714 |
+
},
|
| 243715 |
+
{
|
| 243716 |
+
"epoch": 914.73,
|
| 243717 |
+
"learning_rate": 8.183925686591276e-06,
|
| 243718 |
+
"loss": 0.303,
|
| 243719 |
+
"step": 113460
|
| 243720 |
+
},
|
| 243721 |
+
{
|
| 243722 |
+
"epoch": 914.77,
|
| 243723 |
+
"learning_rate": 8.183844911147012e-06,
|
| 243724 |
+
"loss": 0.443,
|
| 243725 |
+
"step": 113465
|
| 243726 |
+
},
|
| 243727 |
+
{
|
| 243728 |
+
"epoch": 914.81,
|
| 243729 |
+
"learning_rate": 8.183764135702748e-06,
|
| 243730 |
+
"loss": 1.1133,
|
| 243731 |
+
"step": 113470
|
| 243732 |
+
},
|
| 243733 |
+
{
|
| 243734 |
+
"epoch": 914.85,
|
| 243735 |
+
"learning_rate": 8.183683360258482e-06,
|
| 243736 |
+
"loss": 0.3293,
|
| 243737 |
+
"step": 113475
|
| 243738 |
+
},
|
| 243739 |
+
{
|
| 243740 |
+
"epoch": 914.89,
|
| 243741 |
+
"learning_rate": 8.183602584814218e-06,
|
| 243742 |
+
"loss": 0.2448,
|
| 243743 |
+
"step": 113480
|
| 243744 |
+
},
|
| 243745 |
+
{
|
| 243746 |
+
"epoch": 914.93,
|
| 243747 |
+
"learning_rate": 8.183521809369952e-06,
|
| 243748 |
+
"loss": 0.4102,
|
| 243749 |
+
"step": 113485
|
| 243750 |
+
},
|
| 243751 |
+
{
|
| 243752 |
+
"epoch": 914.97,
|
| 243753 |
+
"learning_rate": 8.183441033925688e-06,
|
| 243754 |
+
"loss": 0.5599,
|
| 243755 |
+
"step": 113490
|
| 243756 |
+
},
|
| 243757 |
+
{
|
| 243758 |
+
"epoch": 915.0,
|
| 243759 |
+
"eval_loss": 0.3741353750228882,
|
| 243760 |
+
"eval_runtime": 40.0199,
|
| 243761 |
+
"eval_samples_per_second": 20.915,
|
| 243762 |
+
"eval_steps_per_second": 0.675,
|
| 243763 |
+
"eval_wer": 0.1836039431719339,
|
| 243764 |
+
"step": 113494
|
| 243765 |
+
},
|
| 243766 |
+
{
|
| 243767 |
+
"epoch": 915.01,
|
| 243768 |
+
"learning_rate": 8.183360258481422e-06,
|
| 243769 |
+
"loss": 0.3819,
|
| 243770 |
+
"step": 113495
|
| 243771 |
+
},
|
| 243772 |
+
{
|
| 243773 |
+
"epoch": 915.05,
|
| 243774 |
+
"learning_rate": 8.183279483037158e-06,
|
| 243775 |
+
"loss": 0.2852,
|
| 243776 |
+
"step": 113500
|
| 243777 |
+
},
|
| 243778 |
+
{
|
| 243779 |
+
"epoch": 915.09,
|
| 243780 |
+
"learning_rate": 8.183198707592892e-06,
|
| 243781 |
+
"loss": 0.3385,
|
| 243782 |
+
"step": 113505
|
| 243783 |
+
},
|
| 243784 |
+
{
|
| 243785 |
+
"epoch": 915.13,
|
| 243786 |
+
"learning_rate": 8.183117932148628e-06,
|
| 243787 |
+
"loss": 0.3233,
|
| 243788 |
+
"step": 113510
|
| 243789 |
+
},
|
| 243790 |
+
{
|
| 243791 |
+
"epoch": 915.17,
|
| 243792 |
+
"learning_rate": 8.183037156704362e-06,
|
| 243793 |
+
"loss": 0.5698,
|
| 243794 |
+
"step": 113515
|
| 243795 |
+
},
|
| 243796 |
+
{
|
| 243797 |
+
"epoch": 915.21,
|
| 243798 |
+
"learning_rate": 8.182956381260098e-06,
|
| 243799 |
+
"loss": 1.2247,
|
| 243800 |
+
"step": 113520
|
| 243801 |
+
},
|
| 243802 |
+
{
|
| 243803 |
+
"epoch": 915.25,
|
| 243804 |
+
"learning_rate": 8.182875605815834e-06,
|
| 243805 |
+
"loss": 0.3235,
|
| 243806 |
+
"step": 113525
|
| 243807 |
+
},
|
| 243808 |
+
{
|
| 243809 |
+
"epoch": 915.29,
|
| 243810 |
+
"learning_rate": 8.182794830371568e-06,
|
| 243811 |
+
"loss": 0.3095,
|
| 243812 |
+
"step": 113530
|
| 243813 |
+
},
|
| 243814 |
+
{
|
| 243815 |
+
"epoch": 915.33,
|
| 243816 |
+
"learning_rate": 8.182714054927304e-06,
|
| 243817 |
+
"loss": 0.323,
|
| 243818 |
+
"step": 113535
|
| 243819 |
+
},
|
| 243820 |
+
{
|
| 243821 |
+
"epoch": 915.37,
|
| 243822 |
+
"learning_rate": 8.182633279483038e-06,
|
| 243823 |
+
"loss": 0.4947,
|
| 243824 |
+
"step": 113540
|
| 243825 |
+
},
|
| 243826 |
+
{
|
| 243827 |
+
"epoch": 915.41,
|
| 243828 |
+
"learning_rate": 8.182552504038774e-06,
|
| 243829 |
+
"loss": 1.2059,
|
| 243830 |
+
"step": 113545
|
| 243831 |
+
},
|
| 243832 |
+
{
|
| 243833 |
+
"epoch": 915.45,
|
| 243834 |
+
"learning_rate": 8.182471728594508e-06,
|
| 243835 |
+
"loss": 0.2672,
|
| 243836 |
+
"step": 113550
|
| 243837 |
+
},
|
| 243838 |
+
{
|
| 243839 |
+
"epoch": 915.49,
|
| 243840 |
+
"learning_rate": 8.182390953150244e-06,
|
| 243841 |
+
"loss": 0.3121,
|
| 243842 |
+
"step": 113555
|
| 243843 |
+
},
|
| 243844 |
+
{
|
| 243845 |
+
"epoch": 915.53,
|
| 243846 |
+
"learning_rate": 8.182310177705978e-06,
|
| 243847 |
+
"loss": 0.2868,
|
| 243848 |
+
"step": 113560
|
| 243849 |
+
},
|
| 243850 |
+
{
|
| 243851 |
+
"epoch": 915.57,
|
| 243852 |
+
"learning_rate": 8.182229402261714e-06,
|
| 243853 |
+
"loss": 0.5263,
|
| 243854 |
+
"step": 113565
|
| 243855 |
+
},
|
| 243856 |
+
{
|
| 243857 |
+
"epoch": 915.61,
|
| 243858 |
+
"learning_rate": 8.182148626817448e-06,
|
| 243859 |
+
"loss": 1.0193,
|
| 243860 |
+
"step": 113570
|
| 243861 |
+
},
|
| 243862 |
+
{
|
| 243863 |
+
"epoch": 915.65,
|
| 243864 |
+
"learning_rate": 8.182067851373184e-06,
|
| 243865 |
+
"loss": 0.3044,
|
| 243866 |
+
"step": 113575
|
| 243867 |
+
},
|
| 243868 |
+
{
|
| 243869 |
+
"epoch": 915.69,
|
| 243870 |
+
"learning_rate": 8.18198707592892e-06,
|
| 243871 |
+
"loss": 0.2556,
|
| 243872 |
+
"step": 113580
|
| 243873 |
+
},
|
| 243874 |
+
{
|
| 243875 |
+
"epoch": 915.73,
|
| 243876 |
+
"learning_rate": 8.181906300484654e-06,
|
| 243877 |
+
"loss": 0.3336,
|
| 243878 |
+
"step": 113585
|
| 243879 |
+
},
|
| 243880 |
+
{
|
| 243881 |
+
"epoch": 915.77,
|
| 243882 |
+
"learning_rate": 8.18182552504039e-06,
|
| 243883 |
+
"loss": 0.5224,
|
| 243884 |
+
"step": 113590
|
| 243885 |
+
},
|
| 243886 |
+
{
|
| 243887 |
+
"epoch": 915.81,
|
| 243888 |
+
"learning_rate": 8.181744749596124e-06,
|
| 243889 |
+
"loss": 1.2317,
|
| 243890 |
+
"step": 113595
|
| 243891 |
+
},
|
| 243892 |
+
{
|
| 243893 |
+
"epoch": 915.85,
|
| 243894 |
+
"learning_rate": 8.18166397415186e-06,
|
| 243895 |
+
"loss": 0.303,
|
| 243896 |
+
"step": 113600
|
| 243897 |
+
},
|
| 243898 |
+
{
|
| 243899 |
+
"epoch": 915.89,
|
| 243900 |
+
"learning_rate": 8.181583198707594e-06,
|
| 243901 |
+
"loss": 0.2783,
|
| 243902 |
+
"step": 113605
|
| 243903 |
+
},
|
| 243904 |
+
{
|
| 243905 |
+
"epoch": 915.93,
|
| 243906 |
+
"learning_rate": 8.18150242326333e-06,
|
| 243907 |
+
"loss": 0.3852,
|
| 243908 |
+
"step": 113610
|
| 243909 |
+
},
|
| 243910 |
+
{
|
| 243911 |
+
"epoch": 915.97,
|
| 243912 |
+
"learning_rate": 8.181421647819063e-06,
|
| 243913 |
+
"loss": 0.4534,
|
| 243914 |
+
"step": 113615
|
| 243915 |
+
},
|
| 243916 |
+
{
|
| 243917 |
+
"epoch": 916.0,
|
| 243918 |
+
"eval_loss": 0.466743141412735,
|
| 243919 |
+
"eval_runtime": 39.7713,
|
| 243920 |
+
"eval_samples_per_second": 21.02,
|
| 243921 |
+
"eval_steps_per_second": 0.679,
|
| 243922 |
+
"eval_wer": 0.18429758141586414,
|
| 243923 |
+
"step": 113618
|
| 243924 |
+
},
|
| 243925 |
+
{
|
| 243926 |
+
"epoch": 908.02,
|
| 243927 |
+
"learning_rate": 8.1813408723748e-06,
|
| 243928 |
+
"loss": 0.3404,
|
| 243929 |
+
"step": 113620
|
| 243930 |
+
},
|
| 243931 |
+
{
|
| 243932 |
+
"epoch": 908.06,
|
| 243933 |
+
"learning_rate": 8.181260096930533e-06,
|
| 243934 |
+
"loss": 0.3143,
|
| 243935 |
+
"step": 113625
|
| 243936 |
+
},
|
| 243937 |
+
{
|
| 243938 |
+
"epoch": 908.1,
|
| 243939 |
+
"learning_rate": 8.18117932148627e-06,
|
| 243940 |
+
"loss": 0.2869,
|
| 243941 |
+
"step": 113630
|
| 243942 |
+
},
|
| 243943 |
+
{
|
| 243944 |
+
"epoch": 908.14,
|
| 243945 |
+
"learning_rate": 8.181098546042003e-06,
|
| 243946 |
+
"loss": 0.3377,
|
| 243947 |
+
"step": 113635
|
| 243948 |
+
},
|
| 243949 |
+
{
|
| 243950 |
+
"epoch": 908.18,
|
| 243951 |
+
"learning_rate": 8.18101777059774e-06,
|
| 243952 |
+
"loss": 0.5821,
|
| 243953 |
+
"step": 113640
|
| 243954 |
+
},
|
| 243955 |
+
{
|
| 243956 |
+
"epoch": 908.22,
|
| 243957 |
+
"learning_rate": 8.180936995153475e-06,
|
| 243958 |
+
"loss": 1.1251,
|
| 243959 |
+
"step": 113645
|
| 243960 |
+
},
|
| 243961 |
+
{
|
| 243962 |
+
"epoch": 908.26,
|
| 243963 |
+
"learning_rate": 8.18085621970921e-06,
|
| 243964 |
+
"loss": 0.327,
|
| 243965 |
+
"step": 113650
|
| 243966 |
+
},
|
| 243967 |
+
{
|
| 243968 |
+
"epoch": 908.3,
|
| 243969 |
+
"learning_rate": 8.180775444264945e-06,
|
| 243970 |
+
"loss": 0.2657,
|
| 243971 |
+
"step": 113655
|
| 243972 |
+
},
|
| 243973 |
+
{
|
| 243974 |
+
"epoch": 908.34,
|
| 243975 |
+
"learning_rate": 8.18069466882068e-06,
|
| 243976 |
+
"loss": 0.3922,
|
| 243977 |
+
"step": 113660
|
| 243978 |
+
},
|
| 243979 |
+
{
|
| 243980 |
+
"epoch": 908.38,
|
| 243981 |
+
"learning_rate": 8.180613893376415e-06,
|
| 243982 |
+
"loss": 0.5555,
|
| 243983 |
+
"step": 113665
|
| 243984 |
+
},
|
| 243985 |
+
{
|
| 243986 |
+
"epoch": 908.42,
|
| 243987 |
+
"learning_rate": 8.18053311793215e-06,
|
| 243988 |
+
"loss": 0.8309,
|
| 243989 |
+
"step": 113670
|
| 243990 |
+
},
|
| 243991 |
+
{
|
| 243992 |
+
"epoch": 908.46,
|
| 243993 |
+
"learning_rate": 8.180452342487885e-06,
|
| 243994 |
+
"loss": 0.3719,
|
| 243995 |
+
"step": 113675
|
| 243996 |
+
},
|
| 243997 |
+
{
|
| 243998 |
+
"epoch": 908.5,
|
| 243999 |
+
"learning_rate": 8.18037156704362e-06,
|
| 244000 |
+
"loss": 0.354,
|
| 244001 |
+
"step": 113680
|
| 244002 |
+
},
|
| 244003 |
+
{
|
| 244004 |
+
"epoch": 908.54,
|
| 244005 |
+
"learning_rate": 8.180290791599355e-06,
|
| 244006 |
+
"loss": 0.3629,
|
| 244007 |
+
"step": 113685
|
| 244008 |
+
},
|
| 244009 |
+
{
|
| 244010 |
+
"epoch": 908.58,
|
| 244011 |
+
"learning_rate": 8.180210016155089e-06,
|
| 244012 |
+
"loss": 0.7741,
|
| 244013 |
+
"step": 113690
|
| 244014 |
+
},
|
| 244015 |
+
{
|
| 244016 |
+
"epoch": 908.62,
|
| 244017 |
+
"learning_rate": 8.180129240710825e-06,
|
| 244018 |
+
"loss": 0.8774,
|
| 244019 |
+
"step": 113695
|
| 244020 |
+
},
|
| 244021 |
+
{
|
| 244022 |
+
"epoch": 908.66,
|
| 244023 |
+
"learning_rate": 8.18004846526656e-06,
|
| 244024 |
+
"loss": 0.2846,
|
| 244025 |
+
"step": 113700
|
| 244026 |
+
},
|
| 244027 |
+
{
|
| 244028 |
+
"epoch": 908.7,
|
| 244029 |
+
"learning_rate": 8.179967689822295e-06,
|
| 244030 |
+
"loss": 0.2622,
|
| 244031 |
+
"step": 113705
|
| 244032 |
+
},
|
| 244033 |
+
{
|
| 244034 |
+
"epoch": 908.74,
|
| 244035 |
+
"learning_rate": 8.17988691437803e-06,
|
| 244036 |
+
"loss": 0.4547,
|
| 244037 |
+
"step": 113710
|
| 244038 |
+
},
|
| 244039 |
+
{
|
| 244040 |
+
"epoch": 908.78,
|
| 244041 |
+
"learning_rate": 8.179806138933765e-06,
|
| 244042 |
+
"loss": 0.6174,
|
| 244043 |
+
"step": 113715
|
| 244044 |
+
},
|
| 244045 |
+
{
|
| 244046 |
+
"epoch": 908.82,
|
| 244047 |
+
"learning_rate": 8.1797253634895e-06,
|
| 244048 |
+
"loss": 1.027,
|
| 244049 |
+
"step": 113720
|
| 244050 |
+
},
|
| 244051 |
+
{
|
| 244052 |
+
"epoch": 908.86,
|
| 244053 |
+
"learning_rate": 8.179644588045235e-06,
|
| 244054 |
+
"loss": 0.2858,
|
| 244055 |
+
"step": 113725
|
| 244056 |
+
},
|
| 244057 |
+
{
|
| 244058 |
+
"epoch": 908.9,
|
| 244059 |
+
"learning_rate": 8.17956381260097e-06,
|
| 244060 |
+
"loss": 0.2602,
|
| 244061 |
+
"step": 113730
|
| 244062 |
+
},
|
| 244063 |
+
{
|
| 244064 |
+
"epoch": 908.94,
|
| 244065 |
+
"learning_rate": 8.179483037156705e-06,
|
| 244066 |
+
"loss": 0.347,
|
| 244067 |
+
"step": 113735
|
| 244068 |
+
},
|
| 244069 |
+
{
|
| 244070 |
+
"epoch": 908.98,
|
| 244071 |
+
"learning_rate": 8.17940226171244e-06,
|
| 244072 |
+
"loss": 0.6511,
|
| 244073 |
+
"step": 113740
|
| 244074 |
+
},
|
| 244075 |
+
{
|
| 244076 |
+
"epoch": 909.0,
|
| 244077 |
+
"eval_loss": 0.4806082248687744,
|
| 244078 |
+
"eval_runtime": 40.3137,
|
| 244079 |
+
"eval_samples_per_second": 20.737,
|
| 244080 |
+
"eval_steps_per_second": 0.67,
|
| 244081 |
+
"eval_wer": 0.19115031129558258,
|
| 244082 |
+
"step": 113743
|
| 244083 |
+
},
|
| 244084 |
+
{
|
| 244085 |
+
"epoch": 917.02,
|
| 244086 |
+
"learning_rate": 8.179321486268175e-06,
|
| 244087 |
+
"loss": 0.3522,
|
| 244088 |
+
"step": 113745
|
| 244089 |
+
},
|
| 244090 |
+
{
|
| 244091 |
+
"epoch": 917.06,
|
| 244092 |
+
"learning_rate": 8.17924071082391e-06,
|
| 244093 |
+
"loss": 0.2977,
|
| 244094 |
+
"step": 113750
|
| 244095 |
+
},
|
| 244096 |
+
{
|
| 244097 |
+
"epoch": 917.1,
|
| 244098 |
+
"learning_rate": 8.179159935379647e-06,
|
| 244099 |
+
"loss": 0.3204,
|
| 244100 |
+
"step": 113755
|
| 244101 |
+
},
|
| 244102 |
+
{
|
| 244103 |
+
"epoch": 917.14,
|
| 244104 |
+
"learning_rate": 8.17907915993538e-06,
|
| 244105 |
+
"loss": 0.4102,
|
| 244106 |
+
"step": 113760
|
| 244107 |
+
},
|
| 244108 |
+
{
|
| 244109 |
+
"epoch": 917.18,
|
| 244110 |
+
"learning_rate": 8.178998384491117e-06,
|
| 244111 |
+
"loss": 0.6249,
|
| 244112 |
+
"step": 113765
|
| 244113 |
+
},
|
| 244114 |
+
{
|
| 244115 |
+
"epoch": 917.22,
|
| 244116 |
+
"learning_rate": 8.17891760904685e-06,
|
| 244117 |
+
"loss": 0.9684,
|
| 244118 |
+
"step": 113770
|
| 244119 |
+
},
|
| 244120 |
+
{
|
| 244121 |
+
"epoch": 917.26,
|
| 244122 |
+
"learning_rate": 8.178836833602586e-06,
|
| 244123 |
+
"loss": 0.3437,
|
| 244124 |
+
"step": 113775
|
| 244125 |
+
},
|
| 244126 |
+
{
|
| 244127 |
+
"epoch": 917.3,
|
| 244128 |
+
"learning_rate": 8.17875605815832e-06,
|
| 244129 |
+
"loss": 0.2964,
|
| 244130 |
+
"step": 113780
|
| 244131 |
+
},
|
| 244132 |
+
{
|
| 244133 |
+
"epoch": 917.34,
|
| 244134 |
+
"learning_rate": 8.178675282714056e-06,
|
| 244135 |
+
"loss": 0.3275,
|
| 244136 |
+
"step": 113785
|
| 244137 |
+
},
|
| 244138 |
+
{
|
| 244139 |
+
"epoch": 917.38,
|
| 244140 |
+
"learning_rate": 8.17859450726979e-06,
|
| 244141 |
+
"loss": 0.6473,
|
| 244142 |
+
"step": 113790
|
| 244143 |
+
},
|
| 244144 |
+
{
|
| 244145 |
+
"epoch": 917.42,
|
| 244146 |
+
"learning_rate": 8.178513731825526e-06,
|
| 244147 |
+
"loss": 1.1345,
|
| 244148 |
+
"step": 113795
|
| 244149 |
+
},
|
| 244150 |
+
{
|
| 244151 |
+
"epoch": 917.46,
|
| 244152 |
+
"learning_rate": 8.17843295638126e-06,
|
| 244153 |
+
"loss": 0.2581,
|
| 244154 |
+
"step": 113800
|
| 244155 |
+
},
|
| 244156 |
+
{
|
| 244157 |
+
"epoch": 917.5,
|
| 244158 |
+
"learning_rate": 8.178352180936996e-06,
|
| 244159 |
+
"loss": 0.3575,
|
| 244160 |
+
"step": 113805
|
| 244161 |
+
},
|
| 244162 |
+
{
|
| 244163 |
+
"epoch": 917.54,
|
| 244164 |
+
"learning_rate": 8.17827140549273e-06,
|
| 244165 |
+
"loss": 0.3293,
|
| 244166 |
+
"step": 113810
|
| 244167 |
+
},
|
| 244168 |
+
{
|
| 244169 |
+
"epoch": 917.58,
|
| 244170 |
+
"learning_rate": 8.178190630048466e-06,
|
| 244171 |
+
"loss": 0.5812,
|
| 244172 |
+
"step": 113815
|
| 244173 |
+
},
|
| 244174 |
+
{
|
| 244175 |
+
"epoch": 917.62,
|
| 244176 |
+
"learning_rate": 8.178109854604202e-06,
|
| 244177 |
+
"loss": 1.0217,
|
| 244178 |
+
"step": 113820
|
| 244179 |
+
},
|
| 244180 |
+
{
|
| 244181 |
+
"epoch": 917.66,
|
| 244182 |
+
"learning_rate": 8.178029079159936e-06,
|
| 244183 |
+
"loss": 0.3646,
|
| 244184 |
+
"step": 113825
|
| 244185 |
+
},
|
| 244186 |
+
{
|
| 244187 |
+
"epoch": 917.7,
|
| 244188 |
+
"learning_rate": 8.177948303715672e-06,
|
| 244189 |
+
"loss": 0.2918,
|
| 244190 |
+
"step": 113830
|
| 244191 |
+
},
|
| 244192 |
+
{
|
| 244193 |
+
"epoch": 917.74,
|
| 244194 |
+
"learning_rate": 8.177867528271406e-06,
|
| 244195 |
+
"loss": 0.3763,
|
| 244196 |
+
"step": 113835
|
| 244197 |
+
},
|
| 244198 |
+
{
|
| 244199 |
+
"epoch": 917.78,
|
| 244200 |
+
"learning_rate": 8.177786752827142e-06,
|
| 244201 |
+
"loss": 0.4992,
|
| 244202 |
+
"step": 113840
|
| 244203 |
+
},
|
| 244204 |
+
{
|
| 244205 |
+
"epoch": 917.82,
|
| 244206 |
+
"learning_rate": 8.177705977382876e-06,
|
| 244207 |
+
"loss": 0.8907,
|
| 244208 |
+
"step": 113845
|
| 244209 |
+
},
|
| 244210 |
+
{
|
| 244211 |
+
"epoch": 917.86,
|
| 244212 |
+
"learning_rate": 8.177625201938612e-06,
|
| 244213 |
+
"loss": 0.2465,
|
| 244214 |
+
"step": 113850
|
| 244215 |
+
},
|
| 244216 |
+
{
|
| 244217 |
+
"epoch": 917.9,
|
| 244218 |
+
"learning_rate": 8.177544426494346e-06,
|
| 244219 |
+
"loss": 0.2878,
|
| 244220 |
+
"step": 113855
|
| 244221 |
+
},
|
| 244222 |
+
{
|
| 244223 |
+
"epoch": 917.94,
|
| 244224 |
+
"learning_rate": 8.177463651050082e-06,
|
| 244225 |
+
"loss": 0.3729,
|
| 244226 |
+
"step": 113860
|
| 244227 |
+
},
|
| 244228 |
+
{
|
| 244229 |
+
"epoch": 917.98,
|
| 244230 |
+
"learning_rate": 8.177382875605816e-06,
|
| 244231 |
+
"loss": 0.6637,
|
| 244232 |
+
"step": 113865
|
| 244233 |
+
},
|
| 244234 |
+
{
|
| 244235 |
+
"epoch": 918.0,
|
| 244236 |
+
"eval_loss": 0.38021937012672424,
|
| 244237 |
+
"eval_runtime": 38.5081,
|
| 244238 |
+
"eval_samples_per_second": 21.71,
|
| 244239 |
+
"eval_steps_per_second": 0.701,
|
| 244240 |
+
"eval_wer": 0.18388491353607633,
|
| 244241 |
+
"step": 113867
|
| 244242 |
+
},
|
| 244243 |
+
{
|
| 244244 |
+
"epoch": 910.02,
|
| 244245 |
+
"learning_rate": 8.177302100161552e-06,
|
| 244246 |
+
"loss": 0.35,
|
| 244247 |
+
"step": 113870
|
| 244248 |
+
},
|
| 244249 |
+
{
|
| 244250 |
+
"epoch": 910.06,
|
| 244251 |
+
"learning_rate": 8.177221324717288e-06,
|
| 244252 |
+
"loss": 0.2942,
|
| 244253 |
+
"step": 113875
|
| 244254 |
+
},
|
| 244255 |
+
{
|
| 244256 |
+
"epoch": 910.1,
|
| 244257 |
+
"learning_rate": 8.177140549273022e-06,
|
| 244258 |
+
"loss": 0.2951,
|
| 244259 |
+
"step": 113880
|
| 244260 |
+
},
|
| 244261 |
+
{
|
| 244262 |
+
"epoch": 910.14,
|
| 244263 |
+
"learning_rate": 8.177059773828758e-06,
|
| 244264 |
+
"loss": 0.3673,
|
| 244265 |
+
"step": 113885
|
| 244266 |
+
},
|
| 244267 |
+
{
|
| 244268 |
+
"epoch": 910.18,
|
| 244269 |
+
"learning_rate": 8.176978998384492e-06,
|
| 244270 |
+
"loss": 0.8055,
|
| 244271 |
+
"step": 113890
|
| 244272 |
+
},
|
| 244273 |
+
{
|
| 244274 |
+
"epoch": 910.22,
|
| 244275 |
+
"learning_rate": 8.176898222940228e-06,
|
| 244276 |
+
"loss": 0.8096,
|
| 244277 |
+
"step": 113895
|
| 244278 |
+
},
|
| 244279 |
+
{
|
| 244280 |
+
"epoch": 910.26,
|
| 244281 |
+
"learning_rate": 8.176817447495962e-06,
|
| 244282 |
+
"loss": 0.2766,
|
| 244283 |
+
"step": 113900
|
| 244284 |
+
},
|
| 244285 |
+
{
|
| 244286 |
+
"epoch": 910.3,
|
| 244287 |
+
"learning_rate": 8.176736672051698e-06,
|
| 244288 |
+
"loss": 0.3135,
|
| 244289 |
+
"step": 113905
|
| 244290 |
+
},
|
| 244291 |
+
{
|
| 244292 |
+
"epoch": 910.34,
|
| 244293 |
+
"learning_rate": 8.176655896607432e-06,
|
| 244294 |
+
"loss": 0.3225,
|
| 244295 |
+
"step": 113910
|
| 244296 |
+
},
|
| 244297 |
+
{
|
| 244298 |
+
"epoch": 910.38,
|
| 244299 |
+
"learning_rate": 8.176575121163168e-06,
|
| 244300 |
+
"loss": 0.684,
|
| 244301 |
+
"step": 113915
|
| 244302 |
+
},
|
| 244303 |
+
{
|
| 244304 |
+
"epoch": 910.42,
|
| 244305 |
+
"learning_rate": 8.176494345718902e-06,
|
| 244306 |
+
"loss": 0.9405,
|
| 244307 |
+
"step": 113920
|
| 244308 |
+
},
|
| 244309 |
+
{
|
| 244310 |
+
"epoch": 910.46,
|
| 244311 |
+
"learning_rate": 8.176413570274638e-06,
|
| 244312 |
+
"loss": 0.3006,
|
| 244313 |
+
"step": 113925
|
| 244314 |
+
},
|
| 244315 |
+
{
|
| 244316 |
+
"epoch": 910.5,
|
| 244317 |
+
"learning_rate": 8.176332794830374e-06,
|
| 244318 |
+
"loss": 0.2658,
|
| 244319 |
+
"step": 113930
|
| 244320 |
+
},
|
| 244321 |
+
{
|
| 244322 |
+
"epoch": 910.54,
|
| 244323 |
+
"learning_rate": 8.176252019386108e-06,
|
| 244324 |
+
"loss": 0.36,
|
| 244325 |
+
"step": 113935
|
| 244326 |
+
},
|
| 244327 |
+
{
|
| 244328 |
+
"epoch": 910.58,
|
| 244329 |
+
"learning_rate": 8.176171243941844e-06,
|
| 244330 |
+
"loss": 0.7644,
|
| 244331 |
+
"step": 113940
|
| 244332 |
+
},
|
| 244333 |
+
{
|
| 244334 |
+
"epoch": 910.62,
|
| 244335 |
+
"learning_rate": 8.176090468497578e-06,
|
| 244336 |
+
"loss": 0.8812,
|
| 244337 |
+
"step": 113945
|
| 244338 |
+
},
|
| 244339 |
+
{
|
| 244340 |
+
"epoch": 910.66,
|
| 244341 |
+
"learning_rate": 8.176009693053314e-06,
|
| 244342 |
+
"loss": 0.2433,
|
| 244343 |
+
"step": 113950
|
| 244344 |
+
},
|
| 244345 |
+
{
|
| 244346 |
+
"epoch": 910.7,
|
| 244347 |
+
"learning_rate": 8.175928917609048e-06,
|
| 244348 |
+
"loss": 0.3455,
|
| 244349 |
+
"step": 113955
|
| 244350 |
+
},
|
| 244351 |
+
{
|
| 244352 |
+
"epoch": 910.74,
|
| 244353 |
+
"learning_rate": 8.175848142164784e-06,
|
| 244354 |
+
"loss": 0.3529,
|
| 244355 |
+
"step": 113960
|
| 244356 |
+
},
|
| 244357 |
+
{
|
| 244358 |
+
"epoch": 910.78,
|
| 244359 |
+
"learning_rate": 8.175767366720518e-06,
|
| 244360 |
+
"loss": 0.7352,
|
| 244361 |
+
"step": 113965
|
| 244362 |
+
},
|
| 244363 |
+
{
|
| 244364 |
+
"epoch": 910.82,
|
| 244365 |
+
"learning_rate": 8.175686591276254e-06,
|
| 244366 |
+
"loss": 0.9688,
|
| 244367 |
+
"step": 113970
|
| 244368 |
+
},
|
| 244369 |
+
{
|
| 244370 |
+
"epoch": 910.86,
|
| 244371 |
+
"learning_rate": 8.175605815831988e-06,
|
| 244372 |
+
"loss": 0.258,
|
| 244373 |
+
"step": 113975
|
| 244374 |
+
},
|
| 244375 |
+
{
|
| 244376 |
+
"epoch": 910.9,
|
| 244377 |
+
"learning_rate": 8.175525040387724e-06,
|
| 244378 |
+
"loss": 0.3258,
|
| 244379 |
+
"step": 113980
|
| 244380 |
+
},
|
| 244381 |
+
{
|
| 244382 |
+
"epoch": 910.94,
|
| 244383 |
+
"learning_rate": 8.175444264943458e-06,
|
| 244384 |
+
"loss": 0.352,
|
| 244385 |
+
"step": 113985
|
| 244386 |
+
},
|
| 244387 |
+
{
|
| 244388 |
+
"epoch": 910.98,
|
| 244389 |
+
"learning_rate": 8.175363489499193e-06,
|
| 244390 |
+
"loss": 0.7014,
|
| 244391 |
+
"step": 113990
|
| 244392 |
+
},
|
| 244393 |
+
{
|
| 244394 |
+
"epoch": 911.0,
|
| 244395 |
+
"eval_loss": 0.3746826946735382,
|
| 244396 |
+
"eval_runtime": 40.8605,
|
| 244397 |
+
"eval_samples_per_second": 20.46,
|
| 244398 |
+
"eval_steps_per_second": 0.661,
|
| 244399 |
+
"eval_wer": 0.18297995933778682,
|
| 244400 |
+
"step": 113992
|
| 244401 |
}
|
| 244402 |
],
|
| 244403 |
+
"max_steps": 625000,
|
| 244404 |
"num_train_epochs": 5000,
|
| 244405 |
+
"total_flos": 3.208038482122093e+20,
|
| 244406 |
"trial_name": null,
|
| 244407 |
"trial_params": null
|
| 244408 |
}
|
model-bin/finetune/base/{checkpoint-113370 β checkpoint-113992}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1630085091.0178604/events.out.tfevents.1630085091.dea8124df033.1391.17
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:724fa55c26a31b6861c29464b4aa44e1153ca653da766200cd07b69901843734
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630085509.6253698/events.out.tfevents.1630085509.dea8124df033.1391.19
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e19ed8f3c9103b8c5207f08be039cd324847ae27fba4ec71b43af202f23eecda
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630085938.0146723/events.out.tfevents.1630085938.dea8124df033.1391.21
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6648f0f003792cbf7efb5a8b9943d6758b19291935312169c29dabc59e082d74
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630086359.258517/events.out.tfevents.1630086359.dea8124df033.1391.23
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bf37a00de48407cdbb55345e5c8f388ec1f64d0df3334d272a8bac197445a8ef
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630086783.2838607/events.out.tfevents.1630086783.dea8124df033.1391.25
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c61e3cea3725234da7bb2057146da95fae81d7e869f1e0696bdbae131acc1f8d
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630085091.dea8124df033.1391.16
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bd6c54ecfce5fc0db85f947ad564150626be2eb0a47260e5df8d743898f17576
|
| 3 |
+
size 8462
|
model-bin/finetune/base/log/events.out.tfevents.1630085509.dea8124df033.1391.18
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:58560bed31557c072a19c26fd8177a2df7bd1cb03a2f4fabaf39733f27b49500
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630085938.dea8124df033.1391.20
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6902e15a463f0d14113fe267221727da5c9630167708808ac6153c2c032d4f05
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630086359.dea8124df033.1391.22
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6bc157a93f6a0c1fbb99622f41cd7f704f39c3c0398f6a8cd5729338879cad51
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630086783.dea8124df033.1391.24
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c2a20f8962da390a8f29ada856c0d68c828b0ed1141d6c57fca8ba02669d1cf1
|
| 3 |
+
size 8622
|