"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-141498 β checkpoint-142119}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-141498 β checkpoint-142119}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-141498 β checkpoint-142119}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-141498 β checkpoint-142119}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-141498 β checkpoint-142119}/rng_state.pth +2 -2
- model-bin/finetune/base/{checkpoint-141498 β checkpoint-142119}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-141498 β checkpoint-142119}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-141498 β checkpoint-142119}/trainer_state.json +792 -3
- model-bin/finetune/base/{checkpoint-141498 β checkpoint-142119}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630212588.3931246/events.out.tfevents.1630212588.cc93b136ebf5.1086.39 +3 -0
- model-bin/finetune/base/log/1630213036.9606094/events.out.tfevents.1630213036.cc93b136ebf5.1086.41 +3 -0
- model-bin/finetune/base/log/1630213481.1394386/events.out.tfevents.1630213481.cc93b136ebf5.1086.43 +3 -0
- model-bin/finetune/base/log/1630213924.587934/events.out.tfevents.1630213924.cc93b136ebf5.1086.45 +3 -0
- model-bin/finetune/base/log/1630214361.273169/events.out.tfevents.1630214361.cc93b136ebf5.1086.47 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630212588.cc93b136ebf5.1086.38 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630213036.cc93b136ebf5.1086.40 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630213481.cc93b136ebf5.1086.42 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630213924.cc93b136ebf5.1086.44 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630214361.cc93b136ebf5.1086.46 +3 -0
model-bin/finetune/base/{checkpoint-141498 β checkpoint-142119}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-141498 β checkpoint-142119}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fabeaa58e79bc630b4651cb5de62d3becf8f3d011d2a72a8769bd12ad8ec2b2d
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-141498 β checkpoint-142119}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-141498 β checkpoint-142119}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c8ca5c08add489f83f8fbbe54f7967a3267d0c0a25ffcaf253ac3312f999a598
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-141498 β checkpoint-142119}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3ead055a10b6a20dc53b007fa47951821da9955300fa2f176322d948a4b8ef3b
|
| 3 |
+
size 14503
|
model-bin/finetune/base/{checkpoint-141498 β checkpoint-142119}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:435c7949f76b27d1892fd86718a853997f446c51f752754e7e9bab37c34db055
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-141498 β checkpoint-142119}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0e654bbb2cd6ccd57472ee266ae110834ec44311eb0593d1db337b55a5107ef7
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-141498 β checkpoint-142119}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1689111747851003,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-132910",
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -279393,11 +279393,800 @@
|
|
| 279393 |
"eval_steps_per_second": 0.636,
|
| 279394 |
"eval_wer": 0.17664390945023356,
|
| 279395 |
"step": 141498
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 279396 |
}
|
| 279397 |
],
|
| 279398 |
"max_steps": 620000,
|
| 279399 |
"num_train_epochs": 5000,
|
| 279400 |
-
"total_flos": 3.
|
| 279401 |
"trial_name": null,
|
| 279402 |
"trial_params": null
|
| 279403 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1689111747851003,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-132910",
|
| 4 |
+
"epoch": 1146.0,
|
| 5 |
+
"global_step": 142119,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 279393 |
"eval_steps_per_second": 0.636,
|
| 279394 |
"eval_wer": 0.17664390945023356,
|
| 279395 |
"step": 141498
|
| 279396 |
+
},
|
| 279397 |
+
{
|
| 279398 |
+
"epoch": 1131.02,
|
| 279399 |
+
"learning_rate": 7.731098546042004e-06,
|
| 279400 |
+
"loss": 0.4259,
|
| 279401 |
+
"step": 141500
|
| 279402 |
+
},
|
| 279403 |
+
{
|
| 279404 |
+
"epoch": 1131.06,
|
| 279405 |
+
"learning_rate": 7.731017770597739e-06,
|
| 279406 |
+
"loss": 0.2644,
|
| 279407 |
+
"step": 141505
|
| 279408 |
+
},
|
| 279409 |
+
{
|
| 279410 |
+
"epoch": 1131.1,
|
| 279411 |
+
"learning_rate": 7.730936995153474e-06,
|
| 279412 |
+
"loss": 0.2419,
|
| 279413 |
+
"step": 141510
|
| 279414 |
+
},
|
| 279415 |
+
{
|
| 279416 |
+
"epoch": 1131.14,
|
| 279417 |
+
"learning_rate": 7.730856219709208e-06,
|
| 279418 |
+
"loss": 0.3695,
|
| 279419 |
+
"step": 141515
|
| 279420 |
+
},
|
| 279421 |
+
{
|
| 279422 |
+
"epoch": 1131.18,
|
| 279423 |
+
"learning_rate": 7.730775444264944e-06,
|
| 279424 |
+
"loss": 0.5446,
|
| 279425 |
+
"step": 141520
|
| 279426 |
+
},
|
| 279427 |
+
{
|
| 279428 |
+
"epoch": 1131.22,
|
| 279429 |
+
"learning_rate": 7.73069466882068e-06,
|
| 279430 |
+
"loss": 0.948,
|
| 279431 |
+
"step": 141525
|
| 279432 |
+
},
|
| 279433 |
+
{
|
| 279434 |
+
"epoch": 1131.25,
|
| 279435 |
+
"learning_rate": 7.730613893376414e-06,
|
| 279436 |
+
"loss": 0.277,
|
| 279437 |
+
"step": 141530
|
| 279438 |
+
},
|
| 279439 |
+
{
|
| 279440 |
+
"epoch": 1131.29,
|
| 279441 |
+
"learning_rate": 7.73053311793215e-06,
|
| 279442 |
+
"loss": 0.2816,
|
| 279443 |
+
"step": 141535
|
| 279444 |
+
},
|
| 279445 |
+
{
|
| 279446 |
+
"epoch": 1131.33,
|
| 279447 |
+
"learning_rate": 7.730452342487884e-06,
|
| 279448 |
+
"loss": 0.3233,
|
| 279449 |
+
"step": 141540
|
| 279450 |
+
},
|
| 279451 |
+
{
|
| 279452 |
+
"epoch": 1131.37,
|
| 279453 |
+
"learning_rate": 7.73037156704362e-06,
|
| 279454 |
+
"loss": 0.6062,
|
| 279455 |
+
"step": 141545
|
| 279456 |
+
},
|
| 279457 |
+
{
|
| 279458 |
+
"epoch": 1131.41,
|
| 279459 |
+
"learning_rate": 7.730290791599354e-06,
|
| 279460 |
+
"loss": 1.098,
|
| 279461 |
+
"step": 141550
|
| 279462 |
+
},
|
| 279463 |
+
{
|
| 279464 |
+
"epoch": 1131.45,
|
| 279465 |
+
"learning_rate": 7.73021001615509e-06,
|
| 279466 |
+
"loss": 0.2661,
|
| 279467 |
+
"step": 141555
|
| 279468 |
+
},
|
| 279469 |
+
{
|
| 279470 |
+
"epoch": 1131.49,
|
| 279471 |
+
"learning_rate": 7.730129240710824e-06,
|
| 279472 |
+
"loss": 0.3398,
|
| 279473 |
+
"step": 141560
|
| 279474 |
+
},
|
| 279475 |
+
{
|
| 279476 |
+
"epoch": 1131.53,
|
| 279477 |
+
"learning_rate": 7.73004846526656e-06,
|
| 279478 |
+
"loss": 0.329,
|
| 279479 |
+
"step": 141565
|
| 279480 |
+
},
|
| 279481 |
+
{
|
| 279482 |
+
"epoch": 1131.57,
|
| 279483 |
+
"learning_rate": 7.729967689822294e-06,
|
| 279484 |
+
"loss": 0.6156,
|
| 279485 |
+
"step": 141570
|
| 279486 |
+
},
|
| 279487 |
+
{
|
| 279488 |
+
"epoch": 1131.61,
|
| 279489 |
+
"learning_rate": 7.72988691437803e-06,
|
| 279490 |
+
"loss": 1.1026,
|
| 279491 |
+
"step": 141575
|
| 279492 |
+
},
|
| 279493 |
+
{
|
| 279494 |
+
"epoch": 1131.65,
|
| 279495 |
+
"learning_rate": 7.729806138933766e-06,
|
| 279496 |
+
"loss": 0.2513,
|
| 279497 |
+
"step": 141580
|
| 279498 |
+
},
|
| 279499 |
+
{
|
| 279500 |
+
"epoch": 1131.69,
|
| 279501 |
+
"learning_rate": 7.7297253634895e-06,
|
| 279502 |
+
"loss": 0.284,
|
| 279503 |
+
"step": 141585
|
| 279504 |
+
},
|
| 279505 |
+
{
|
| 279506 |
+
"epoch": 1131.73,
|
| 279507 |
+
"learning_rate": 7.729644588045236e-06,
|
| 279508 |
+
"loss": 0.3635,
|
| 279509 |
+
"step": 141590
|
| 279510 |
+
},
|
| 279511 |
+
{
|
| 279512 |
+
"epoch": 1131.77,
|
| 279513 |
+
"learning_rate": 7.72956381260097e-06,
|
| 279514 |
+
"loss": 0.5597,
|
| 279515 |
+
"step": 141595
|
| 279516 |
+
},
|
| 279517 |
+
{
|
| 279518 |
+
"epoch": 1131.81,
|
| 279519 |
+
"learning_rate": 7.729483037156706e-06,
|
| 279520 |
+
"loss": 0.9859,
|
| 279521 |
+
"step": 141600
|
| 279522 |
+
},
|
| 279523 |
+
{
|
| 279524 |
+
"epoch": 1131.85,
|
| 279525 |
+
"learning_rate": 7.72940226171244e-06,
|
| 279526 |
+
"loss": 0.247,
|
| 279527 |
+
"step": 141605
|
| 279528 |
+
},
|
| 279529 |
+
{
|
| 279530 |
+
"epoch": 1131.89,
|
| 279531 |
+
"learning_rate": 7.729321486268176e-06,
|
| 279532 |
+
"loss": 0.2809,
|
| 279533 |
+
"step": 141610
|
| 279534 |
+
},
|
| 279535 |
+
{
|
| 279536 |
+
"epoch": 1131.93,
|
| 279537 |
+
"learning_rate": 7.72924071082391e-06,
|
| 279538 |
+
"loss": 0.3225,
|
| 279539 |
+
"step": 141615
|
| 279540 |
+
},
|
| 279541 |
+
{
|
| 279542 |
+
"epoch": 1131.97,
|
| 279543 |
+
"learning_rate": 7.729159935379646e-06,
|
| 279544 |
+
"loss": 0.6184,
|
| 279545 |
+
"step": 141620
|
| 279546 |
+
},
|
| 279547 |
+
{
|
| 279548 |
+
"epoch": 1132.0,
|
| 279549 |
+
"eval_loss": 0.39268848299980164,
|
| 279550 |
+
"eval_runtime": 42.1482,
|
| 279551 |
+
"eval_samples_per_second": 19.858,
|
| 279552 |
+
"eval_steps_per_second": 0.641,
|
| 279553 |
+
"eval_wer": 0.18137723230690087,
|
| 279554 |
+
"step": 141623
|
| 279555 |
+
},
|
| 279556 |
+
{
|
| 279557 |
+
"epoch": 1142.02,
|
| 279558 |
+
"learning_rate": 7.72907915993538e-06,
|
| 279559 |
+
"loss": 0.3626,
|
| 279560 |
+
"step": 141625
|
| 279561 |
+
},
|
| 279562 |
+
{
|
| 279563 |
+
"epoch": 1142.06,
|
| 279564 |
+
"learning_rate": 7.728998384491116e-06,
|
| 279565 |
+
"loss": 0.3136,
|
| 279566 |
+
"step": 141630
|
| 279567 |
+
},
|
| 279568 |
+
{
|
| 279569 |
+
"epoch": 1142.1,
|
| 279570 |
+
"learning_rate": 7.72891760904685e-06,
|
| 279571 |
+
"loss": 0.2508,
|
| 279572 |
+
"step": 141635
|
| 279573 |
+
},
|
| 279574 |
+
{
|
| 279575 |
+
"epoch": 1142.14,
|
| 279576 |
+
"learning_rate": 7.728836833602586e-06,
|
| 279577 |
+
"loss": 0.3972,
|
| 279578 |
+
"step": 141640
|
| 279579 |
+
},
|
| 279580 |
+
{
|
| 279581 |
+
"epoch": 1142.18,
|
| 279582 |
+
"learning_rate": 7.728756058158322e-06,
|
| 279583 |
+
"loss": 0.6265,
|
| 279584 |
+
"step": 141645
|
| 279585 |
+
},
|
| 279586 |
+
{
|
| 279587 |
+
"epoch": 1142.22,
|
| 279588 |
+
"learning_rate": 7.728675282714056e-06,
|
| 279589 |
+
"loss": 0.9015,
|
| 279590 |
+
"step": 141650
|
| 279591 |
+
},
|
| 279592 |
+
{
|
| 279593 |
+
"epoch": 1142.26,
|
| 279594 |
+
"learning_rate": 7.728594507269792e-06,
|
| 279595 |
+
"loss": 0.2899,
|
| 279596 |
+
"step": 141655
|
| 279597 |
+
},
|
| 279598 |
+
{
|
| 279599 |
+
"epoch": 1142.3,
|
| 279600 |
+
"learning_rate": 7.728513731825526e-06,
|
| 279601 |
+
"loss": 0.3102,
|
| 279602 |
+
"step": 141660
|
| 279603 |
+
},
|
| 279604 |
+
{
|
| 279605 |
+
"epoch": 1142.34,
|
| 279606 |
+
"learning_rate": 7.728432956381261e-06,
|
| 279607 |
+
"loss": 0.3435,
|
| 279608 |
+
"step": 141665
|
| 279609 |
+
},
|
| 279610 |
+
{
|
| 279611 |
+
"epoch": 1142.38,
|
| 279612 |
+
"learning_rate": 7.728352180936996e-06,
|
| 279613 |
+
"loss": 0.6388,
|
| 279614 |
+
"step": 141670
|
| 279615 |
+
},
|
| 279616 |
+
{
|
| 279617 |
+
"epoch": 1142.42,
|
| 279618 |
+
"learning_rate": 7.728271405492731e-06,
|
| 279619 |
+
"loss": 0.9743,
|
| 279620 |
+
"step": 141675
|
| 279621 |
+
},
|
| 279622 |
+
{
|
| 279623 |
+
"epoch": 1142.46,
|
| 279624 |
+
"learning_rate": 7.728190630048466e-06,
|
| 279625 |
+
"loss": 0.2384,
|
| 279626 |
+
"step": 141680
|
| 279627 |
+
},
|
| 279628 |
+
{
|
| 279629 |
+
"epoch": 1142.5,
|
| 279630 |
+
"learning_rate": 7.728109854604201e-06,
|
| 279631 |
+
"loss": 0.4136,
|
| 279632 |
+
"step": 141685
|
| 279633 |
+
},
|
| 279634 |
+
{
|
| 279635 |
+
"epoch": 1142.54,
|
| 279636 |
+
"learning_rate": 7.728029079159936e-06,
|
| 279637 |
+
"loss": 0.3579,
|
| 279638 |
+
"step": 141690
|
| 279639 |
+
},
|
| 279640 |
+
{
|
| 279641 |
+
"epoch": 1142.58,
|
| 279642 |
+
"learning_rate": 7.727948303715671e-06,
|
| 279643 |
+
"loss": 0.5829,
|
| 279644 |
+
"step": 141695
|
| 279645 |
+
},
|
| 279646 |
+
{
|
| 279647 |
+
"epoch": 1142.62,
|
| 279648 |
+
"learning_rate": 7.727867528271407e-06,
|
| 279649 |
+
"loss": 0.9261,
|
| 279650 |
+
"step": 141700
|
| 279651 |
+
},
|
| 279652 |
+
{
|
| 279653 |
+
"epoch": 1142.66,
|
| 279654 |
+
"learning_rate": 7.727786752827141e-06,
|
| 279655 |
+
"loss": 0.2647,
|
| 279656 |
+
"step": 141705
|
| 279657 |
+
},
|
| 279658 |
+
{
|
| 279659 |
+
"epoch": 1142.7,
|
| 279660 |
+
"learning_rate": 7.727705977382877e-06,
|
| 279661 |
+
"loss": 0.2991,
|
| 279662 |
+
"step": 141710
|
| 279663 |
+
},
|
| 279664 |
+
{
|
| 279665 |
+
"epoch": 1142.74,
|
| 279666 |
+
"learning_rate": 7.727625201938611e-06,
|
| 279667 |
+
"loss": 0.3412,
|
| 279668 |
+
"step": 141715
|
| 279669 |
+
},
|
| 279670 |
+
{
|
| 279671 |
+
"epoch": 1142.78,
|
| 279672 |
+
"learning_rate": 7.727544426494347e-06,
|
| 279673 |
+
"loss": 0.5808,
|
| 279674 |
+
"step": 141720
|
| 279675 |
+
},
|
| 279676 |
+
{
|
| 279677 |
+
"epoch": 1142.82,
|
| 279678 |
+
"learning_rate": 7.727463651050081e-06,
|
| 279679 |
+
"loss": 0.9713,
|
| 279680 |
+
"step": 141725
|
| 279681 |
+
},
|
| 279682 |
+
{
|
| 279683 |
+
"epoch": 1142.86,
|
| 279684 |
+
"learning_rate": 7.727382875605817e-06,
|
| 279685 |
+
"loss": 0.2904,
|
| 279686 |
+
"step": 141730
|
| 279687 |
+
},
|
| 279688 |
+
{
|
| 279689 |
+
"epoch": 1142.9,
|
| 279690 |
+
"learning_rate": 7.727302100161551e-06,
|
| 279691 |
+
"loss": 0.2996,
|
| 279692 |
+
"step": 141735
|
| 279693 |
+
},
|
| 279694 |
+
{
|
| 279695 |
+
"epoch": 1142.94,
|
| 279696 |
+
"learning_rate": 7.727221324717287e-06,
|
| 279697 |
+
"loss": 0.3428,
|
| 279698 |
+
"step": 141740
|
| 279699 |
+
},
|
| 279700 |
+
{
|
| 279701 |
+
"epoch": 1142.98,
|
| 279702 |
+
"learning_rate": 7.727140549273021e-06,
|
| 279703 |
+
"loss": 0.6433,
|
| 279704 |
+
"step": 141745
|
| 279705 |
+
},
|
| 279706 |
+
{
|
| 279707 |
+
"epoch": 1143.0,
|
| 279708 |
+
"eval_loss": 0.3831433951854706,
|
| 279709 |
+
"eval_runtime": 41.1179,
|
| 279710 |
+
"eval_samples_per_second": 20.356,
|
| 279711 |
+
"eval_steps_per_second": 0.657,
|
| 279712 |
+
"eval_wer": 0.19271791099668997,
|
| 279713 |
+
"step": 141747
|
| 279714 |
+
},
|
| 279715 |
+
{
|
| 279716 |
+
"epoch": 1143.02,
|
| 279717 |
+
"learning_rate": 7.727059773828757e-06,
|
| 279718 |
+
"loss": 0.3706,
|
| 279719 |
+
"step": 141750
|
| 279720 |
+
},
|
| 279721 |
+
{
|
| 279722 |
+
"epoch": 1143.06,
|
| 279723 |
+
"learning_rate": 7.726978998384493e-06,
|
| 279724 |
+
"loss": 0.2449,
|
| 279725 |
+
"step": 141755
|
| 279726 |
+
},
|
| 279727 |
+
{
|
| 279728 |
+
"epoch": 1143.1,
|
| 279729 |
+
"learning_rate": 7.726898222940227e-06,
|
| 279730 |
+
"loss": 0.2845,
|
| 279731 |
+
"step": 141760
|
| 279732 |
+
},
|
| 279733 |
+
{
|
| 279734 |
+
"epoch": 1143.14,
|
| 279735 |
+
"learning_rate": 7.726817447495963e-06,
|
| 279736 |
+
"loss": 0.3353,
|
| 279737 |
+
"step": 141765
|
| 279738 |
+
},
|
| 279739 |
+
{
|
| 279740 |
+
"epoch": 1143.18,
|
| 279741 |
+
"learning_rate": 7.726736672051697e-06,
|
| 279742 |
+
"loss": 0.7831,
|
| 279743 |
+
"step": 141770
|
| 279744 |
+
},
|
| 279745 |
+
{
|
| 279746 |
+
"epoch": 1143.22,
|
| 279747 |
+
"learning_rate": 7.726655896607433e-06,
|
| 279748 |
+
"loss": 0.9689,
|
| 279749 |
+
"step": 141775
|
| 279750 |
+
},
|
| 279751 |
+
{
|
| 279752 |
+
"epoch": 1143.27,
|
| 279753 |
+
"learning_rate": 7.726575121163167e-06,
|
| 279754 |
+
"loss": 0.3101,
|
| 279755 |
+
"step": 141780
|
| 279756 |
+
},
|
| 279757 |
+
{
|
| 279758 |
+
"epoch": 1143.31,
|
| 279759 |
+
"learning_rate": 7.726494345718903e-06,
|
| 279760 |
+
"loss": 0.2913,
|
| 279761 |
+
"step": 141785
|
| 279762 |
+
},
|
| 279763 |
+
{
|
| 279764 |
+
"epoch": 1143.35,
|
| 279765 |
+
"learning_rate": 7.726413570274637e-06,
|
| 279766 |
+
"loss": 0.3821,
|
| 279767 |
+
"step": 141790
|
| 279768 |
+
},
|
| 279769 |
+
{
|
| 279770 |
+
"epoch": 1143.39,
|
| 279771 |
+
"learning_rate": 7.726332794830373e-06,
|
| 279772 |
+
"loss": 0.7098,
|
| 279773 |
+
"step": 141795
|
| 279774 |
+
},
|
| 279775 |
+
{
|
| 279776 |
+
"epoch": 1143.43,
|
| 279777 |
+
"learning_rate": 7.726252019386107e-06,
|
| 279778 |
+
"loss": 0.8485,
|
| 279779 |
+
"step": 141800
|
| 279780 |
+
},
|
| 279781 |
+
{
|
| 279782 |
+
"epoch": 1143.47,
|
| 279783 |
+
"learning_rate": 7.726171243941843e-06,
|
| 279784 |
+
"loss": 0.2593,
|
| 279785 |
+
"step": 141805
|
| 279786 |
+
},
|
| 279787 |
+
{
|
| 279788 |
+
"epoch": 1143.51,
|
| 279789 |
+
"learning_rate": 7.726090468497577e-06,
|
| 279790 |
+
"loss": 0.2944,
|
| 279791 |
+
"step": 141810
|
| 279792 |
+
},
|
| 279793 |
+
{
|
| 279794 |
+
"epoch": 1143.55,
|
| 279795 |
+
"learning_rate": 7.726009693053313e-06,
|
| 279796 |
+
"loss": 0.2959,
|
| 279797 |
+
"step": 141815
|
| 279798 |
+
},
|
| 279799 |
+
{
|
| 279800 |
+
"epoch": 1143.59,
|
| 279801 |
+
"learning_rate": 7.725928917609049e-06,
|
| 279802 |
+
"loss": 0.7449,
|
| 279803 |
+
"step": 141820
|
| 279804 |
+
},
|
| 279805 |
+
{
|
| 279806 |
+
"epoch": 1143.63,
|
| 279807 |
+
"learning_rate": 7.725848142164783e-06,
|
| 279808 |
+
"loss": 0.7156,
|
| 279809 |
+
"step": 141825
|
| 279810 |
+
},
|
| 279811 |
+
{
|
| 279812 |
+
"epoch": 1143.67,
|
| 279813 |
+
"learning_rate": 7.725767366720519e-06,
|
| 279814 |
+
"loss": 0.3058,
|
| 279815 |
+
"step": 141830
|
| 279816 |
+
},
|
| 279817 |
+
{
|
| 279818 |
+
"epoch": 1143.71,
|
| 279819 |
+
"learning_rate": 7.725686591276253e-06,
|
| 279820 |
+
"loss": 0.2875,
|
| 279821 |
+
"step": 141835
|
| 279822 |
+
},
|
| 279823 |
+
{
|
| 279824 |
+
"epoch": 1143.75,
|
| 279825 |
+
"learning_rate": 7.725605815831989e-06,
|
| 279826 |
+
"loss": 0.2819,
|
| 279827 |
+
"step": 141840
|
| 279828 |
+
},
|
| 279829 |
+
{
|
| 279830 |
+
"epoch": 1143.79,
|
| 279831 |
+
"learning_rate": 7.725525040387723e-06,
|
| 279832 |
+
"loss": 0.6082,
|
| 279833 |
+
"step": 141845
|
| 279834 |
+
},
|
| 279835 |
+
{
|
| 279836 |
+
"epoch": 1143.83,
|
| 279837 |
+
"learning_rate": 7.725444264943459e-06,
|
| 279838 |
+
"loss": 0.7674,
|
| 279839 |
+
"step": 141850
|
| 279840 |
+
},
|
| 279841 |
+
{
|
| 279842 |
+
"epoch": 1143.87,
|
| 279843 |
+
"learning_rate": 7.725363489499193e-06,
|
| 279844 |
+
"loss": 0.2629,
|
| 279845 |
+
"step": 141855
|
| 279846 |
+
},
|
| 279847 |
+
{
|
| 279848 |
+
"epoch": 1143.91,
|
| 279849 |
+
"learning_rate": 7.725282714054929e-06,
|
| 279850 |
+
"loss": 0.2882,
|
| 279851 |
+
"step": 141860
|
| 279852 |
+
},
|
| 279853 |
+
{
|
| 279854 |
+
"epoch": 1143.95,
|
| 279855 |
+
"learning_rate": 7.725201938610663e-06,
|
| 279856 |
+
"loss": 0.3376,
|
| 279857 |
+
"step": 141865
|
| 279858 |
+
},
|
| 279859 |
+
{
|
| 279860 |
+
"epoch": 1143.99,
|
| 279861 |
+
"learning_rate": 7.725121163166399e-06,
|
| 279862 |
+
"loss": 0.7656,
|
| 279863 |
+
"step": 141870
|
| 279864 |
+
},
|
| 279865 |
+
{
|
| 279866 |
+
"epoch": 1144.0,
|
| 279867 |
+
"eval_loss": 0.32772520184516907,
|
| 279868 |
+
"eval_runtime": 41.4072,
|
| 279869 |
+
"eval_samples_per_second": 20.214,
|
| 279870 |
+
"eval_steps_per_second": 0.652,
|
| 279871 |
+
"eval_wer": 0.18056568012829857,
|
| 279872 |
+
"step": 141871
|
| 279873 |
+
},
|
| 279874 |
+
{
|
| 279875 |
+
"epoch": 1144.03,
|
| 279876 |
+
"learning_rate": 7.725040387722134e-06,
|
| 279877 |
+
"loss": 0.2765,
|
| 279878 |
+
"step": 141875
|
| 279879 |
+
},
|
| 279880 |
+
{
|
| 279881 |
+
"epoch": 1144.07,
|
| 279882 |
+
"learning_rate": 7.724959612277868e-06,
|
| 279883 |
+
"loss": 0.2441,
|
| 279884 |
+
"step": 141880
|
| 279885 |
+
},
|
| 279886 |
+
{
|
| 279887 |
+
"epoch": 1144.11,
|
| 279888 |
+
"learning_rate": 7.724878836833604e-06,
|
| 279889 |
+
"loss": 0.311,
|
| 279890 |
+
"step": 141885
|
| 279891 |
+
},
|
| 279892 |
+
{
|
| 279893 |
+
"epoch": 1144.15,
|
| 279894 |
+
"learning_rate": 7.724798061389338e-06,
|
| 279895 |
+
"loss": 0.3957,
|
| 279896 |
+
"step": 141890
|
| 279897 |
+
},
|
| 279898 |
+
{
|
| 279899 |
+
"epoch": 1144.19,
|
| 279900 |
+
"learning_rate": 7.724717285945074e-06,
|
| 279901 |
+
"loss": 0.8882,
|
| 279902 |
+
"step": 141895
|
| 279903 |
+
},
|
| 279904 |
+
{
|
| 279905 |
+
"epoch": 1144.23,
|
| 279906 |
+
"learning_rate": 7.724636510500808e-06,
|
| 279907 |
+
"loss": 0.5858,
|
| 279908 |
+
"step": 141900
|
| 279909 |
+
},
|
| 279910 |
+
{
|
| 279911 |
+
"epoch": 1144.27,
|
| 279912 |
+
"learning_rate": 7.724555735056544e-06,
|
| 279913 |
+
"loss": 0.2691,
|
| 279914 |
+
"step": 141905
|
| 279915 |
+
},
|
| 279916 |
+
{
|
| 279917 |
+
"epoch": 1144.31,
|
| 279918 |
+
"learning_rate": 7.724474959612278e-06,
|
| 279919 |
+
"loss": 0.2628,
|
| 279920 |
+
"step": 141910
|
| 279921 |
+
},
|
| 279922 |
+
{
|
| 279923 |
+
"epoch": 1144.35,
|
| 279924 |
+
"learning_rate": 7.724394184168014e-06,
|
| 279925 |
+
"loss": 0.3303,
|
| 279926 |
+
"step": 141915
|
| 279927 |
+
},
|
| 279928 |
+
{
|
| 279929 |
+
"epoch": 1144.39,
|
| 279930 |
+
"learning_rate": 7.724313408723748e-06,
|
| 279931 |
+
"loss": 0.979,
|
| 279932 |
+
"step": 141920
|
| 279933 |
+
},
|
| 279934 |
+
{
|
| 279935 |
+
"epoch": 1144.43,
|
| 279936 |
+
"learning_rate": 7.724232633279484e-06,
|
| 279937 |
+
"loss": 1.0384,
|
| 279938 |
+
"step": 141925
|
| 279939 |
+
},
|
| 279940 |
+
{
|
| 279941 |
+
"epoch": 1144.47,
|
| 279942 |
+
"learning_rate": 7.72415185783522e-06,
|
| 279943 |
+
"loss": 0.2295,
|
| 279944 |
+
"step": 141930
|
| 279945 |
+
},
|
| 279946 |
+
{
|
| 279947 |
+
"epoch": 1144.51,
|
| 279948 |
+
"learning_rate": 7.724071082390954e-06,
|
| 279949 |
+
"loss": 0.3122,
|
| 279950 |
+
"step": 141935
|
| 279951 |
+
},
|
| 279952 |
+
{
|
| 279953 |
+
"epoch": 1144.55,
|
| 279954 |
+
"learning_rate": 7.72399030694669e-06,
|
| 279955 |
+
"loss": 0.3926,
|
| 279956 |
+
"step": 141940
|
| 279957 |
+
},
|
| 279958 |
+
{
|
| 279959 |
+
"epoch": 1144.59,
|
| 279960 |
+
"learning_rate": 7.723909531502424e-06,
|
| 279961 |
+
"loss": 0.7148,
|
| 279962 |
+
"step": 141945
|
| 279963 |
+
},
|
| 279964 |
+
{
|
| 279965 |
+
"epoch": 1144.63,
|
| 279966 |
+
"learning_rate": 7.72382875605816e-06,
|
| 279967 |
+
"loss": 0.591,
|
| 279968 |
+
"step": 141950
|
| 279969 |
+
},
|
| 279970 |
+
{
|
| 279971 |
+
"epoch": 1144.67,
|
| 279972 |
+
"learning_rate": 7.723747980613894e-06,
|
| 279973 |
+
"loss": 0.2606,
|
| 279974 |
+
"step": 141955
|
| 279975 |
+
},
|
| 279976 |
+
{
|
| 279977 |
+
"epoch": 1144.71,
|
| 279978 |
+
"learning_rate": 7.72366720516963e-06,
|
| 279979 |
+
"loss": 0.272,
|
| 279980 |
+
"step": 141960
|
| 279981 |
+
},
|
| 279982 |
+
{
|
| 279983 |
+
"epoch": 1144.76,
|
| 279984 |
+
"learning_rate": 7.723586429725364e-06,
|
| 279985 |
+
"loss": 0.402,
|
| 279986 |
+
"step": 141965
|
| 279987 |
+
},
|
| 279988 |
+
{
|
| 279989 |
+
"epoch": 1144.8,
|
| 279990 |
+
"learning_rate": 7.7235056542811e-06,
|
| 279991 |
+
"loss": 0.7544,
|
| 279992 |
+
"step": 141970
|
| 279993 |
+
},
|
| 279994 |
+
{
|
| 279995 |
+
"epoch": 1144.84,
|
| 279996 |
+
"learning_rate": 7.723424878836834e-06,
|
| 279997 |
+
"loss": 0.6907,
|
| 279998 |
+
"step": 141975
|
| 279999 |
+
},
|
| 280000 |
+
{
|
| 280001 |
+
"epoch": 1144.88,
|
| 280002 |
+
"learning_rate": 7.72334410339257e-06,
|
| 280003 |
+
"loss": 0.3154,
|
| 280004 |
+
"step": 141980
|
| 280005 |
+
},
|
| 280006 |
+
{
|
| 280007 |
+
"epoch": 1144.92,
|
| 280008 |
+
"learning_rate": 7.723263327948304e-06,
|
| 280009 |
+
"loss": 0.314,
|
| 280010 |
+
"step": 141985
|
| 280011 |
+
},
|
| 280012 |
+
{
|
| 280013 |
+
"epoch": 1144.96,
|
| 280014 |
+
"learning_rate": 7.72318255250404e-06,
|
| 280015 |
+
"loss": 0.4343,
|
| 280016 |
+
"step": 141990
|
| 280017 |
+
},
|
| 280018 |
+
{
|
| 280019 |
+
"epoch": 1145.0,
|
| 280020 |
+
"learning_rate": 7.723101777059776e-06,
|
| 280021 |
+
"loss": 1.0367,
|
| 280022 |
+
"step": 141995
|
| 280023 |
+
},
|
| 280024 |
+
{
|
| 280025 |
+
"epoch": 1145.0,
|
| 280026 |
+
"eval_loss": 0.36632317304611206,
|
| 280027 |
+
"eval_runtime": 41.4842,
|
| 280028 |
+
"eval_samples_per_second": 20.2,
|
| 280029 |
+
"eval_steps_per_second": 0.651,
|
| 280030 |
+
"eval_wer": 0.17683603283780785,
|
| 280031 |
+
"step": 141995
|
| 280032 |
+
},
|
| 280033 |
+
{
|
| 280034 |
+
"epoch": 1145.04,
|
| 280035 |
+
"learning_rate": 7.72302100161551e-06,
|
| 280036 |
+
"loss": 0.3223,
|
| 280037 |
+
"step": 142000
|
| 280038 |
+
},
|
| 280039 |
+
{
|
| 280040 |
+
"epoch": 1145.08,
|
| 280041 |
+
"learning_rate": 7.722940226171246e-06,
|
| 280042 |
+
"loss": 0.3433,
|
| 280043 |
+
"step": 142005
|
| 280044 |
+
},
|
| 280045 |
+
{
|
| 280046 |
+
"epoch": 1145.12,
|
| 280047 |
+
"learning_rate": 7.72285945072698e-06,
|
| 280048 |
+
"loss": 0.3129,
|
| 280049 |
+
"step": 142010
|
| 280050 |
+
},
|
| 280051 |
+
{
|
| 280052 |
+
"epoch": 1145.16,
|
| 280053 |
+
"learning_rate": 7.722778675282716e-06,
|
| 280054 |
+
"loss": 0.3968,
|
| 280055 |
+
"step": 142015
|
| 280056 |
+
},
|
| 280057 |
+
{
|
| 280058 |
+
"epoch": 1145.2,
|
| 280059 |
+
"learning_rate": 7.72269789983845e-06,
|
| 280060 |
+
"loss": 1.1603,
|
| 280061 |
+
"step": 142020
|
| 280062 |
+
},
|
| 280063 |
+
{
|
| 280064 |
+
"epoch": 1145.24,
|
| 280065 |
+
"learning_rate": 7.722617124394186e-06,
|
| 280066 |
+
"loss": 0.2974,
|
| 280067 |
+
"step": 142025
|
| 280068 |
+
},
|
| 280069 |
+
{
|
| 280070 |
+
"epoch": 1145.28,
|
| 280071 |
+
"learning_rate": 7.72253634894992e-06,
|
| 280072 |
+
"loss": 0.3043,
|
| 280073 |
+
"step": 142030
|
| 280074 |
+
},
|
| 280075 |
+
{
|
| 280076 |
+
"epoch": 1145.32,
|
| 280077 |
+
"learning_rate": 7.722455573505656e-06,
|
| 280078 |
+
"loss": 0.2808,
|
| 280079 |
+
"step": 142035
|
| 280080 |
+
},
|
| 280081 |
+
{
|
| 280082 |
+
"epoch": 1145.36,
|
| 280083 |
+
"learning_rate": 7.72237479806139e-06,
|
| 280084 |
+
"loss": 0.418,
|
| 280085 |
+
"step": 142040
|
| 280086 |
+
},
|
| 280087 |
+
{
|
| 280088 |
+
"epoch": 1145.4,
|
| 280089 |
+
"learning_rate": 7.722294022617126e-06,
|
| 280090 |
+
"loss": 1.1368,
|
| 280091 |
+
"step": 142045
|
| 280092 |
+
},
|
| 280093 |
+
{
|
| 280094 |
+
"epoch": 1145.44,
|
| 280095 |
+
"learning_rate": 7.72221324717286e-06,
|
| 280096 |
+
"loss": 0.305,
|
| 280097 |
+
"step": 142050
|
| 280098 |
+
},
|
| 280099 |
+
{
|
| 280100 |
+
"epoch": 1145.48,
|
| 280101 |
+
"learning_rate": 7.722132471728596e-06,
|
| 280102 |
+
"loss": 0.2486,
|
| 280103 |
+
"step": 142055
|
| 280104 |
+
},
|
| 280105 |
+
{
|
| 280106 |
+
"epoch": 1145.52,
|
| 280107 |
+
"learning_rate": 7.722051696284331e-06,
|
| 280108 |
+
"loss": 0.276,
|
| 280109 |
+
"step": 142060
|
| 280110 |
+
},
|
| 280111 |
+
{
|
| 280112 |
+
"epoch": 1145.56,
|
| 280113 |
+
"learning_rate": 7.721970920840066e-06,
|
| 280114 |
+
"loss": 0.3786,
|
| 280115 |
+
"step": 142065
|
| 280116 |
+
},
|
| 280117 |
+
{
|
| 280118 |
+
"epoch": 1145.6,
|
| 280119 |
+
"learning_rate": 7.721890145395801e-06,
|
| 280120 |
+
"loss": 1.0858,
|
| 280121 |
+
"step": 142070
|
| 280122 |
+
},
|
| 280123 |
+
{
|
| 280124 |
+
"epoch": 1145.65,
|
| 280125 |
+
"learning_rate": 7.721809369951536e-06,
|
| 280126 |
+
"loss": 0.3283,
|
| 280127 |
+
"step": 142075
|
| 280128 |
+
},
|
| 280129 |
+
{
|
| 280130 |
+
"epoch": 1145.69,
|
| 280131 |
+
"learning_rate": 7.721728594507271e-06,
|
| 280132 |
+
"loss": 0.269,
|
| 280133 |
+
"step": 142080
|
| 280134 |
+
},
|
| 280135 |
+
{
|
| 280136 |
+
"epoch": 1145.73,
|
| 280137 |
+
"learning_rate": 7.721647819063006e-06,
|
| 280138 |
+
"loss": 0.3025,
|
| 280139 |
+
"step": 142085
|
| 280140 |
+
},
|
| 280141 |
+
{
|
| 280142 |
+
"epoch": 1145.77,
|
| 280143 |
+
"learning_rate": 7.721567043618741e-06,
|
| 280144 |
+
"loss": 0.459,
|
| 280145 |
+
"step": 142090
|
| 280146 |
+
},
|
| 280147 |
+
{
|
| 280148 |
+
"epoch": 1145.81,
|
| 280149 |
+
"learning_rate": 7.721486268174475e-06,
|
| 280150 |
+
"loss": 1.3132,
|
| 280151 |
+
"step": 142095
|
| 280152 |
+
},
|
| 280153 |
+
{
|
| 280154 |
+
"epoch": 1145.85,
|
| 280155 |
+
"learning_rate": 7.721405492730211e-06,
|
| 280156 |
+
"loss": 0.5061,
|
| 280157 |
+
"step": 142100
|
| 280158 |
+
},
|
| 280159 |
+
{
|
| 280160 |
+
"epoch": 1145.89,
|
| 280161 |
+
"learning_rate": 7.721324717285945e-06,
|
| 280162 |
+
"loss": 0.3975,
|
| 280163 |
+
"step": 142105
|
| 280164 |
+
},
|
| 280165 |
+
{
|
| 280166 |
+
"epoch": 1145.93,
|
| 280167 |
+
"learning_rate": 7.721243941841681e-06,
|
| 280168 |
+
"loss": 0.2731,
|
| 280169 |
+
"step": 142110
|
| 280170 |
+
},
|
| 280171 |
+
{
|
| 280172 |
+
"epoch": 1145.97,
|
| 280173 |
+
"learning_rate": 7.721163166397415e-06,
|
| 280174 |
+
"loss": 0.4579,
|
| 280175 |
+
"step": 142115
|
| 280176 |
+
},
|
| 280177 |
+
{
|
| 280178 |
+
"epoch": 1146.0,
|
| 280179 |
+
"eval_loss": 0.3085727393627167,
|
| 280180 |
+
"eval_runtime": 41.3912,
|
| 280181 |
+
"eval_samples_per_second": 20.246,
|
| 280182 |
+
"eval_steps_per_second": 0.652,
|
| 280183 |
+
"eval_wer": 0.18064516129032257,
|
| 280184 |
+
"step": 142119
|
| 280185 |
}
|
| 280186 |
],
|
| 280187 |
"max_steps": 620000,
|
| 280188 |
"num_train_epochs": 5000,
|
| 280189 |
+
"total_flos": 3.999266636171066e+20,
|
| 280190 |
"trial_name": null,
|
| 280191 |
"trial_params": null
|
| 280192 |
}
|
model-bin/finetune/base/{checkpoint-141498 β checkpoint-142119}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1630212588.3931246/events.out.tfevents.1630212588.cc93b136ebf5.1086.39
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0c5e861e922cdc7f53de2c00486af0f74996318b37fc1d9a494d8bf793692a91
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630213036.9606094/events.out.tfevents.1630213036.cc93b136ebf5.1086.41
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:021df7086f0e001322c25f1ee0bfc9a11e6cc059a7a4bb73d75890a874f6cdff
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630213481.1394386/events.out.tfevents.1630213481.cc93b136ebf5.1086.43
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d78ef73847fcf4c078b856415372a2e0d6576f9df97aa1685ddaf3c9f50c4817
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630213924.587934/events.out.tfevents.1630213924.cc93b136ebf5.1086.45
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:553de5a9a6a00d490406270cb5e07813c91d013e4de78cf533963a3c61620cd3
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630214361.273169/events.out.tfevents.1630214361.cc93b136ebf5.1086.47
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1e16b44e2fa2d95584801fdcfbc14d177de5fd3b3d3a763c5ef239f82dc43921
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630212588.cc93b136ebf5.1086.38
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:503b8170234cacdf6c83700bcb0029868147641f33e3037d43424ab2d405a5ec
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630213036.cc93b136ebf5.1086.40
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:55d2b14d9c8d7738fa05f69362035ff2b99c1dd41be2078b1cd300bd8297638f
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630213481.cc93b136ebf5.1086.42
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ff7c368ef5da476643bc4e149343cae1f0d21f86a868ba6b7cf452073289150f
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630213924.cc93b136ebf5.1086.44
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:128d43720ab01b9d11b9073331af63a176f4d66056fbe6d2548778e1103e271b
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630214361.cc93b136ebf5.1086.46
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:812e218f1a368ab0f3db92f2f3fcae7cef4c84f56b8baab5cb1d5f6eb0ed0510
|
| 3 |
+
size 8462
|