"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/trainer_state.json +793 -4
- model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630244012.7530751/events.out.tfevents.1630244012.cc93b136ebf5.1086.179 +3 -0
- model-bin/finetune/base/log/1630244448.5792234/events.out.tfevents.1630244448.cc93b136ebf5.1086.181 +3 -0
- model-bin/finetune/base/log/1630244993.356596/events.out.tfevents.1630244993.cc93b136ebf5.1086.183 +3 -0
- model-bin/finetune/base/log/1630245427.5586958/events.out.tfevents.1630245427.cc93b136ebf5.1086.185 +3 -0
- model-bin/finetune/base/log/1630245858.8737807/events.out.tfevents.1630245858.cc93b136ebf5.1086.187 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630244012.cc93b136ebf5.1086.178 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630244448.cc93b136ebf5.1086.180 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630244993.cc93b136ebf5.1086.182 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630245427.cc93b136ebf5.1086.184 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630245858.cc93b136ebf5.1086.186 +3 -0
model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c9665c6e1c6f2987ac982b4f02510d642ccf49cb18e475af98729715643f2744
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:83712bb0bb146694098ef77fead19cec258379ae2b0f648e4edb18652a55565a
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3f44a922c3d45f1ec6984b47ad3b841b6c8eca8288bc65f7fa757f180153fd40
|
| 3 |
size 14503
|
model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:adf29ee28b173423b576c771747dd93d436f5c3f2a3af9a31f91c508ff5dd1b2
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ee134784171ae4f4005cf971653fad3f3e37c1a6c7b565c91f04852ab760d0e4
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1668011442822563,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -290475,11 +290475,800 @@
|
|
| 290475 |
"eval_steps_per_second": 0.644,
|
| 290476 |
"eval_wer": 0.17267648552564754,
|
| 290477 |
"step": 150205
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 290478 |
}
|
| 290479 |
],
|
| 290480 |
-
"max_steps":
|
| 290481 |
"num_train_epochs": 5000,
|
| 290482 |
-
"total_flos": 4.
|
| 290483 |
"trial_name": null,
|
| 290484 |
"trial_params": null
|
| 290485 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1668011442822563,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
| 4 |
+
"epoch": 1206.0,
|
| 5 |
+
"global_step": 150827,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 290475 |
"eval_steps_per_second": 0.644,
|
| 290476 |
"eval_wer": 0.17267648552564754,
|
| 290477 |
"step": 150205
|
| 290478 |
+
},
|
| 290479 |
+
{
|
| 290480 |
+
"epoch": 1211.04,
|
| 290481 |
+
"learning_rate": 7.590452342487884e-06,
|
| 290482 |
+
"loss": 0.3201,
|
| 290483 |
+
"step": 150210
|
| 290484 |
+
},
|
| 290485 |
+
{
|
| 290486 |
+
"epoch": 1211.08,
|
| 290487 |
+
"learning_rate": 7.59037156704362e-06,
|
| 290488 |
+
"loss": 0.2736,
|
| 290489 |
+
"step": 150215
|
| 290490 |
+
},
|
| 290491 |
+
{
|
| 290492 |
+
"epoch": 1211.12,
|
| 290493 |
+
"learning_rate": 7.590290791599354e-06,
|
| 290494 |
+
"loss": 0.2423,
|
| 290495 |
+
"step": 150220
|
| 290496 |
+
},
|
| 290497 |
+
{
|
| 290498 |
+
"epoch": 1211.16,
|
| 290499 |
+
"learning_rate": 7.59021001615509e-06,
|
| 290500 |
+
"loss": 0.4379,
|
| 290501 |
+
"step": 150225
|
| 290502 |
+
},
|
| 290503 |
+
{
|
| 290504 |
+
"epoch": 1211.2,
|
| 290505 |
+
"learning_rate": 7.590129240710824e-06,
|
| 290506 |
+
"loss": 0.9755,
|
| 290507 |
+
"step": 150230
|
| 290508 |
+
},
|
| 290509 |
+
{
|
| 290510 |
+
"epoch": 1211.24,
|
| 290511 |
+
"learning_rate": 7.59004846526656e-06,
|
| 290512 |
+
"loss": 0.3417,
|
| 290513 |
+
"step": 150235
|
| 290514 |
+
},
|
| 290515 |
+
{
|
| 290516 |
+
"epoch": 1211.28,
|
| 290517 |
+
"learning_rate": 7.5899676898222955e-06,
|
| 290518 |
+
"loss": 0.2895,
|
| 290519 |
+
"step": 150240
|
| 290520 |
+
},
|
| 290521 |
+
{
|
| 290522 |
+
"epoch": 1211.32,
|
| 290523 |
+
"learning_rate": 7.58988691437803e-06,
|
| 290524 |
+
"loss": 0.2984,
|
| 290525 |
+
"step": 150245
|
| 290526 |
+
},
|
| 290527 |
+
{
|
| 290528 |
+
"epoch": 1211.36,
|
| 290529 |
+
"learning_rate": 7.5898061389337655e-06,
|
| 290530 |
+
"loss": 0.3888,
|
| 290531 |
+
"step": 150250
|
| 290532 |
+
},
|
| 290533 |
+
{
|
| 290534 |
+
"epoch": 1211.4,
|
| 290535 |
+
"learning_rate": 7.5897253634895e-06,
|
| 290536 |
+
"loss": 1.1877,
|
| 290537 |
+
"step": 150255
|
| 290538 |
+
},
|
| 290539 |
+
{
|
| 290540 |
+
"epoch": 1211.44,
|
| 290541 |
+
"learning_rate": 7.5896445880452355e-06,
|
| 290542 |
+
"loss": 0.2986,
|
| 290543 |
+
"step": 150260
|
| 290544 |
+
},
|
| 290545 |
+
{
|
| 290546 |
+
"epoch": 1211.48,
|
| 290547 |
+
"learning_rate": 7.58956381260097e-06,
|
| 290548 |
+
"loss": 0.2978,
|
| 290549 |
+
"step": 150265
|
| 290550 |
+
},
|
| 290551 |
+
{
|
| 290552 |
+
"epoch": 1211.52,
|
| 290553 |
+
"learning_rate": 7.5894830371567055e-06,
|
| 290554 |
+
"loss": 0.3136,
|
| 290555 |
+
"step": 150270
|
| 290556 |
+
},
|
| 290557 |
+
{
|
| 290558 |
+
"epoch": 1211.56,
|
| 290559 |
+
"learning_rate": 7.58940226171244e-06,
|
| 290560 |
+
"loss": 0.4057,
|
| 290561 |
+
"step": 150275
|
| 290562 |
+
},
|
| 290563 |
+
{
|
| 290564 |
+
"epoch": 1211.6,
|
| 290565 |
+
"learning_rate": 7.5893214862681754e-06,
|
| 290566 |
+
"loss": 1.0684,
|
| 290567 |
+
"step": 150280
|
| 290568 |
+
},
|
| 290569 |
+
{
|
| 290570 |
+
"epoch": 1211.64,
|
| 290571 |
+
"learning_rate": 7.5892407108239096e-06,
|
| 290572 |
+
"loss": 0.3111,
|
| 290573 |
+
"step": 150285
|
| 290574 |
+
},
|
| 290575 |
+
{
|
| 290576 |
+
"epoch": 1211.68,
|
| 290577 |
+
"learning_rate": 7.589159935379645e-06,
|
| 290578 |
+
"loss": 0.3245,
|
| 290579 |
+
"step": 150290
|
| 290580 |
+
},
|
| 290581 |
+
{
|
| 290582 |
+
"epoch": 1211.72,
|
| 290583 |
+
"learning_rate": 7.58907915993538e-06,
|
| 290584 |
+
"loss": 0.3207,
|
| 290585 |
+
"step": 150295
|
| 290586 |
+
},
|
| 290587 |
+
{
|
| 290588 |
+
"epoch": 1211.76,
|
| 290589 |
+
"learning_rate": 7.588998384491115e-06,
|
| 290590 |
+
"loss": 0.3932,
|
| 290591 |
+
"step": 150300
|
| 290592 |
+
},
|
| 290593 |
+
{
|
| 290594 |
+
"epoch": 1211.8,
|
| 290595 |
+
"learning_rate": 7.588917609046851e-06,
|
| 290596 |
+
"loss": 1.1473,
|
| 290597 |
+
"step": 150305
|
| 290598 |
+
},
|
| 290599 |
+
{
|
| 290600 |
+
"epoch": 1211.84,
|
| 290601 |
+
"learning_rate": 7.588836833602585e-06,
|
| 290602 |
+
"loss": 0.2849,
|
| 290603 |
+
"step": 150310
|
| 290604 |
+
},
|
| 290605 |
+
{
|
| 290606 |
+
"epoch": 1211.88,
|
| 290607 |
+
"learning_rate": 7.588756058158321e-06,
|
| 290608 |
+
"loss": 0.3008,
|
| 290609 |
+
"step": 150315
|
| 290610 |
+
},
|
| 290611 |
+
{
|
| 290612 |
+
"epoch": 1211.92,
|
| 290613 |
+
"learning_rate": 7.588675282714055e-06,
|
| 290614 |
+
"loss": 0.3498,
|
| 290615 |
+
"step": 150320
|
| 290616 |
+
},
|
| 290617 |
+
{
|
| 290618 |
+
"epoch": 1211.96,
|
| 290619 |
+
"learning_rate": 7.588594507269791e-06,
|
| 290620 |
+
"loss": 0.528,
|
| 290621 |
+
"step": 150325
|
| 290622 |
+
},
|
| 290623 |
+
{
|
| 290624 |
+
"epoch": 1212.0,
|
| 290625 |
+
"eval_loss": 0.38405805826187134,
|
| 290626 |
+
"eval_runtime": 42.6893,
|
| 290627 |
+
"eval_samples_per_second": 19.7,
|
| 290628 |
+
"eval_steps_per_second": 0.632,
|
| 290629 |
+
"eval_wer": 0.17788736420811893,
|
| 290630 |
+
"step": 150329
|
| 290631 |
+
},
|
| 290632 |
+
{
|
| 290633 |
+
"epoch": 1212.01,
|
| 290634 |
+
"learning_rate": 7.588513731825525e-06,
|
| 290635 |
+
"loss": 0.4414,
|
| 290636 |
+
"step": 150330
|
| 290637 |
+
},
|
| 290638 |
+
{
|
| 290639 |
+
"epoch": 1212.05,
|
| 290640 |
+
"learning_rate": 7.588432956381261e-06,
|
| 290641 |
+
"loss": 0.3031,
|
| 290642 |
+
"step": 150335
|
| 290643 |
+
},
|
| 290644 |
+
{
|
| 290645 |
+
"epoch": 1212.09,
|
| 290646 |
+
"learning_rate": 7.588352180936995e-06,
|
| 290647 |
+
"loss": 0.2945,
|
| 290648 |
+
"step": 150340
|
| 290649 |
+
},
|
| 290650 |
+
{
|
| 290651 |
+
"epoch": 1212.13,
|
| 290652 |
+
"learning_rate": 7.588271405492731e-06,
|
| 290653 |
+
"loss": 0.3029,
|
| 290654 |
+
"step": 150345
|
| 290655 |
+
},
|
| 290656 |
+
{
|
| 290657 |
+
"epoch": 1212.17,
|
| 290658 |
+
"learning_rate": 7.588190630048465e-06,
|
| 290659 |
+
"loss": 0.4425,
|
| 290660 |
+
"step": 150350
|
| 290661 |
+
},
|
| 290662 |
+
{
|
| 290663 |
+
"epoch": 1212.21,
|
| 290664 |
+
"learning_rate": 7.588109854604201e-06,
|
| 290665 |
+
"loss": 1.0489,
|
| 290666 |
+
"step": 150355
|
| 290667 |
+
},
|
| 290668 |
+
{
|
| 290669 |
+
"epoch": 1212.25,
|
| 290670 |
+
"learning_rate": 7.588029079159936e-06,
|
| 290671 |
+
"loss": 0.2773,
|
| 290672 |
+
"step": 150360
|
| 290673 |
+
},
|
| 290674 |
+
{
|
| 290675 |
+
"epoch": 1212.29,
|
| 290676 |
+
"learning_rate": 7.587948303715671e-06,
|
| 290677 |
+
"loss": 0.267,
|
| 290678 |
+
"step": 150365
|
| 290679 |
+
},
|
| 290680 |
+
{
|
| 290681 |
+
"epoch": 1212.33,
|
| 290682 |
+
"learning_rate": 7.587867528271407e-06,
|
| 290683 |
+
"loss": 0.3697,
|
| 290684 |
+
"step": 150370
|
| 290685 |
+
},
|
| 290686 |
+
{
|
| 290687 |
+
"epoch": 1212.37,
|
| 290688 |
+
"learning_rate": 7.587786752827141e-06,
|
| 290689 |
+
"loss": 0.4697,
|
| 290690 |
+
"step": 150375
|
| 290691 |
+
},
|
| 290692 |
+
{
|
| 290693 |
+
"epoch": 1212.41,
|
| 290694 |
+
"learning_rate": 7.587705977382877e-06,
|
| 290695 |
+
"loss": 1.0624,
|
| 290696 |
+
"step": 150380
|
| 290697 |
+
},
|
| 290698 |
+
{
|
| 290699 |
+
"epoch": 1212.45,
|
| 290700 |
+
"learning_rate": 7.587625201938611e-06,
|
| 290701 |
+
"loss": 0.2572,
|
| 290702 |
+
"step": 150385
|
| 290703 |
+
},
|
| 290704 |
+
{
|
| 290705 |
+
"epoch": 1212.49,
|
| 290706 |
+
"learning_rate": 7.587544426494347e-06,
|
| 290707 |
+
"loss": 0.3007,
|
| 290708 |
+
"step": 150390
|
| 290709 |
+
},
|
| 290710 |
+
{
|
| 290711 |
+
"epoch": 1212.53,
|
| 290712 |
+
"learning_rate": 7.587463651050081e-06,
|
| 290713 |
+
"loss": 0.2912,
|
| 290714 |
+
"step": 150395
|
| 290715 |
+
},
|
| 290716 |
+
{
|
| 290717 |
+
"epoch": 1212.57,
|
| 290718 |
+
"learning_rate": 7.587382875605817e-06,
|
| 290719 |
+
"loss": 0.5251,
|
| 290720 |
+
"step": 150400
|
| 290721 |
+
},
|
| 290722 |
+
{
|
| 290723 |
+
"epoch": 1212.61,
|
| 290724 |
+
"learning_rate": 7.587302100161551e-06,
|
| 290725 |
+
"loss": 1.0472,
|
| 290726 |
+
"step": 150405
|
| 290727 |
+
},
|
| 290728 |
+
{
|
| 290729 |
+
"epoch": 1212.65,
|
| 290730 |
+
"learning_rate": 7.587221324717287e-06,
|
| 290731 |
+
"loss": 0.3116,
|
| 290732 |
+
"step": 150410
|
| 290733 |
+
},
|
| 290734 |
+
{
|
| 290735 |
+
"epoch": 1212.69,
|
| 290736 |
+
"learning_rate": 7.587140549273022e-06,
|
| 290737 |
+
"loss": 0.2427,
|
| 290738 |
+
"step": 150415
|
| 290739 |
+
},
|
| 290740 |
+
{
|
| 290741 |
+
"epoch": 1212.73,
|
| 290742 |
+
"learning_rate": 7.587059773828757e-06,
|
| 290743 |
+
"loss": 0.3504,
|
| 290744 |
+
"step": 150420
|
| 290745 |
+
},
|
| 290746 |
+
{
|
| 290747 |
+
"epoch": 1212.77,
|
| 290748 |
+
"learning_rate": 7.586978998384492e-06,
|
| 290749 |
+
"loss": 0.4429,
|
| 290750 |
+
"step": 150425
|
| 290751 |
+
},
|
| 290752 |
+
{
|
| 290753 |
+
"epoch": 1212.81,
|
| 290754 |
+
"learning_rate": 7.586898222940227e-06,
|
| 290755 |
+
"loss": 0.945,
|
| 290756 |
+
"step": 150430
|
| 290757 |
+
},
|
| 290758 |
+
{
|
| 290759 |
+
"epoch": 1212.85,
|
| 290760 |
+
"learning_rate": 7.586817447495963e-06,
|
| 290761 |
+
"loss": 0.3042,
|
| 290762 |
+
"step": 150435
|
| 290763 |
+
},
|
| 290764 |
+
{
|
| 290765 |
+
"epoch": 1212.9,
|
| 290766 |
+
"learning_rate": 7.586736672051697e-06,
|
| 290767 |
+
"loss": 0.2662,
|
| 290768 |
+
"step": 150440
|
| 290769 |
+
},
|
| 290770 |
+
{
|
| 290771 |
+
"epoch": 1212.94,
|
| 290772 |
+
"learning_rate": 7.5866558966074326e-06,
|
| 290773 |
+
"loss": 0.3849,
|
| 290774 |
+
"step": 150445
|
| 290775 |
+
},
|
| 290776 |
+
{
|
| 290777 |
+
"epoch": 1212.98,
|
| 290778 |
+
"learning_rate": 7.586575121163167e-06,
|
| 290779 |
+
"loss": 0.6568,
|
| 290780 |
+
"step": 150450
|
| 290781 |
+
},
|
| 290782 |
+
{
|
| 290783 |
+
"epoch": 1213.0,
|
| 290784 |
+
"eval_loss": 0.3243919909000397,
|
| 290785 |
+
"eval_runtime": 45.4766,
|
| 290786 |
+
"eval_samples_per_second": 18.493,
|
| 290787 |
+
"eval_steps_per_second": 0.594,
|
| 290788 |
+
"eval_wer": 0.17089605734767024,
|
| 290789 |
+
"step": 150453
|
| 290790 |
+
},
|
| 290791 |
+
{
|
| 290792 |
+
"epoch": 1203.02,
|
| 290793 |
+
"learning_rate": 7.5864943457189025e-06,
|
| 290794 |
+
"loss": 0.2572,
|
| 290795 |
+
"step": 150455
|
| 290796 |
+
},
|
| 290797 |
+
{
|
| 290798 |
+
"epoch": 1203.06,
|
| 290799 |
+
"learning_rate": 7.586413570274637e-06,
|
| 290800 |
+
"loss": 0.2823,
|
| 290801 |
+
"step": 150460
|
| 290802 |
+
},
|
| 290803 |
+
{
|
| 290804 |
+
"epoch": 1203.1,
|
| 290805 |
+
"learning_rate": 7.5863327948303725e-06,
|
| 290806 |
+
"loss": 0.2366,
|
| 290807 |
+
"step": 150465
|
| 290808 |
+
},
|
| 290809 |
+
{
|
| 290810 |
+
"epoch": 1203.14,
|
| 290811 |
+
"learning_rate": 7.5862520193861075e-06,
|
| 290812 |
+
"loss": 0.3375,
|
| 290813 |
+
"step": 150470
|
| 290814 |
+
},
|
| 290815 |
+
{
|
| 290816 |
+
"epoch": 1203.18,
|
| 290817 |
+
"learning_rate": 7.5861712439418425e-06,
|
| 290818 |
+
"loss": 0.6214,
|
| 290819 |
+
"step": 150475
|
| 290820 |
+
},
|
| 290821 |
+
{
|
| 290822 |
+
"epoch": 1203.22,
|
| 290823 |
+
"learning_rate": 7.5860904684975775e-06,
|
| 290824 |
+
"loss": 0.9851,
|
| 290825 |
+
"step": 150480
|
| 290826 |
+
},
|
| 290827 |
+
{
|
| 290828 |
+
"epoch": 1203.26,
|
| 290829 |
+
"learning_rate": 7.5860096930533125e-06,
|
| 290830 |
+
"loss": 0.2788,
|
| 290831 |
+
"step": 150485
|
| 290832 |
+
},
|
| 290833 |
+
{
|
| 290834 |
+
"epoch": 1203.3,
|
| 290835 |
+
"learning_rate": 7.5859289176090474e-06,
|
| 290836 |
+
"loss": 0.2879,
|
| 290837 |
+
"step": 150490
|
| 290838 |
+
},
|
| 290839 |
+
{
|
| 290840 |
+
"epoch": 1203.34,
|
| 290841 |
+
"learning_rate": 7.5858481421647824e-06,
|
| 290842 |
+
"loss": 0.3532,
|
| 290843 |
+
"step": 150495
|
| 290844 |
+
},
|
| 290845 |
+
{
|
| 290846 |
+
"epoch": 1203.38,
|
| 290847 |
+
"learning_rate": 7.585767366720517e-06,
|
| 290848 |
+
"loss": 0.5532,
|
| 290849 |
+
"step": 150500
|
| 290850 |
+
},
|
| 290851 |
+
{
|
| 290852 |
+
"epoch": 1203.42,
|
| 290853 |
+
"learning_rate": 7.585686591276252e-06,
|
| 290854 |
+
"loss": 1.016,
|
| 290855 |
+
"step": 150505
|
| 290856 |
+
},
|
| 290857 |
+
{
|
| 290858 |
+
"epoch": 1203.46,
|
| 290859 |
+
"learning_rate": 7.585605815831988e-06,
|
| 290860 |
+
"loss": 0.303,
|
| 290861 |
+
"step": 150510
|
| 290862 |
+
},
|
| 290863 |
+
{
|
| 290864 |
+
"epoch": 1203.5,
|
| 290865 |
+
"learning_rate": 7.585525040387722e-06,
|
| 290866 |
+
"loss": 0.2581,
|
| 290867 |
+
"step": 150515
|
| 290868 |
+
},
|
| 290869 |
+
{
|
| 290870 |
+
"epoch": 1203.54,
|
| 290871 |
+
"learning_rate": 7.585444264943458e-06,
|
| 290872 |
+
"loss": 0.3192,
|
| 290873 |
+
"step": 150520
|
| 290874 |
+
},
|
| 290875 |
+
{
|
| 290876 |
+
"epoch": 1203.58,
|
| 290877 |
+
"learning_rate": 7.585363489499192e-06,
|
| 290878 |
+
"loss": 0.5658,
|
| 290879 |
+
"step": 150525
|
| 290880 |
+
},
|
| 290881 |
+
{
|
| 290882 |
+
"epoch": 1203.62,
|
| 290883 |
+
"learning_rate": 7.585282714054928e-06,
|
| 290884 |
+
"loss": 1.1043,
|
| 290885 |
+
"step": 150530
|
| 290886 |
+
},
|
| 290887 |
+
{
|
| 290888 |
+
"epoch": 1203.66,
|
| 290889 |
+
"learning_rate": 7.585201938610663e-06,
|
| 290890 |
+
"loss": 0.3109,
|
| 290891 |
+
"step": 150535
|
| 290892 |
+
},
|
| 290893 |
+
{
|
| 290894 |
+
"epoch": 1203.7,
|
| 290895 |
+
"learning_rate": 7.585121163166398e-06,
|
| 290896 |
+
"loss": 0.3193,
|
| 290897 |
+
"step": 150540
|
| 290898 |
+
},
|
| 290899 |
+
{
|
| 290900 |
+
"epoch": 1203.74,
|
| 290901 |
+
"learning_rate": 7.585040387722133e-06,
|
| 290902 |
+
"loss": 0.2706,
|
| 290903 |
+
"step": 150545
|
| 290904 |
+
},
|
| 290905 |
+
{
|
| 290906 |
+
"epoch": 1203.78,
|
| 290907 |
+
"learning_rate": 7.584959612277868e-06,
|
| 290908 |
+
"loss": 0.6651,
|
| 290909 |
+
"step": 150550
|
| 290910 |
+
},
|
| 290911 |
+
{
|
| 290912 |
+
"epoch": 1203.82,
|
| 290913 |
+
"learning_rate": 7.584878836833603e-06,
|
| 290914 |
+
"loss": 0.9646,
|
| 290915 |
+
"step": 150555
|
| 290916 |
+
},
|
| 290917 |
+
{
|
| 290918 |
+
"epoch": 1203.86,
|
| 290919 |
+
"learning_rate": 7.584798061389338e-06,
|
| 290920 |
+
"loss": 0.284,
|
| 290921 |
+
"step": 150560
|
| 290922 |
+
},
|
| 290923 |
+
{
|
| 290924 |
+
"epoch": 1203.9,
|
| 290925 |
+
"learning_rate": 7.584717285945073e-06,
|
| 290926 |
+
"loss": 0.2926,
|
| 290927 |
+
"step": 150565
|
| 290928 |
+
},
|
| 290929 |
+
{
|
| 290930 |
+
"epoch": 1203.94,
|
| 290931 |
+
"learning_rate": 7.584636510500808e-06,
|
| 290932 |
+
"loss": 0.427,
|
| 290933 |
+
"step": 150570
|
| 290934 |
+
},
|
| 290935 |
+
{
|
| 290936 |
+
"epoch": 1203.98,
|
| 290937 |
+
"learning_rate": 7.584555735056544e-06,
|
| 290938 |
+
"loss": 0.6464,
|
| 290939 |
+
"step": 150575
|
| 290940 |
+
},
|
| 290941 |
+
{
|
| 290942 |
+
"epoch": 1204.0,
|
| 290943 |
+
"eval_loss": 0.5391775965690613,
|
| 290944 |
+
"eval_runtime": 39.7033,
|
| 290945 |
+
"eval_samples_per_second": 21.056,
|
| 290946 |
+
"eval_steps_per_second": 0.68,
|
| 290947 |
+
"eval_wer": 0.1843401803986963,
|
| 290948 |
+
"step": 150578
|
| 290949 |
+
},
|
| 290950 |
+
{
|
| 290951 |
+
"epoch": 1214.02,
|
| 290952 |
+
"learning_rate": 7.584474959612278e-06,
|
| 290953 |
+
"loss": 0.3288,
|
| 290954 |
+
"step": 150580
|
| 290955 |
+
},
|
| 290956 |
+
{
|
| 290957 |
+
"epoch": 1214.06,
|
| 290958 |
+
"learning_rate": 7.584394184168014e-06,
|
| 290959 |
+
"loss": 0.9793,
|
| 290960 |
+
"step": 150585
|
| 290961 |
+
},
|
| 290962 |
+
{
|
| 290963 |
+
"epoch": 1214.1,
|
| 290964 |
+
"learning_rate": 7.584313408723749e-06,
|
| 290965 |
+
"loss": 0.2777,
|
| 290966 |
+
"step": 150590
|
| 290967 |
+
},
|
| 290968 |
+
{
|
| 290969 |
+
"epoch": 1214.14,
|
| 290970 |
+
"learning_rate": 7.584232633279484e-06,
|
| 290971 |
+
"loss": 0.3273,
|
| 290972 |
+
"step": 150595
|
| 290973 |
+
},
|
| 290974 |
+
{
|
| 290975 |
+
"epoch": 1214.18,
|
| 290976 |
+
"learning_rate": 7.584151857835219e-06,
|
| 290977 |
+
"loss": 0.5917,
|
| 290978 |
+
"step": 150600
|
| 290979 |
+
},
|
| 290980 |
+
{
|
| 290981 |
+
"epoch": 1214.22,
|
| 290982 |
+
"learning_rate": 7.584071082390954e-06,
|
| 290983 |
+
"loss": 0.8979,
|
| 290984 |
+
"step": 150605
|
| 290985 |
+
},
|
| 290986 |
+
{
|
| 290987 |
+
"epoch": 1214.26,
|
| 290988 |
+
"learning_rate": 7.583990306946689e-06,
|
| 290989 |
+
"loss": 0.3145,
|
| 290990 |
+
"step": 150610
|
| 290991 |
+
},
|
| 290992 |
+
{
|
| 290993 |
+
"epoch": 1214.3,
|
| 290994 |
+
"learning_rate": 7.583909531502424e-06,
|
| 290995 |
+
"loss": 0.2725,
|
| 290996 |
+
"step": 150615
|
| 290997 |
+
},
|
| 290998 |
+
{
|
| 290999 |
+
"epoch": 1214.34,
|
| 291000 |
+
"learning_rate": 7.583828756058159e-06,
|
| 291001 |
+
"loss": 0.3622,
|
| 291002 |
+
"step": 150620
|
| 291003 |
+
},
|
| 291004 |
+
{
|
| 291005 |
+
"epoch": 1214.38,
|
| 291006 |
+
"learning_rate": 7.583747980613894e-06,
|
| 291007 |
+
"loss": 0.5381,
|
| 291008 |
+
"step": 150625
|
| 291009 |
+
},
|
| 291010 |
+
{
|
| 291011 |
+
"epoch": 1214.42,
|
| 291012 |
+
"learning_rate": 7.583667205169629e-06,
|
| 291013 |
+
"loss": 1.0317,
|
| 291014 |
+
"step": 150630
|
| 291015 |
+
},
|
| 291016 |
+
{
|
| 291017 |
+
"epoch": 1214.46,
|
| 291018 |
+
"learning_rate": 7.583586429725364e-06,
|
| 291019 |
+
"loss": 0.3174,
|
| 291020 |
+
"step": 150635
|
| 291021 |
+
},
|
| 291022 |
+
{
|
| 291023 |
+
"epoch": 1214.5,
|
| 291024 |
+
"learning_rate": 7.5835056542811e-06,
|
| 291025 |
+
"loss": 1.059,
|
| 291026 |
+
"step": 150640
|
| 291027 |
+
},
|
| 291028 |
+
{
|
| 291029 |
+
"epoch": 1214.54,
|
| 291030 |
+
"learning_rate": 7.583424878836835e-06,
|
| 291031 |
+
"loss": 0.355,
|
| 291032 |
+
"step": 150645
|
| 291033 |
+
},
|
| 291034 |
+
{
|
| 291035 |
+
"epoch": 1214.58,
|
| 291036 |
+
"learning_rate": 7.58334410339257e-06,
|
| 291037 |
+
"loss": 0.6424,
|
| 291038 |
+
"step": 150650
|
| 291039 |
+
},
|
| 291040 |
+
{
|
| 291041 |
+
"epoch": 1214.62,
|
| 291042 |
+
"learning_rate": 7.5832633279483046e-06,
|
| 291043 |
+
"loss": 1.0002,
|
| 291044 |
+
"step": 150655
|
| 291045 |
+
},
|
| 291046 |
+
{
|
| 291047 |
+
"epoch": 1214.66,
|
| 291048 |
+
"learning_rate": 7.5831825525040396e-06,
|
| 291049 |
+
"loss": 0.2473,
|
| 291050 |
+
"step": 150660
|
| 291051 |
+
},
|
| 291052 |
+
{
|
| 291053 |
+
"epoch": 1214.7,
|
| 291054 |
+
"learning_rate": 7.5831017770597745e-06,
|
| 291055 |
+
"loss": 0.2639,
|
| 291056 |
+
"step": 150665
|
| 291057 |
+
},
|
| 291058 |
+
{
|
| 291059 |
+
"epoch": 1214.74,
|
| 291060 |
+
"learning_rate": 7.5830210016155095e-06,
|
| 291061 |
+
"loss": 0.3395,
|
| 291062 |
+
"step": 150670
|
| 291063 |
+
},
|
| 291064 |
+
{
|
| 291065 |
+
"epoch": 1214.78,
|
| 291066 |
+
"learning_rate": 7.5829402261712445e-06,
|
| 291067 |
+
"loss": 0.5267,
|
| 291068 |
+
"step": 150675
|
| 291069 |
+
},
|
| 291070 |
+
{
|
| 291071 |
+
"epoch": 1214.82,
|
| 291072 |
+
"learning_rate": 7.5828594507269795e-06,
|
| 291073 |
+
"loss": 1.1767,
|
| 291074 |
+
"step": 150680
|
| 291075 |
+
},
|
| 291076 |
+
{
|
| 291077 |
+
"epoch": 1214.86,
|
| 291078 |
+
"learning_rate": 7.5827786752827145e-06,
|
| 291079 |
+
"loss": 0.2499,
|
| 291080 |
+
"step": 150685
|
| 291081 |
+
},
|
| 291082 |
+
{
|
| 291083 |
+
"epoch": 1214.9,
|
| 291084 |
+
"learning_rate": 7.5826978998384495e-06,
|
| 291085 |
+
"loss": 0.3152,
|
| 291086 |
+
"step": 150690
|
| 291087 |
+
},
|
| 291088 |
+
{
|
| 291089 |
+
"epoch": 1214.94,
|
| 291090 |
+
"learning_rate": 7.5826171243941845e-06,
|
| 291091 |
+
"loss": 0.3273,
|
| 291092 |
+
"step": 150695
|
| 291093 |
+
},
|
| 291094 |
+
{
|
| 291095 |
+
"epoch": 1214.98,
|
| 291096 |
+
"learning_rate": 7.5825363489499195e-06,
|
| 291097 |
+
"loss": 0.5964,
|
| 291098 |
+
"step": 150700
|
| 291099 |
+
},
|
| 291100 |
+
{
|
| 291101 |
+
"epoch": 1215.0,
|
| 291102 |
+
"eval_loss": 0.3669835031032562,
|
| 291103 |
+
"eval_runtime": 42.421,
|
| 291104 |
+
"eval_samples_per_second": 19.707,
|
| 291105 |
+
"eval_steps_per_second": 0.636,
|
| 291106 |
+
"eval_wer": 0.16976108232584916,
|
| 291107 |
+
"step": 150702
|
| 291108 |
+
},
|
| 291109 |
+
{
|
| 291110 |
+
"epoch": 1205.02,
|
| 291111 |
+
"learning_rate": 7.5824555735056544e-06,
|
| 291112 |
+
"loss": 0.3077,
|
| 291113 |
+
"step": 150705
|
| 291114 |
+
},
|
| 291115 |
+
{
|
| 291116 |
+
"epoch": 1205.06,
|
| 291117 |
+
"learning_rate": 7.58237479806139e-06,
|
| 291118 |
+
"loss": 0.294,
|
| 291119 |
+
"step": 150710
|
| 291120 |
+
},
|
| 291121 |
+
{
|
| 291122 |
+
"epoch": 1205.1,
|
| 291123 |
+
"learning_rate": 7.582294022617125e-06,
|
| 291124 |
+
"loss": 0.3353,
|
| 291125 |
+
"step": 150715
|
| 291126 |
+
},
|
| 291127 |
+
{
|
| 291128 |
+
"epoch": 1205.14,
|
| 291129 |
+
"learning_rate": 7.58221324717286e-06,
|
| 291130 |
+
"loss": 0.3076,
|
| 291131 |
+
"step": 150720
|
| 291132 |
+
},
|
| 291133 |
+
{
|
| 291134 |
+
"epoch": 1205.18,
|
| 291135 |
+
"learning_rate": 7.582132471728595e-06,
|
| 291136 |
+
"loss": 0.6049,
|
| 291137 |
+
"step": 150725
|
| 291138 |
+
},
|
| 291139 |
+
{
|
| 291140 |
+
"epoch": 1205.22,
|
| 291141 |
+
"learning_rate": 7.58205169628433e-06,
|
| 291142 |
+
"loss": 0.8056,
|
| 291143 |
+
"step": 150730
|
| 291144 |
+
},
|
| 291145 |
+
{
|
| 291146 |
+
"epoch": 1205.26,
|
| 291147 |
+
"learning_rate": 7.581970920840065e-06,
|
| 291148 |
+
"loss": 0.2615,
|
| 291149 |
+
"step": 150735
|
| 291150 |
+
},
|
| 291151 |
+
{
|
| 291152 |
+
"epoch": 1205.3,
|
| 291153 |
+
"learning_rate": 7.5818901453958e-06,
|
| 291154 |
+
"loss": 0.2697,
|
| 291155 |
+
"step": 150740
|
| 291156 |
+
},
|
| 291157 |
+
{
|
| 291158 |
+
"epoch": 1205.34,
|
| 291159 |
+
"learning_rate": 7.581809369951535e-06,
|
| 291160 |
+
"loss": 0.3135,
|
| 291161 |
+
"step": 150745
|
| 291162 |
+
},
|
| 291163 |
+
{
|
| 291164 |
+
"epoch": 1205.38,
|
| 291165 |
+
"learning_rate": 7.58172859450727e-06,
|
| 291166 |
+
"loss": 0.7361,
|
| 291167 |
+
"step": 150750
|
| 291168 |
+
},
|
| 291169 |
+
{
|
| 291170 |
+
"epoch": 1205.42,
|
| 291171 |
+
"learning_rate": 7.581647819063005e-06,
|
| 291172 |
+
"loss": 0.778,
|
| 291173 |
+
"step": 150755
|
| 291174 |
+
},
|
| 291175 |
+
{
|
| 291176 |
+
"epoch": 1205.46,
|
| 291177 |
+
"learning_rate": 7.58156704361874e-06,
|
| 291178 |
+
"loss": 0.2723,
|
| 291179 |
+
"step": 150760
|
| 291180 |
+
},
|
| 291181 |
+
{
|
| 291182 |
+
"epoch": 1205.5,
|
| 291183 |
+
"learning_rate": 7.581486268174476e-06,
|
| 291184 |
+
"loss": 0.3067,
|
| 291185 |
+
"step": 150765
|
| 291186 |
+
},
|
| 291187 |
+
{
|
| 291188 |
+
"epoch": 1205.54,
|
| 291189 |
+
"learning_rate": 7.58140549273021e-06,
|
| 291190 |
+
"loss": 0.2966,
|
| 291191 |
+
"step": 150770
|
| 291192 |
+
},
|
| 291193 |
+
{
|
| 291194 |
+
"epoch": 1205.58,
|
| 291195 |
+
"learning_rate": 7.581324717285946e-06,
|
| 291196 |
+
"loss": 0.7155,
|
| 291197 |
+
"step": 150775
|
| 291198 |
+
},
|
| 291199 |
+
{
|
| 291200 |
+
"epoch": 1205.62,
|
| 291201 |
+
"learning_rate": 7.581243941841681e-06,
|
| 291202 |
+
"loss": 0.9073,
|
| 291203 |
+
"step": 150780
|
| 291204 |
+
},
|
| 291205 |
+
{
|
| 291206 |
+
"epoch": 1205.66,
|
| 291207 |
+
"learning_rate": 7.581163166397416e-06,
|
| 291208 |
+
"loss": 0.2664,
|
| 291209 |
+
"step": 150785
|
| 291210 |
+
},
|
| 291211 |
+
{
|
| 291212 |
+
"epoch": 1205.7,
|
| 291213 |
+
"learning_rate": 7.581082390953151e-06,
|
| 291214 |
+
"loss": 0.2695,
|
| 291215 |
+
"step": 150790
|
| 291216 |
+
},
|
| 291217 |
+
{
|
| 291218 |
+
"epoch": 1205.74,
|
| 291219 |
+
"learning_rate": 7.581001615508886e-06,
|
| 291220 |
+
"loss": 0.3299,
|
| 291221 |
+
"step": 150795
|
| 291222 |
+
},
|
| 291223 |
+
{
|
| 291224 |
+
"epoch": 1205.78,
|
| 291225 |
+
"learning_rate": 7.580920840064621e-06,
|
| 291226 |
+
"loss": 0.7078,
|
| 291227 |
+
"step": 150800
|
| 291228 |
+
},
|
| 291229 |
+
{
|
| 291230 |
+
"epoch": 1205.82,
|
| 291231 |
+
"learning_rate": 7.580840064620356e-06,
|
| 291232 |
+
"loss": 0.9233,
|
| 291233 |
+
"step": 150805
|
| 291234 |
+
},
|
| 291235 |
+
{
|
| 291236 |
+
"epoch": 1205.86,
|
| 291237 |
+
"learning_rate": 7.580759289176091e-06,
|
| 291238 |
+
"loss": 0.2724,
|
| 291239 |
+
"step": 150810
|
| 291240 |
+
},
|
| 291241 |
+
{
|
| 291242 |
+
"epoch": 1205.9,
|
| 291243 |
+
"learning_rate": 7.580678513731826e-06,
|
| 291244 |
+
"loss": 0.3251,
|
| 291245 |
+
"step": 150815
|
| 291246 |
+
},
|
| 291247 |
+
{
|
| 291248 |
+
"epoch": 1205.94,
|
| 291249 |
+
"learning_rate": 7.580597738287562e-06,
|
| 291250 |
+
"loss": 0.3044,
|
| 291251 |
+
"step": 150820
|
| 291252 |
+
},
|
| 291253 |
+
{
|
| 291254 |
+
"epoch": 1205.98,
|
| 291255 |
+
"learning_rate": 7.580516962843296e-06,
|
| 291256 |
+
"loss": 0.614,
|
| 291257 |
+
"step": 150825
|
| 291258 |
+
},
|
| 291259 |
+
{
|
| 291260 |
+
"epoch": 1206.0,
|
| 291261 |
+
"eval_loss": 0.3523024320602417,
|
| 291262 |
+
"eval_runtime": 40.2575,
|
| 291263 |
+
"eval_samples_per_second": 20.766,
|
| 291264 |
+
"eval_steps_per_second": 0.671,
|
| 291265 |
+
"eval_wer": 0.182829413960433,
|
| 291266 |
+
"step": 150827
|
| 291267 |
}
|
| 291268 |
],
|
| 291269 |
+
"max_steps": 625000,
|
| 291270 |
"num_train_epochs": 5000,
|
| 291271 |
+
"total_flos": 4.2444315493114164e+20,
|
| 291272 |
"trial_name": null,
|
| 291273 |
"trial_params": null
|
| 291274 |
}
|
model-bin/finetune/base/{checkpoint-150205 β checkpoint-150827}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1630244012.7530751/events.out.tfevents.1630244012.cc93b136ebf5.1086.179
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:be26c9daead194841c91de81f386a51de6ca433a7e1908af359da8e170158788
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630244448.5792234/events.out.tfevents.1630244448.cc93b136ebf5.1086.181
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:52d266b5f028fc38c9d1a6a414c1c393506f6832953cb7488df1efa39b8bf893
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630244993.356596/events.out.tfevents.1630244993.cc93b136ebf5.1086.183
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:db7a02b0a48cfebf6119ec88b66e2fb707cbbd8d04c1f4c7556763fea904dbd8
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630245427.5586958/events.out.tfevents.1630245427.cc93b136ebf5.1086.185
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c397d04b557890d4c7955c897cc3fe2286c2fbf5b49d17e2f205594fa1ca5726
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630245858.8737807/events.out.tfevents.1630245858.cc93b136ebf5.1086.187
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:816366c9a07099ac79c7220c098da88f70aabc2b8cfd69ef728187b81a69a4f1
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630244012.cc93b136ebf5.1086.178
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0aa3f0de0578fd1d28673ee4fdbb0ebfb66f41ea2387c6285d9df7e4d33ecad0
|
| 3 |
+
size 8462
|
model-bin/finetune/base/log/events.out.tfevents.1630244448.cc93b136ebf5.1086.180
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e96fea5cd10e9f7da659de83310ebaee8c9de0e99e1f2c10547f556dcea96d12
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630244993.cc93b136ebf5.1086.182
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:afc99b9f992b08112289a38952d17cabf09acb3de0e8da1ccd48ac875887989b
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630245427.cc93b136ebf5.1086.184
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0d7955bc7c27f04e99fcd1c98830cccfa12ac8c2b0edbaff0d6a30315c0b3e1d
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630245858.cc93b136ebf5.1086.186
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ef4e6f78f20a1befa18e737a50ec4c4d90bf6a887d10552ee89424c85ea0395d
|
| 3 |
+
size 8622
|