"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-113992 β checkpoint-114613}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-113992 β checkpoint-114613}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-113992 β checkpoint-114613}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-113992 β checkpoint-114613}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-113992 β checkpoint-114613}/rng_state.pth +2 -2
- model-bin/finetune/base/{checkpoint-113992 β checkpoint-114613}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-113992 β checkpoint-114613}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-113992 β checkpoint-114613}/trainer_state.json +793 -4
- model-bin/finetune/base/{checkpoint-113992 β checkpoint-114613}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630107172.831585/events.out.tfevents.1630107172.86bb0ddabf9b.1042.1 +3 -0
- model-bin/finetune/base/log/1630107597.3949003/events.out.tfevents.1630107597.86bb0ddabf9b.1042.3 +3 -0
- model-bin/finetune/base/log/1630108127.983279/events.out.tfevents.1630108127.86bb0ddabf9b.1042.5 +3 -0
- model-bin/finetune/base/log/1630108542.9879258/events.out.tfevents.1630108542.86bb0ddabf9b.1042.7 +3 -0
- model-bin/finetune/base/log/1630108960.8081708/events.out.tfevents.1630108960.86bb0ddabf9b.1042.9 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630107172.86bb0ddabf9b.1042.0 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630107597.86bb0ddabf9b.1042.2 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630108127.86bb0ddabf9b.1042.4 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630108542.86bb0ddabf9b.1042.6 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630108960.86bb0ddabf9b.1042.8 +3 -0
model-bin/finetune/base/{checkpoint-113992 β checkpoint-114613}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-113992 β checkpoint-114613}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3dc554dec301ccc8dca1e759108d10f280284ba56da831c7924e4f12d8241e81
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-113992 β checkpoint-114613}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-113992 β checkpoint-114613}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5144d589a8f3be429110fd3438d71c50cbe2b5d96f43ebff77a52854f4e4c466
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-113992 β checkpoint-114613}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:06429aa76c5c31c9f61e7d1a7c32603b81cc59ed4d9203338065436cd62cc39b
|
| 3 |
+
size 14567
|
model-bin/finetune/base/{checkpoint-113992 β checkpoint-114613}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d93d43bbeecc2e2dd56d18ef4fdffea967e794dca54efdfd4a801ca408b71336
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-113992 β checkpoint-114613}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ffe03c0c156b224ba155a03dcb28f915736509e98b9e680245bcc2241ec2d0d4
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-113992 β checkpoint-114613}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1743826049391605,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -244398,11 +244398,800 @@
|
|
| 244398 |
"eval_steps_per_second": 0.661,
|
| 244399 |
"eval_wer": 0.18297995933778682,
|
| 244400 |
"step": 113992
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 244401 |
}
|
| 244402 |
],
|
| 244403 |
-
"max_steps":
|
| 244404 |
"num_train_epochs": 5000,
|
| 244405 |
-
"total_flos": 3.
|
| 244406 |
"trial_name": null,
|
| 244407 |
"trial_params": null
|
| 244408 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1743826049391605,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
|
| 4 |
+
"epoch": 923.995983935743,
|
| 5 |
+
"global_step": 114613,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 244398 |
"eval_steps_per_second": 0.661,
|
| 244399 |
"eval_wer": 0.18297995933778682,
|
| 244400 |
"step": 113992
|
| 244401 |
+
},
|
| 244402 |
+
{
|
| 244403 |
+
"epoch": 911.02,
|
| 244404 |
+
"learning_rate": 8.189903846153846e-06,
|
| 244405 |
+
"loss": 0.3429,
|
| 244406 |
+
"step": 113995
|
| 244407 |
+
},
|
| 244408 |
+
{
|
| 244409 |
+
"epoch": 911.06,
|
| 244410 |
+
"learning_rate": 8.189823717948718e-06,
|
| 244411 |
+
"loss": 0.3061,
|
| 244412 |
+
"step": 114000
|
| 244413 |
+
},
|
| 244414 |
+
{
|
| 244415 |
+
"epoch": 911.1,
|
| 244416 |
+
"learning_rate": 8.189743589743591e-06,
|
| 244417 |
+
"loss": 0.2955,
|
| 244418 |
+
"step": 114005
|
| 244419 |
+
},
|
| 244420 |
+
{
|
| 244421 |
+
"epoch": 911.14,
|
| 244422 |
+
"learning_rate": 8.189663461538462e-06,
|
| 244423 |
+
"loss": 0.3999,
|
| 244424 |
+
"step": 114010
|
| 244425 |
+
},
|
| 244426 |
+
{
|
| 244427 |
+
"epoch": 911.18,
|
| 244428 |
+
"learning_rate": 8.189583333333334e-06,
|
| 244429 |
+
"loss": 0.7552,
|
| 244430 |
+
"step": 114015
|
| 244431 |
+
},
|
| 244432 |
+
{
|
| 244433 |
+
"epoch": 911.22,
|
| 244434 |
+
"learning_rate": 8.189503205128205e-06,
|
| 244435 |
+
"loss": 0.7735,
|
| 244436 |
+
"step": 114020
|
| 244437 |
+
},
|
| 244438 |
+
{
|
| 244439 |
+
"epoch": 911.26,
|
| 244440 |
+
"learning_rate": 8.189423076923078e-06,
|
| 244441 |
+
"loss": 0.2605,
|
| 244442 |
+
"step": 114025
|
| 244443 |
+
},
|
| 244444 |
+
{
|
| 244445 |
+
"epoch": 911.3,
|
| 244446 |
+
"learning_rate": 8.18934294871795e-06,
|
| 244447 |
+
"loss": 0.3436,
|
| 244448 |
+
"step": 114030
|
| 244449 |
+
},
|
| 244450 |
+
{
|
| 244451 |
+
"epoch": 911.34,
|
| 244452 |
+
"learning_rate": 8.18926282051282e-06,
|
| 244453 |
+
"loss": 0.3272,
|
| 244454 |
+
"step": 114035
|
| 244455 |
+
},
|
| 244456 |
+
{
|
| 244457 |
+
"epoch": 911.38,
|
| 244458 |
+
"learning_rate": 8.189182692307694e-06,
|
| 244459 |
+
"loss": 0.7182,
|
| 244460 |
+
"step": 114040
|
| 244461 |
+
},
|
| 244462 |
+
{
|
| 244463 |
+
"epoch": 911.42,
|
| 244464 |
+
"learning_rate": 8.189102564102565e-06,
|
| 244465 |
+
"loss": 0.8876,
|
| 244466 |
+
"step": 114045
|
| 244467 |
+
},
|
| 244468 |
+
{
|
| 244469 |
+
"epoch": 911.46,
|
| 244470 |
+
"learning_rate": 8.189022435897437e-06,
|
| 244471 |
+
"loss": 0.2841,
|
| 244472 |
+
"step": 114050
|
| 244473 |
+
},
|
| 244474 |
+
{
|
| 244475 |
+
"epoch": 911.5,
|
| 244476 |
+
"learning_rate": 8.188942307692308e-06,
|
| 244477 |
+
"loss": 0.2787,
|
| 244478 |
+
"step": 114055
|
| 244479 |
+
},
|
| 244480 |
+
{
|
| 244481 |
+
"epoch": 911.54,
|
| 244482 |
+
"learning_rate": 8.188862179487181e-06,
|
| 244483 |
+
"loss": 0.3579,
|
| 244484 |
+
"step": 114060
|
| 244485 |
+
},
|
| 244486 |
+
{
|
| 244487 |
+
"epoch": 911.58,
|
| 244488 |
+
"learning_rate": 8.188782051282052e-06,
|
| 244489 |
+
"loss": 0.7733,
|
| 244490 |
+
"step": 114065
|
| 244491 |
+
},
|
| 244492 |
+
{
|
| 244493 |
+
"epoch": 911.62,
|
| 244494 |
+
"learning_rate": 8.188701923076924e-06,
|
| 244495 |
+
"loss": 0.9861,
|
| 244496 |
+
"step": 114070
|
| 244497 |
+
},
|
| 244498 |
+
{
|
| 244499 |
+
"epoch": 911.66,
|
| 244500 |
+
"learning_rate": 8.188621794871797e-06,
|
| 244501 |
+
"loss": 0.2945,
|
| 244502 |
+
"step": 114075
|
| 244503 |
+
},
|
| 244504 |
+
{
|
| 244505 |
+
"epoch": 911.7,
|
| 244506 |
+
"learning_rate": 8.188541666666668e-06,
|
| 244507 |
+
"loss": 0.3552,
|
| 244508 |
+
"step": 114080
|
| 244509 |
+
},
|
| 244510 |
+
{
|
| 244511 |
+
"epoch": 911.74,
|
| 244512 |
+
"learning_rate": 8.18846153846154e-06,
|
| 244513 |
+
"loss": 0.3424,
|
| 244514 |
+
"step": 114085
|
| 244515 |
+
},
|
| 244516 |
+
{
|
| 244517 |
+
"epoch": 911.78,
|
| 244518 |
+
"learning_rate": 8.18838141025641e-06,
|
| 244519 |
+
"loss": 0.6975,
|
| 244520 |
+
"step": 114090
|
| 244521 |
+
},
|
| 244522 |
+
{
|
| 244523 |
+
"epoch": 911.82,
|
| 244524 |
+
"learning_rate": 8.188301282051284e-06,
|
| 244525 |
+
"loss": 0.8733,
|
| 244526 |
+
"step": 114095
|
| 244527 |
+
},
|
| 244528 |
+
{
|
| 244529 |
+
"epoch": 911.86,
|
| 244530 |
+
"learning_rate": 8.188221153846153e-06,
|
| 244531 |
+
"loss": 0.298,
|
| 244532 |
+
"step": 114100
|
| 244533 |
+
},
|
| 244534 |
+
{
|
| 244535 |
+
"epoch": 911.9,
|
| 244536 |
+
"learning_rate": 8.188141025641027e-06,
|
| 244537 |
+
"loss": 0.298,
|
| 244538 |
+
"step": 114105
|
| 244539 |
+
},
|
| 244540 |
+
{
|
| 244541 |
+
"epoch": 911.94,
|
| 244542 |
+
"learning_rate": 8.188060897435898e-06,
|
| 244543 |
+
"loss": 0.4223,
|
| 244544 |
+
"step": 114110
|
| 244545 |
+
},
|
| 244546 |
+
{
|
| 244547 |
+
"epoch": 911.98,
|
| 244548 |
+
"learning_rate": 8.18798076923077e-06,
|
| 244549 |
+
"loss": 0.7582,
|
| 244550 |
+
"step": 114115
|
| 244551 |
+
},
|
| 244552 |
+
{
|
| 244553 |
+
"epoch": 912.0,
|
| 244554 |
+
"eval_loss": 0.4367474317550659,
|
| 244555 |
+
"eval_runtime": 39.6856,
|
| 244556 |
+
"eval_samples_per_second": 21.066,
|
| 244557 |
+
"eval_steps_per_second": 0.68,
|
| 244558 |
+
"eval_wer": 0.1899942163100058,
|
| 244559 |
+
"step": 114117
|
| 244560 |
+
},
|
| 244561 |
+
{
|
| 244562 |
+
"epoch": 920.02,
|
| 244563 |
+
"learning_rate": 8.18790064102564e-06,
|
| 244564 |
+
"loss": 0.2965,
|
| 244565 |
+
"step": 114120
|
| 244566 |
+
},
|
| 244567 |
+
{
|
| 244568 |
+
"epoch": 920.06,
|
| 244569 |
+
"learning_rate": 8.187820512820514e-06,
|
| 244570 |
+
"loss": 0.3006,
|
| 244571 |
+
"step": 114125
|
| 244572 |
+
},
|
| 244573 |
+
{
|
| 244574 |
+
"epoch": 920.1,
|
| 244575 |
+
"learning_rate": 8.187740384615385e-06,
|
| 244576 |
+
"loss": 0.2453,
|
| 244577 |
+
"step": 114130
|
| 244578 |
+
},
|
| 244579 |
+
{
|
| 244580 |
+
"epoch": 920.14,
|
| 244581 |
+
"learning_rate": 8.187660256410256e-06,
|
| 244582 |
+
"loss": 0.3644,
|
| 244583 |
+
"step": 114135
|
| 244584 |
+
},
|
| 244585 |
+
{
|
| 244586 |
+
"epoch": 920.18,
|
| 244587 |
+
"learning_rate": 8.18758012820513e-06,
|
| 244588 |
+
"loss": 0.8205,
|
| 244589 |
+
"step": 114140
|
| 244590 |
+
},
|
| 244591 |
+
{
|
| 244592 |
+
"epoch": 920.22,
|
| 244593 |
+
"learning_rate": 8.1875e-06,
|
| 244594 |
+
"loss": 0.8451,
|
| 244595 |
+
"step": 114145
|
| 244596 |
+
},
|
| 244597 |
+
{
|
| 244598 |
+
"epoch": 920.27,
|
| 244599 |
+
"learning_rate": 8.187419871794872e-06,
|
| 244600 |
+
"loss": 0.2713,
|
| 244601 |
+
"step": 114150
|
| 244602 |
+
},
|
| 244603 |
+
{
|
| 244604 |
+
"epoch": 920.31,
|
| 244605 |
+
"learning_rate": 8.187339743589744e-06,
|
| 244606 |
+
"loss": 0.2795,
|
| 244607 |
+
"step": 114155
|
| 244608 |
+
},
|
| 244609 |
+
{
|
| 244610 |
+
"epoch": 920.35,
|
| 244611 |
+
"learning_rate": 8.187259615384617e-06,
|
| 244612 |
+
"loss": 0.3686,
|
| 244613 |
+
"step": 114160
|
| 244614 |
+
},
|
| 244615 |
+
{
|
| 244616 |
+
"epoch": 920.39,
|
| 244617 |
+
"learning_rate": 8.187179487179488e-06,
|
| 244618 |
+
"loss": 0.8049,
|
| 244619 |
+
"step": 114165
|
| 244620 |
+
},
|
| 244621 |
+
{
|
| 244622 |
+
"epoch": 920.43,
|
| 244623 |
+
"learning_rate": 8.18709935897436e-06,
|
| 244624 |
+
"loss": 0.9125,
|
| 244625 |
+
"step": 114170
|
| 244626 |
+
},
|
| 244627 |
+
{
|
| 244628 |
+
"epoch": 920.47,
|
| 244629 |
+
"learning_rate": 8.18701923076923e-06,
|
| 244630 |
+
"loss": 0.237,
|
| 244631 |
+
"step": 114175
|
| 244632 |
+
},
|
| 244633 |
+
{
|
| 244634 |
+
"epoch": 920.51,
|
| 244635 |
+
"learning_rate": 8.186939102564104e-06,
|
| 244636 |
+
"loss": 0.3213,
|
| 244637 |
+
"step": 114180
|
| 244638 |
+
},
|
| 244639 |
+
{
|
| 244640 |
+
"epoch": 920.55,
|
| 244641 |
+
"learning_rate": 8.186858974358975e-06,
|
| 244642 |
+
"loss": 0.3379,
|
| 244643 |
+
"step": 114185
|
| 244644 |
+
},
|
| 244645 |
+
{
|
| 244646 |
+
"epoch": 920.59,
|
| 244647 |
+
"learning_rate": 8.186778846153846e-06,
|
| 244648 |
+
"loss": 0.6896,
|
| 244649 |
+
"step": 114190
|
| 244650 |
+
},
|
| 244651 |
+
{
|
| 244652 |
+
"epoch": 920.63,
|
| 244653 |
+
"learning_rate": 8.18669871794872e-06,
|
| 244654 |
+
"loss": 0.8142,
|
| 244655 |
+
"step": 114195
|
| 244656 |
+
},
|
| 244657 |
+
{
|
| 244658 |
+
"epoch": 920.67,
|
| 244659 |
+
"learning_rate": 8.18661858974359e-06,
|
| 244660 |
+
"loss": 0.327,
|
| 244661 |
+
"step": 114200
|
| 244662 |
+
},
|
| 244663 |
+
{
|
| 244664 |
+
"epoch": 920.71,
|
| 244665 |
+
"learning_rate": 8.186538461538462e-06,
|
| 244666 |
+
"loss": 0.3287,
|
| 244667 |
+
"step": 114205
|
| 244668 |
+
},
|
| 244669 |
+
{
|
| 244670 |
+
"epoch": 920.75,
|
| 244671 |
+
"learning_rate": 8.186458333333334e-06,
|
| 244672 |
+
"loss": 0.35,
|
| 244673 |
+
"step": 114210
|
| 244674 |
+
},
|
| 244675 |
+
{
|
| 244676 |
+
"epoch": 920.79,
|
| 244677 |
+
"learning_rate": 8.186378205128207e-06,
|
| 244678 |
+
"loss": 0.8775,
|
| 244679 |
+
"step": 114215
|
| 244680 |
+
},
|
| 244681 |
+
{
|
| 244682 |
+
"epoch": 920.83,
|
| 244683 |
+
"learning_rate": 8.186298076923076e-06,
|
| 244684 |
+
"loss": 0.8496,
|
| 244685 |
+
"step": 114220
|
| 244686 |
+
},
|
| 244687 |
+
{
|
| 244688 |
+
"epoch": 920.87,
|
| 244689 |
+
"learning_rate": 8.18621794871795e-06,
|
| 244690 |
+
"loss": 0.2979,
|
| 244691 |
+
"step": 114225
|
| 244692 |
+
},
|
| 244693 |
+
{
|
| 244694 |
+
"epoch": 920.91,
|
| 244695 |
+
"learning_rate": 8.186137820512822e-06,
|
| 244696 |
+
"loss": 0.2862,
|
| 244697 |
+
"step": 114230
|
| 244698 |
+
},
|
| 244699 |
+
{
|
| 244700 |
+
"epoch": 920.95,
|
| 244701 |
+
"learning_rate": 8.186057692307692e-06,
|
| 244702 |
+
"loss": 0.3481,
|
| 244703 |
+
"step": 114235
|
| 244704 |
+
},
|
| 244705 |
+
{
|
| 244706 |
+
"epoch": 920.99,
|
| 244707 |
+
"learning_rate": 8.185977564102565e-06,
|
| 244708 |
+
"loss": 0.7926,
|
| 244709 |
+
"step": 114240
|
| 244710 |
+
},
|
| 244711 |
+
{
|
| 244712 |
+
"epoch": 921.0,
|
| 244713 |
+
"eval_loss": 0.47359028458595276,
|
| 244714 |
+
"eval_runtime": 48.9621,
|
| 244715 |
+
"eval_samples_per_second": 17.095,
|
| 244716 |
+
"eval_steps_per_second": 0.551,
|
| 244717 |
+
"eval_wer": 0.17835125448028674,
|
| 244718 |
+
"step": 114241
|
| 244719 |
+
},
|
| 244720 |
+
{
|
| 244721 |
+
"epoch": 921.03,
|
| 244722 |
+
"learning_rate": 8.185897435897436e-06,
|
| 244723 |
+
"loss": 0.3033,
|
| 244724 |
+
"step": 114245
|
| 244725 |
+
},
|
| 244726 |
+
{
|
| 244727 |
+
"epoch": 921.07,
|
| 244728 |
+
"learning_rate": 8.185817307692308e-06,
|
| 244729 |
+
"loss": 0.2835,
|
| 244730 |
+
"step": 114250
|
| 244731 |
+
},
|
| 244732 |
+
{
|
| 244733 |
+
"epoch": 921.11,
|
| 244734 |
+
"learning_rate": 8.18573717948718e-06,
|
| 244735 |
+
"loss": 0.2973,
|
| 244736 |
+
"step": 114255
|
| 244737 |
+
},
|
| 244738 |
+
{
|
| 244739 |
+
"epoch": 921.15,
|
| 244740 |
+
"learning_rate": 8.185657051282052e-06,
|
| 244741 |
+
"loss": 0.3924,
|
| 244742 |
+
"step": 114260
|
| 244743 |
+
},
|
| 244744 |
+
{
|
| 244745 |
+
"epoch": 921.19,
|
| 244746 |
+
"learning_rate": 8.185576923076924e-06,
|
| 244747 |
+
"loss": 0.8991,
|
| 244748 |
+
"step": 114265
|
| 244749 |
+
},
|
| 244750 |
+
{
|
| 244751 |
+
"epoch": 921.23,
|
| 244752 |
+
"learning_rate": 8.185496794871795e-06,
|
| 244753 |
+
"loss": 0.7195,
|
| 244754 |
+
"step": 114270
|
| 244755 |
+
},
|
| 244756 |
+
{
|
| 244757 |
+
"epoch": 921.27,
|
| 244758 |
+
"learning_rate": 8.185416666666666e-06,
|
| 244759 |
+
"loss": 0.2874,
|
| 244760 |
+
"step": 114275
|
| 244761 |
+
},
|
| 244762 |
+
{
|
| 244763 |
+
"epoch": 921.31,
|
| 244764 |
+
"learning_rate": 8.18533653846154e-06,
|
| 244765 |
+
"loss": 0.3402,
|
| 244766 |
+
"step": 114280
|
| 244767 |
+
},
|
| 244768 |
+
{
|
| 244769 |
+
"epoch": 921.35,
|
| 244770 |
+
"learning_rate": 8.18525641025641e-06,
|
| 244771 |
+
"loss": 0.4055,
|
| 244772 |
+
"step": 114285
|
| 244773 |
+
},
|
| 244774 |
+
{
|
| 244775 |
+
"epoch": 921.39,
|
| 244776 |
+
"learning_rate": 8.185176282051282e-06,
|
| 244777 |
+
"loss": 1.2911,
|
| 244778 |
+
"step": 114290
|
| 244779 |
+
},
|
| 244780 |
+
{
|
| 244781 |
+
"epoch": 921.43,
|
| 244782 |
+
"learning_rate": 8.185096153846155e-06,
|
| 244783 |
+
"loss": 0.6954,
|
| 244784 |
+
"step": 114295
|
| 244785 |
+
},
|
| 244786 |
+
{
|
| 244787 |
+
"epoch": 921.47,
|
| 244788 |
+
"learning_rate": 8.185016025641026e-06,
|
| 244789 |
+
"loss": 0.298,
|
| 244790 |
+
"step": 114300
|
| 244791 |
+
},
|
| 244792 |
+
{
|
| 244793 |
+
"epoch": 921.51,
|
| 244794 |
+
"learning_rate": 8.184935897435898e-06,
|
| 244795 |
+
"loss": 0.2934,
|
| 244796 |
+
"step": 114305
|
| 244797 |
+
},
|
| 244798 |
+
{
|
| 244799 |
+
"epoch": 921.55,
|
| 244800 |
+
"learning_rate": 8.18485576923077e-06,
|
| 244801 |
+
"loss": 0.4285,
|
| 244802 |
+
"step": 114310
|
| 244803 |
+
},
|
| 244804 |
+
{
|
| 244805 |
+
"epoch": 921.59,
|
| 244806 |
+
"learning_rate": 8.184775641025642e-06,
|
| 244807 |
+
"loss": 0.8933,
|
| 244808 |
+
"step": 114315
|
| 244809 |
+
},
|
| 244810 |
+
{
|
| 244811 |
+
"epoch": 921.63,
|
| 244812 |
+
"learning_rate": 8.184695512820514e-06,
|
| 244813 |
+
"loss": 0.9413,
|
| 244814 |
+
"step": 114320
|
| 244815 |
+
},
|
| 244816 |
+
{
|
| 244817 |
+
"epoch": 921.67,
|
| 244818 |
+
"learning_rate": 8.184615384615385e-06,
|
| 244819 |
+
"loss": 0.2615,
|
| 244820 |
+
"step": 114325
|
| 244821 |
+
},
|
| 244822 |
+
{
|
| 244823 |
+
"epoch": 921.71,
|
| 244824 |
+
"learning_rate": 8.184535256410258e-06,
|
| 244825 |
+
"loss": 0.3053,
|
| 244826 |
+
"step": 114330
|
| 244827 |
+
},
|
| 244828 |
+
{
|
| 244829 |
+
"epoch": 921.76,
|
| 244830 |
+
"learning_rate": 8.18445512820513e-06,
|
| 244831 |
+
"loss": 0.4418,
|
| 244832 |
+
"step": 114335
|
| 244833 |
+
},
|
| 244834 |
+
{
|
| 244835 |
+
"epoch": 921.8,
|
| 244836 |
+
"learning_rate": 8.184375e-06,
|
| 244837 |
+
"loss": 0.7972,
|
| 244838 |
+
"step": 114340
|
| 244839 |
+
},
|
| 244840 |
+
{
|
| 244841 |
+
"epoch": 921.84,
|
| 244842 |
+
"learning_rate": 8.184294871794872e-06,
|
| 244843 |
+
"loss": 0.697,
|
| 244844 |
+
"step": 114345
|
| 244845 |
+
},
|
| 244846 |
+
{
|
| 244847 |
+
"epoch": 921.88,
|
| 244848 |
+
"learning_rate": 8.184214743589745e-06,
|
| 244849 |
+
"loss": 0.2833,
|
| 244850 |
+
"step": 114350
|
| 244851 |
+
},
|
| 244852 |
+
{
|
| 244853 |
+
"epoch": 921.92,
|
| 244854 |
+
"learning_rate": 8.184134615384617e-06,
|
| 244855 |
+
"loss": 0.3443,
|
| 244856 |
+
"step": 114355
|
| 244857 |
+
},
|
| 244858 |
+
{
|
| 244859 |
+
"epoch": 921.96,
|
| 244860 |
+
"learning_rate": 8.184054487179488e-06,
|
| 244861 |
+
"loss": 0.4554,
|
| 244862 |
+
"step": 114360
|
| 244863 |
+
},
|
| 244864 |
+
{
|
| 244865 |
+
"epoch": 922.0,
|
| 244866 |
+
"learning_rate": 8.18397435897436e-06,
|
| 244867 |
+
"loss": 1.1524,
|
| 244868 |
+
"step": 114365
|
| 244869 |
+
},
|
| 244870 |
+
{
|
| 244871 |
+
"epoch": 922.0,
|
| 244872 |
+
"eval_loss": 0.4351113736629486,
|
| 244873 |
+
"eval_runtime": 39.0293,
|
| 244874 |
+
"eval_samples_per_second": 21.599,
|
| 244875 |
+
"eval_steps_per_second": 0.692,
|
| 244876 |
+
"eval_wer": 0.18565555718690355,
|
| 244877 |
+
"step": 114365
|
| 244878 |
+
},
|
| 244879 |
+
{
|
| 244880 |
+
"epoch": 922.04,
|
| 244881 |
+
"learning_rate": 8.183894230769232e-06,
|
| 244882 |
+
"loss": 0.3181,
|
| 244883 |
+
"step": 114370
|
| 244884 |
+
},
|
| 244885 |
+
{
|
| 244886 |
+
"epoch": 922.08,
|
| 244887 |
+
"learning_rate": 8.183814102564102e-06,
|
| 244888 |
+
"loss": 0.287,
|
| 244889 |
+
"step": 114375
|
| 244890 |
+
},
|
| 244891 |
+
{
|
| 244892 |
+
"epoch": 922.12,
|
| 244893 |
+
"learning_rate": 8.183733974358975e-06,
|
| 244894 |
+
"loss": 0.3079,
|
| 244895 |
+
"step": 114380
|
| 244896 |
+
},
|
| 244897 |
+
{
|
| 244898 |
+
"epoch": 922.16,
|
| 244899 |
+
"learning_rate": 8.183653846153848e-06,
|
| 244900 |
+
"loss": 0.3877,
|
| 244901 |
+
"step": 114385
|
| 244902 |
+
},
|
| 244903 |
+
{
|
| 244904 |
+
"epoch": 922.2,
|
| 244905 |
+
"learning_rate": 8.183573717948718e-06,
|
| 244906 |
+
"loss": 1.0876,
|
| 244907 |
+
"step": 114390
|
| 244908 |
+
},
|
| 244909 |
+
{
|
| 244910 |
+
"epoch": 922.24,
|
| 244911 |
+
"learning_rate": 8.18349358974359e-06,
|
| 244912 |
+
"loss": 0.315,
|
| 244913 |
+
"step": 114395
|
| 244914 |
+
},
|
| 244915 |
+
{
|
| 244916 |
+
"epoch": 922.28,
|
| 244917 |
+
"learning_rate": 8.183413461538462e-06,
|
| 244918 |
+
"loss": 0.2827,
|
| 244919 |
+
"step": 114400
|
| 244920 |
+
},
|
| 244921 |
+
{
|
| 244922 |
+
"epoch": 922.32,
|
| 244923 |
+
"learning_rate": 8.183333333333333e-06,
|
| 244924 |
+
"loss": 0.3118,
|
| 244925 |
+
"step": 114405
|
| 244926 |
+
},
|
| 244927 |
+
{
|
| 244928 |
+
"epoch": 922.36,
|
| 244929 |
+
"learning_rate": 8.183253205128205e-06,
|
| 244930 |
+
"loss": 0.4451,
|
| 244931 |
+
"step": 114410
|
| 244932 |
+
},
|
| 244933 |
+
{
|
| 244934 |
+
"epoch": 922.4,
|
| 244935 |
+
"learning_rate": 8.183189102564102e-06,
|
| 244936 |
+
"loss": 1.3298,
|
| 244937 |
+
"step": 114415
|
| 244938 |
+
},
|
| 244939 |
+
{
|
| 244940 |
+
"epoch": 922.44,
|
| 244941 |
+
"learning_rate": 8.183108974358975e-06,
|
| 244942 |
+
"loss": 0.2886,
|
| 244943 |
+
"step": 114420
|
| 244944 |
+
},
|
| 244945 |
+
{
|
| 244946 |
+
"epoch": 922.48,
|
| 244947 |
+
"learning_rate": 8.183028846153847e-06,
|
| 244948 |
+
"loss": 0.2573,
|
| 244949 |
+
"step": 114425
|
| 244950 |
+
},
|
| 244951 |
+
{
|
| 244952 |
+
"epoch": 922.52,
|
| 244953 |
+
"learning_rate": 8.182948717948718e-06,
|
| 244954 |
+
"loss": 0.3227,
|
| 244955 |
+
"step": 114430
|
| 244956 |
+
},
|
| 244957 |
+
{
|
| 244958 |
+
"epoch": 922.56,
|
| 244959 |
+
"learning_rate": 8.18286858974359e-06,
|
| 244960 |
+
"loss": 0.5028,
|
| 244961 |
+
"step": 114435
|
| 244962 |
+
},
|
| 244963 |
+
{
|
| 244964 |
+
"epoch": 922.6,
|
| 244965 |
+
"learning_rate": 8.182788461538462e-06,
|
| 244966 |
+
"loss": 1.246,
|
| 244967 |
+
"step": 114440
|
| 244968 |
+
},
|
| 244969 |
+
{
|
| 244970 |
+
"epoch": 922.64,
|
| 244971 |
+
"learning_rate": 8.182708333333334e-06,
|
| 244972 |
+
"loss": 0.4493,
|
| 244973 |
+
"step": 114445
|
| 244974 |
+
},
|
| 244975 |
+
{
|
| 244976 |
+
"epoch": 922.68,
|
| 244977 |
+
"learning_rate": 8.182628205128205e-06,
|
| 244978 |
+
"loss": 0.3027,
|
| 244979 |
+
"step": 114450
|
| 244980 |
+
},
|
| 244981 |
+
{
|
| 244982 |
+
"epoch": 922.72,
|
| 244983 |
+
"learning_rate": 8.182548076923078e-06,
|
| 244984 |
+
"loss": 0.3715,
|
| 244985 |
+
"step": 114455
|
| 244986 |
+
},
|
| 244987 |
+
{
|
| 244988 |
+
"epoch": 922.76,
|
| 244989 |
+
"learning_rate": 8.18246794871795e-06,
|
| 244990 |
+
"loss": 0.4394,
|
| 244991 |
+
"step": 114460
|
| 244992 |
+
},
|
| 244993 |
+
{
|
| 244994 |
+
"epoch": 922.8,
|
| 244995 |
+
"learning_rate": 8.182387820512821e-06,
|
| 244996 |
+
"loss": 1.129,
|
| 244997 |
+
"step": 114465
|
| 244998 |
+
},
|
| 244999 |
+
{
|
| 245000 |
+
"epoch": 922.84,
|
| 245001 |
+
"learning_rate": 8.182307692307692e-06,
|
| 245002 |
+
"loss": 0.3216,
|
| 245003 |
+
"step": 114470
|
| 245004 |
+
},
|
| 245005 |
+
{
|
| 245006 |
+
"epoch": 922.88,
|
| 245007 |
+
"learning_rate": 8.182227564102565e-06,
|
| 245008 |
+
"loss": 0.3146,
|
| 245009 |
+
"step": 114475
|
| 245010 |
+
},
|
| 245011 |
+
{
|
| 245012 |
+
"epoch": 922.92,
|
| 245013 |
+
"learning_rate": 8.182147435897437e-06,
|
| 245014 |
+
"loss": 0.3341,
|
| 245015 |
+
"step": 114480
|
| 245016 |
+
},
|
| 245017 |
+
{
|
| 245018 |
+
"epoch": 922.96,
|
| 245019 |
+
"learning_rate": 8.182067307692308e-06,
|
| 245020 |
+
"loss": 0.4186,
|
| 245021 |
+
"step": 114485
|
| 245022 |
+
},
|
| 245023 |
+
{
|
| 245024 |
+
"epoch": 923.0,
|
| 245025 |
+
"eval_loss": 0.4149532616138458,
|
| 245026 |
+
"eval_runtime": 39.8191,
|
| 245027 |
+
"eval_samples_per_second": 21.171,
|
| 245028 |
+
"eval_steps_per_second": 0.678,
|
| 245029 |
+
"eval_wer": 0.18946368551804765,
|
| 245030 |
+
"step": 114489
|
| 245031 |
+
},
|
| 245032 |
+
{
|
| 245033 |
+
"epoch": 923.01,
|
| 245034 |
+
"learning_rate": 8.18198717948718e-06,
|
| 245035 |
+
"loss": 0.3366,
|
| 245036 |
+
"step": 114490
|
| 245037 |
+
},
|
| 245038 |
+
{
|
| 245039 |
+
"epoch": 923.05,
|
| 245040 |
+
"learning_rate": 8.181907051282052e-06,
|
| 245041 |
+
"loss": 0.2907,
|
| 245042 |
+
"step": 114495
|
| 245043 |
+
},
|
| 245044 |
+
{
|
| 245045 |
+
"epoch": 923.09,
|
| 245046 |
+
"learning_rate": 8.181826923076924e-06,
|
| 245047 |
+
"loss": 0.3085,
|
| 245048 |
+
"step": 114500
|
| 245049 |
+
},
|
| 245050 |
+
{
|
| 245051 |
+
"epoch": 923.13,
|
| 245052 |
+
"learning_rate": 8.181746794871795e-06,
|
| 245053 |
+
"loss": 0.3446,
|
| 245054 |
+
"step": 114505
|
| 245055 |
+
},
|
| 245056 |
+
{
|
| 245057 |
+
"epoch": 923.17,
|
| 245058 |
+
"learning_rate": 8.181666666666668e-06,
|
| 245059 |
+
"loss": 0.4964,
|
| 245060 |
+
"step": 114510
|
| 245061 |
+
},
|
| 245062 |
+
{
|
| 245063 |
+
"epoch": 923.21,
|
| 245064 |
+
"learning_rate": 8.18158653846154e-06,
|
| 245065 |
+
"loss": 1.0242,
|
| 245066 |
+
"step": 114515
|
| 245067 |
+
},
|
| 245068 |
+
{
|
| 245069 |
+
"epoch": 923.25,
|
| 245070 |
+
"learning_rate": 8.181506410256411e-06,
|
| 245071 |
+
"loss": 0.3272,
|
| 245072 |
+
"step": 114520
|
| 245073 |
+
},
|
| 245074 |
+
{
|
| 245075 |
+
"epoch": 923.29,
|
| 245076 |
+
"learning_rate": 8.181426282051282e-06,
|
| 245077 |
+
"loss": 0.325,
|
| 245078 |
+
"step": 114525
|
| 245079 |
+
},
|
| 245080 |
+
{
|
| 245081 |
+
"epoch": 923.33,
|
| 245082 |
+
"learning_rate": 8.181346153846155e-06,
|
| 245083 |
+
"loss": 0.3969,
|
| 245084 |
+
"step": 114530
|
| 245085 |
+
},
|
| 245086 |
+
{
|
| 245087 |
+
"epoch": 923.37,
|
| 245088 |
+
"learning_rate": 8.181266025641027e-06,
|
| 245089 |
+
"loss": 0.4844,
|
| 245090 |
+
"step": 114535
|
| 245091 |
+
},
|
| 245092 |
+
{
|
| 245093 |
+
"epoch": 923.41,
|
| 245094 |
+
"learning_rate": 8.181185897435898e-06,
|
| 245095 |
+
"loss": 1.1032,
|
| 245096 |
+
"step": 114540
|
| 245097 |
+
},
|
| 245098 |
+
{
|
| 245099 |
+
"epoch": 923.45,
|
| 245100 |
+
"learning_rate": 8.18110576923077e-06,
|
| 245101 |
+
"loss": 0.2643,
|
| 245102 |
+
"step": 114545
|
| 245103 |
+
},
|
| 245104 |
+
{
|
| 245105 |
+
"epoch": 923.49,
|
| 245106 |
+
"learning_rate": 8.181025641025642e-06,
|
| 245107 |
+
"loss": 0.2698,
|
| 245108 |
+
"step": 114550
|
| 245109 |
+
},
|
| 245110 |
+
{
|
| 245111 |
+
"epoch": 923.53,
|
| 245112 |
+
"learning_rate": 8.180945512820514e-06,
|
| 245113 |
+
"loss": 0.2943,
|
| 245114 |
+
"step": 114555
|
| 245115 |
+
},
|
| 245116 |
+
{
|
| 245117 |
+
"epoch": 923.57,
|
| 245118 |
+
"learning_rate": 8.180865384615385e-06,
|
| 245119 |
+
"loss": 0.4995,
|
| 245120 |
+
"step": 114560
|
| 245121 |
+
},
|
| 245122 |
+
{
|
| 245123 |
+
"epoch": 923.61,
|
| 245124 |
+
"learning_rate": 8.180785256410258e-06,
|
| 245125 |
+
"loss": 1.1026,
|
| 245126 |
+
"step": 114565
|
| 245127 |
+
},
|
| 245128 |
+
{
|
| 245129 |
+
"epoch": 923.65,
|
| 245130 |
+
"learning_rate": 8.180705128205128e-06,
|
| 245131 |
+
"loss": 0.2717,
|
| 245132 |
+
"step": 114570
|
| 245133 |
+
},
|
| 245134 |
+
{
|
| 245135 |
+
"epoch": 923.69,
|
| 245136 |
+
"learning_rate": 8.180625000000001e-06,
|
| 245137 |
+
"loss": 0.2867,
|
| 245138 |
+
"step": 114575
|
| 245139 |
+
},
|
| 245140 |
+
{
|
| 245141 |
+
"epoch": 923.73,
|
| 245142 |
+
"learning_rate": 8.180544871794872e-06,
|
| 245143 |
+
"loss": 0.3375,
|
| 245144 |
+
"step": 114580
|
| 245145 |
+
},
|
| 245146 |
+
{
|
| 245147 |
+
"epoch": 923.77,
|
| 245148 |
+
"learning_rate": 8.180464743589744e-06,
|
| 245149 |
+
"loss": 0.5839,
|
| 245150 |
+
"step": 114585
|
| 245151 |
+
},
|
| 245152 |
+
{
|
| 245153 |
+
"epoch": 923.81,
|
| 245154 |
+
"learning_rate": 8.180384615384615e-06,
|
| 245155 |
+
"loss": 1.0434,
|
| 245156 |
+
"step": 114590
|
| 245157 |
+
},
|
| 245158 |
+
{
|
| 245159 |
+
"epoch": 923.85,
|
| 245160 |
+
"learning_rate": 8.180304487179488e-06,
|
| 245161 |
+
"loss": 0.3545,
|
| 245162 |
+
"step": 114595
|
| 245163 |
+
},
|
| 245164 |
+
{
|
| 245165 |
+
"epoch": 923.89,
|
| 245166 |
+
"learning_rate": 8.18022435897436e-06,
|
| 245167 |
+
"loss": 0.3341,
|
| 245168 |
+
"step": 114600
|
| 245169 |
+
},
|
| 245170 |
+
{
|
| 245171 |
+
"epoch": 923.93,
|
| 245172 |
+
"learning_rate": 8.18014423076923e-06,
|
| 245173 |
+
"loss": 0.3615,
|
| 245174 |
+
"step": 114605
|
| 245175 |
+
},
|
| 245176 |
+
{
|
| 245177 |
+
"epoch": 923.97,
|
| 245178 |
+
"learning_rate": 8.180064102564104e-06,
|
| 245179 |
+
"loss": 0.6272,
|
| 245180 |
+
"step": 114610
|
| 245181 |
+
},
|
| 245182 |
+
{
|
| 245183 |
+
"epoch": 924.0,
|
| 245184 |
+
"eval_loss": 0.3545716106891632,
|
| 245185 |
+
"eval_runtime": 39.8654,
|
| 245186 |
+
"eval_samples_per_second": 21.146,
|
| 245187 |
+
"eval_steps_per_second": 0.677,
|
| 245188 |
+
"eval_wer": 0.18582472364713531,
|
| 245189 |
+
"step": 114613
|
| 245190 |
}
|
| 245191 |
],
|
| 245192 |
+
"max_steps": 620000,
|
| 245193 |
"num_train_epochs": 5000,
|
| 245194 |
+
"total_flos": 3.225517960789425e+20,
|
| 245195 |
"trial_name": null,
|
| 245196 |
"trial_params": null
|
| 245197 |
}
|
model-bin/finetune/base/{checkpoint-113992 β checkpoint-114613}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1630107172.831585/events.out.tfevents.1630107172.86bb0ddabf9b.1042.1
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6aeda773d1e5f19ef742dd91a92e470ccc30e14faaca77578dfa690f8ba7e209
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630107597.3949003/events.out.tfevents.1630107597.86bb0ddabf9b.1042.3
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8b6f8c9b033347f38e86107122356cffdfc05824f5b13d69389580da12c3af0f
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630108127.983279/events.out.tfevents.1630108127.86bb0ddabf9b.1042.5
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0d1387b80bfffc1109dc85dc636d6e602a32e5cdb20e80fd35c0154a6fc7bfaa
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630108542.9879258/events.out.tfevents.1630108542.86bb0ddabf9b.1042.7
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b13b91412aeecb26c21fb11c54caaeef8910b368662e63848bdab048df4974e0
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630108960.8081708/events.out.tfevents.1630108960.86bb0ddabf9b.1042.9
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9de072a4f71eac89e5082ec377cee3ad05553b5e574f5235aebd7b9ffe492130
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630107172.86bb0ddabf9b.1042.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6dafb199912054a3d32e56492b11806f1f40499977fb9baca81bc4b2db159a66
|
| 3 |
+
size 8630
|
model-bin/finetune/base/log/events.out.tfevents.1630107597.86bb0ddabf9b.1042.2
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:85a44d609ba96a704a9993014a97d1360f7ebfd4e37098729a5ebcf531ad8613
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630108127.86bb0ddabf9b.1042.4
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4249aa82cc943be75382adc2fa3444c76cd80d9275bb29f3810df2c65a5c7d41
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630108542.86bb0ddabf9b.1042.6
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:98e40bee306491bc7675a24ae3340bdc28af0293266cc25fe946014ef80e9200
|
| 3 |
+
size 8462
|
model-bin/finetune/base/log/events.out.tfevents.1630108960.86bb0ddabf9b.1042.8
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2cded565932eb070edcb2a75c1ab2265dd5eb139681ba17383a44fb5241dfad1
|
| 3 |
+
size 8622
|