"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-147094 β checkpoint-147718}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-147094 β checkpoint-147718}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-147094 β checkpoint-147718}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-147094 β checkpoint-147718}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-147094 β checkpoint-147718}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-147094 β checkpoint-147718}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-147094 β checkpoint-147718}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-147094 β checkpoint-147718}/trainer_state.json +799 -4
- model-bin/finetune/base/{checkpoint-147094 β checkpoint-147718}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630232727.7092497/events.out.tfevents.1630232727.cc93b136ebf5.1086.129 +3 -0
- model-bin/finetune/base/log/1630233167.6375728/events.out.tfevents.1630233167.cc93b136ebf5.1086.131 +3 -0
- model-bin/finetune/base/log/1630233719.7330806/events.out.tfevents.1630233719.cc93b136ebf5.1086.133 +3 -0
- model-bin/finetune/base/log/1630234160.490352/events.out.tfevents.1630234160.cc93b136ebf5.1086.135 +3 -0
- model-bin/finetune/base/log/1630234598.790891/events.out.tfevents.1630234598.cc93b136ebf5.1086.137 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630232727.cc93b136ebf5.1086.128 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630233167.cc93b136ebf5.1086.130 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630233719.cc93b136ebf5.1086.132 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630234160.cc93b136ebf5.1086.134 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630234598.cc93b136ebf5.1086.136 +3 -0
model-bin/finetune/base/{checkpoint-147094 β checkpoint-147718}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-147094 β checkpoint-147718}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f9879be8cb38bc67c6efdbd6babaa521bdbb8f6513689ac3b5362caa80ec8a7d
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-147094 β checkpoint-147718}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-147094 β checkpoint-147718}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a4e0b5fa704ba7e3355f2a928cf0e19f664a24aa3d711784e533409ee00530bc
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-147094 β checkpoint-147718}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4a54ecd65da7fe967b38ffea8488cc27f6d9ddf00ed51cf5cc4cf7db84c3d919
|
| 3 |
size 14503
|
model-bin/finetune/base/{checkpoint-147094 β checkpoint-147718}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:484db9061b9a889b93aa271524f19897ef827c165a544339fe99b80d7f7cf6ea
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-147094 β checkpoint-147718}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ca25f6e59009333e7cccfa2c086b44a91ea478af3091c81fc83feafb6082dc2f
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-147094 β checkpoint-147718}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1668011442822563,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -286512,11 +286512,806 @@
|
|
| 286512 |
"eval_steps_per_second": 0.657,
|
| 286513 |
"eval_wer": 0.1767975743574935,
|
| 286514 |
"step": 147094
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 286515 |
}
|
| 286516 |
],
|
| 286517 |
-
"max_steps":
|
| 286518 |
"num_train_epochs": 5000,
|
| 286519 |
-
"total_flos": 4.
|
| 286520 |
"trial_name": null,
|
| 286521 |
"trial_params": null
|
| 286522 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1668011442822563,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
| 4 |
+
"epoch": 1181.0,
|
| 5 |
+
"global_step": 147718,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 286512 |
"eval_steps_per_second": 0.657,
|
| 286513 |
"eval_wer": 0.1767975743574935,
|
| 286514 |
"step": 147094
|
| 286515 |
+
},
|
| 286516 |
+
{
|
| 286517 |
+
"epoch": 1176.01,
|
| 286518 |
+
"learning_rate": 7.640743134087239e-06,
|
| 286519 |
+
"loss": 0.2419,
|
| 286520 |
+
"step": 147095
|
| 286521 |
+
},
|
| 286522 |
+
{
|
| 286523 |
+
"epoch": 1176.05,
|
| 286524 |
+
"learning_rate": 7.640662358642973e-06,
|
| 286525 |
+
"loss": 0.3109,
|
| 286526 |
+
"step": 147100
|
| 286527 |
+
},
|
| 286528 |
+
{
|
| 286529 |
+
"epoch": 1176.09,
|
| 286530 |
+
"learning_rate": 7.640581583198709e-06,
|
| 286531 |
+
"loss": 0.304,
|
| 286532 |
+
"step": 147105
|
| 286533 |
+
},
|
| 286534 |
+
{
|
| 286535 |
+
"epoch": 1176.13,
|
| 286536 |
+
"learning_rate": 7.640500807754443e-06,
|
| 286537 |
+
"loss": 0.3394,
|
| 286538 |
+
"step": 147110
|
| 286539 |
+
},
|
| 286540 |
+
{
|
| 286541 |
+
"epoch": 1176.17,
|
| 286542 |
+
"learning_rate": 7.640420032310179e-06,
|
| 286543 |
+
"loss": 0.4986,
|
| 286544 |
+
"step": 147115
|
| 286545 |
+
},
|
| 286546 |
+
{
|
| 286547 |
+
"epoch": 1176.21,
|
| 286548 |
+
"learning_rate": 7.640339256865913e-06,
|
| 286549 |
+
"loss": 1.0938,
|
| 286550 |
+
"step": 147120
|
| 286551 |
+
},
|
| 286552 |
+
{
|
| 286553 |
+
"epoch": 1176.25,
|
| 286554 |
+
"learning_rate": 7.640258481421649e-06,
|
| 286555 |
+
"loss": 0.2599,
|
| 286556 |
+
"step": 147125
|
| 286557 |
+
},
|
| 286558 |
+
{
|
| 286559 |
+
"epoch": 1176.29,
|
| 286560 |
+
"learning_rate": 7.640193861066236e-06,
|
| 286561 |
+
"loss": 0.9439,
|
| 286562 |
+
"step": 147130
|
| 286563 |
+
},
|
| 286564 |
+
{
|
| 286565 |
+
"epoch": 1176.33,
|
| 286566 |
+
"learning_rate": 7.640113085621972e-06,
|
| 286567 |
+
"loss": 0.3323,
|
| 286568 |
+
"step": 147135
|
| 286569 |
+
},
|
| 286570 |
+
{
|
| 286571 |
+
"epoch": 1176.37,
|
| 286572 |
+
"learning_rate": 7.640032310177706e-06,
|
| 286573 |
+
"loss": 0.4905,
|
| 286574 |
+
"step": 147140
|
| 286575 |
+
},
|
| 286576 |
+
{
|
| 286577 |
+
"epoch": 1176.41,
|
| 286578 |
+
"learning_rate": 7.639951534733442e-06,
|
| 286579 |
+
"loss": 1.1409,
|
| 286580 |
+
"step": 147145
|
| 286581 |
+
},
|
| 286582 |
+
{
|
| 286583 |
+
"epoch": 1176.45,
|
| 286584 |
+
"learning_rate": 7.639870759289176e-06,
|
| 286585 |
+
"loss": 0.288,
|
| 286586 |
+
"step": 147150
|
| 286587 |
+
},
|
| 286588 |
+
{
|
| 286589 |
+
"epoch": 1176.49,
|
| 286590 |
+
"learning_rate": 7.639789983844912e-06,
|
| 286591 |
+
"loss": 0.324,
|
| 286592 |
+
"step": 147155
|
| 286593 |
+
},
|
| 286594 |
+
{
|
| 286595 |
+
"epoch": 1176.53,
|
| 286596 |
+
"learning_rate": 7.639709208400646e-06,
|
| 286597 |
+
"loss": 0.3126,
|
| 286598 |
+
"step": 147160
|
| 286599 |
+
},
|
| 286600 |
+
{
|
| 286601 |
+
"epoch": 1176.57,
|
| 286602 |
+
"learning_rate": 7.639628432956382e-06,
|
| 286603 |
+
"loss": 0.4731,
|
| 286604 |
+
"step": 147165
|
| 286605 |
+
},
|
| 286606 |
+
{
|
| 286607 |
+
"epoch": 1176.61,
|
| 286608 |
+
"learning_rate": 7.639547657512118e-06,
|
| 286609 |
+
"loss": 1.0284,
|
| 286610 |
+
"step": 147170
|
| 286611 |
+
},
|
| 286612 |
+
{
|
| 286613 |
+
"epoch": 1176.65,
|
| 286614 |
+
"learning_rate": 7.639466882067852e-06,
|
| 286615 |
+
"loss": 0.2644,
|
| 286616 |
+
"step": 147175
|
| 286617 |
+
},
|
| 286618 |
+
{
|
| 286619 |
+
"epoch": 1176.69,
|
| 286620 |
+
"learning_rate": 7.639386106623588e-06,
|
| 286621 |
+
"loss": 0.2493,
|
| 286622 |
+
"step": 147180
|
| 286623 |
+
},
|
| 286624 |
+
{
|
| 286625 |
+
"epoch": 1176.73,
|
| 286626 |
+
"learning_rate": 7.639305331179322e-06,
|
| 286627 |
+
"loss": 0.3207,
|
| 286628 |
+
"step": 147185
|
| 286629 |
+
},
|
| 286630 |
+
{
|
| 286631 |
+
"epoch": 1176.76,
|
| 286632 |
+
"learning_rate": 7.639224555735058e-06,
|
| 286633 |
+
"loss": 0.4982,
|
| 286634 |
+
"step": 147190
|
| 286635 |
+
},
|
| 286636 |
+
{
|
| 286637 |
+
"epoch": 1176.8,
|
| 286638 |
+
"learning_rate": 7.639143780290792e-06,
|
| 286639 |
+
"loss": 0.9975,
|
| 286640 |
+
"step": 147195
|
| 286641 |
+
},
|
| 286642 |
+
{
|
| 286643 |
+
"epoch": 1176.84,
|
| 286644 |
+
"learning_rate": 7.639063004846528e-06,
|
| 286645 |
+
"loss": 0.2807,
|
| 286646 |
+
"step": 147200
|
| 286647 |
+
},
|
| 286648 |
+
{
|
| 286649 |
+
"epoch": 1176.88,
|
| 286650 |
+
"learning_rate": 7.638982229402262e-06,
|
| 286651 |
+
"loss": 0.2536,
|
| 286652 |
+
"step": 147205
|
| 286653 |
+
},
|
| 286654 |
+
{
|
| 286655 |
+
"epoch": 1176.92,
|
| 286656 |
+
"learning_rate": 7.638901453957998e-06,
|
| 286657 |
+
"loss": 0.2847,
|
| 286658 |
+
"step": 147210
|
| 286659 |
+
},
|
| 286660 |
+
{
|
| 286661 |
+
"epoch": 1176.96,
|
| 286662 |
+
"learning_rate": 7.638820678513732e-06,
|
| 286663 |
+
"loss": 0.449,
|
| 286664 |
+
"step": 147215
|
| 286665 |
+
},
|
| 286666 |
+
{
|
| 286667 |
+
"epoch": 1177.0,
|
| 286668 |
+
"eval_loss": 0.3598257005214691,
|
| 286669 |
+
"eval_runtime": 41.7013,
|
| 286670 |
+
"eval_samples_per_second": 20.071,
|
| 286671 |
+
"eval_steps_per_second": 0.647,
|
| 286672 |
+
"eval_wer": 0.17469967770290068,
|
| 286673 |
+
"step": 147219
|
| 286674 |
+
},
|
| 286675 |
+
{
|
| 286676 |
+
"epoch": 1187.01,
|
| 286677 |
+
"learning_rate": 7.638739903069468e-06,
|
| 286678 |
+
"loss": 0.3002,
|
| 286679 |
+
"step": 147220
|
| 286680 |
+
},
|
| 286681 |
+
{
|
| 286682 |
+
"epoch": 1187.05,
|
| 286683 |
+
"learning_rate": 7.638659127625202e-06,
|
| 286684 |
+
"loss": 0.2696,
|
| 286685 |
+
"step": 147225
|
| 286686 |
+
},
|
| 286687 |
+
{
|
| 286688 |
+
"epoch": 1187.09,
|
| 286689 |
+
"learning_rate": 7.638578352180938e-06,
|
| 286690 |
+
"loss": 0.3002,
|
| 286691 |
+
"step": 147230
|
| 286692 |
+
},
|
| 286693 |
+
{
|
| 286694 |
+
"epoch": 1187.13,
|
| 286695 |
+
"learning_rate": 7.638497576736673e-06,
|
| 286696 |
+
"loss": 0.2989,
|
| 286697 |
+
"step": 147235
|
| 286698 |
+
},
|
| 286699 |
+
{
|
| 286700 |
+
"epoch": 1187.17,
|
| 286701 |
+
"learning_rate": 7.638416801292407e-06,
|
| 286702 |
+
"loss": 0.5234,
|
| 286703 |
+
"step": 147240
|
| 286704 |
+
},
|
| 286705 |
+
{
|
| 286706 |
+
"epoch": 1187.21,
|
| 286707 |
+
"learning_rate": 7.638336025848143e-06,
|
| 286708 |
+
"loss": 0.9627,
|
| 286709 |
+
"step": 147245
|
| 286710 |
+
},
|
| 286711 |
+
{
|
| 286712 |
+
"epoch": 1187.25,
|
| 286713 |
+
"learning_rate": 7.638255250403877e-06,
|
| 286714 |
+
"loss": 0.3225,
|
| 286715 |
+
"step": 147250
|
| 286716 |
+
},
|
| 286717 |
+
{
|
| 286718 |
+
"epoch": 1187.29,
|
| 286719 |
+
"learning_rate": 7.638174474959613e-06,
|
| 286720 |
+
"loss": 0.3235,
|
| 286721 |
+
"step": 147255
|
| 286722 |
+
},
|
| 286723 |
+
{
|
| 286724 |
+
"epoch": 1187.33,
|
| 286725 |
+
"learning_rate": 7.638093699515347e-06,
|
| 286726 |
+
"loss": 0.4524,
|
| 286727 |
+
"step": 147260
|
| 286728 |
+
},
|
| 286729 |
+
{
|
| 286730 |
+
"epoch": 1187.37,
|
| 286731 |
+
"learning_rate": 7.638012924071083e-06,
|
| 286732 |
+
"loss": 0.5689,
|
| 286733 |
+
"step": 147265
|
| 286734 |
+
},
|
| 286735 |
+
{
|
| 286736 |
+
"epoch": 1187.41,
|
| 286737 |
+
"learning_rate": 7.637932148626817e-06,
|
| 286738 |
+
"loss": 1.2074,
|
| 286739 |
+
"step": 147270
|
| 286740 |
+
},
|
| 286741 |
+
{
|
| 286742 |
+
"epoch": 1187.45,
|
| 286743 |
+
"learning_rate": 7.637851373182553e-06,
|
| 286744 |
+
"loss": 0.2922,
|
| 286745 |
+
"step": 147275
|
| 286746 |
+
},
|
| 286747 |
+
{
|
| 286748 |
+
"epoch": 1187.49,
|
| 286749 |
+
"learning_rate": 7.637770597738287e-06,
|
| 286750 |
+
"loss": 0.2583,
|
| 286751 |
+
"step": 147280
|
| 286752 |
+
},
|
| 286753 |
+
{
|
| 286754 |
+
"epoch": 1187.53,
|
| 286755 |
+
"learning_rate": 7.637689822294023e-06,
|
| 286756 |
+
"loss": 0.3084,
|
| 286757 |
+
"step": 147285
|
| 286758 |
+
},
|
| 286759 |
+
{
|
| 286760 |
+
"epoch": 1187.57,
|
| 286761 |
+
"learning_rate": 7.637609046849759e-06,
|
| 286762 |
+
"loss": 0.4638,
|
| 286763 |
+
"step": 147290
|
| 286764 |
+
},
|
| 286765 |
+
{
|
| 286766 |
+
"epoch": 1187.61,
|
| 286767 |
+
"learning_rate": 7.637528271405493e-06,
|
| 286768 |
+
"loss": 1.0976,
|
| 286769 |
+
"step": 147295
|
| 286770 |
+
},
|
| 286771 |
+
{
|
| 286772 |
+
"epoch": 1187.65,
|
| 286773 |
+
"learning_rate": 7.637447495961229e-06,
|
| 286774 |
+
"loss": 0.2513,
|
| 286775 |
+
"step": 147300
|
| 286776 |
+
},
|
| 286777 |
+
{
|
| 286778 |
+
"epoch": 1187.69,
|
| 286779 |
+
"learning_rate": 7.637366720516963e-06,
|
| 286780 |
+
"loss": 0.2977,
|
| 286781 |
+
"step": 147305
|
| 286782 |
+
},
|
| 286783 |
+
{
|
| 286784 |
+
"epoch": 1187.73,
|
| 286785 |
+
"learning_rate": 7.637285945072699e-06,
|
| 286786 |
+
"loss": 0.3194,
|
| 286787 |
+
"step": 147310
|
| 286788 |
+
},
|
| 286789 |
+
{
|
| 286790 |
+
"epoch": 1187.77,
|
| 286791 |
+
"learning_rate": 7.637205169628433e-06,
|
| 286792 |
+
"loss": 0.5155,
|
| 286793 |
+
"step": 147315
|
| 286794 |
+
},
|
| 286795 |
+
{
|
| 286796 |
+
"epoch": 1187.81,
|
| 286797 |
+
"learning_rate": 7.637124394184169e-06,
|
| 286798 |
+
"loss": 1.0884,
|
| 286799 |
+
"step": 147320
|
| 286800 |
+
},
|
| 286801 |
+
{
|
| 286802 |
+
"epoch": 1187.85,
|
| 286803 |
+
"learning_rate": 7.637043618739903e-06,
|
| 286804 |
+
"loss": 0.264,
|
| 286805 |
+
"step": 147325
|
| 286806 |
+
},
|
| 286807 |
+
{
|
| 286808 |
+
"epoch": 1187.89,
|
| 286809 |
+
"learning_rate": 7.636962843295639e-06,
|
| 286810 |
+
"loss": 0.2722,
|
| 286811 |
+
"step": 147330
|
| 286812 |
+
},
|
| 286813 |
+
{
|
| 286814 |
+
"epoch": 1187.93,
|
| 286815 |
+
"learning_rate": 7.636882067851373e-06,
|
| 286816 |
+
"loss": 0.3605,
|
| 286817 |
+
"step": 147335
|
| 286818 |
+
},
|
| 286819 |
+
{
|
| 286820 |
+
"epoch": 1187.97,
|
| 286821 |
+
"learning_rate": 7.636801292407109e-06,
|
| 286822 |
+
"loss": 0.5894,
|
| 286823 |
+
"step": 147340
|
| 286824 |
+
},
|
| 286825 |
+
{
|
| 286826 |
+
"epoch": 1188.0,
|
| 286827 |
+
"eval_loss": 0.4666043221950531,
|
| 286828 |
+
"eval_runtime": 42.1952,
|
| 286829 |
+
"eval_samples_per_second": 19.836,
|
| 286830 |
+
"eval_steps_per_second": 0.64,
|
| 286831 |
+
"eval_wer": 0.18497540617081532,
|
| 286832 |
+
"step": 147343
|
| 286833 |
+
},
|
| 286834 |
+
{
|
| 286835 |
+
"epoch": 1178.02,
|
| 286836 |
+
"learning_rate": 7.636720516962845e-06,
|
| 286837 |
+
"loss": 0.2956,
|
| 286838 |
+
"step": 147345
|
| 286839 |
+
},
|
| 286840 |
+
{
|
| 286841 |
+
"epoch": 1178.06,
|
| 286842 |
+
"learning_rate": 7.636639741518579e-06,
|
| 286843 |
+
"loss": 0.26,
|
| 286844 |
+
"step": 147350
|
| 286845 |
+
},
|
| 286846 |
+
{
|
| 286847 |
+
"epoch": 1178.1,
|
| 286848 |
+
"learning_rate": 7.636558966074315e-06,
|
| 286849 |
+
"loss": 0.302,
|
| 286850 |
+
"step": 147355
|
| 286851 |
+
},
|
| 286852 |
+
{
|
| 286853 |
+
"epoch": 1178.14,
|
| 286854 |
+
"learning_rate": 7.636478190630049e-06,
|
| 286855 |
+
"loss": 0.2915,
|
| 286856 |
+
"step": 147360
|
| 286857 |
+
},
|
| 286858 |
+
{
|
| 286859 |
+
"epoch": 1178.18,
|
| 286860 |
+
"learning_rate": 7.636397415185785e-06,
|
| 286861 |
+
"loss": 0.5731,
|
| 286862 |
+
"step": 147365
|
| 286863 |
+
},
|
| 286864 |
+
{
|
| 286865 |
+
"epoch": 1178.22,
|
| 286866 |
+
"learning_rate": 7.636316639741519e-06,
|
| 286867 |
+
"loss": 1.0072,
|
| 286868 |
+
"step": 147370
|
| 286869 |
+
},
|
| 286870 |
+
{
|
| 286871 |
+
"epoch": 1178.26,
|
| 286872 |
+
"learning_rate": 7.636235864297255e-06,
|
| 286873 |
+
"loss": 0.3019,
|
| 286874 |
+
"step": 147375
|
| 286875 |
+
},
|
| 286876 |
+
{
|
| 286877 |
+
"epoch": 1178.3,
|
| 286878 |
+
"learning_rate": 7.636155088852989e-06,
|
| 286879 |
+
"loss": 0.2681,
|
| 286880 |
+
"step": 147380
|
| 286881 |
+
},
|
| 286882 |
+
{
|
| 286883 |
+
"epoch": 1178.34,
|
| 286884 |
+
"learning_rate": 7.636074313408725e-06,
|
| 286885 |
+
"loss": 0.2926,
|
| 286886 |
+
"step": 147385
|
| 286887 |
+
},
|
| 286888 |
+
{
|
| 286889 |
+
"epoch": 1178.38,
|
| 286890 |
+
"learning_rate": 7.635993537964459e-06,
|
| 286891 |
+
"loss": 0.56,
|
| 286892 |
+
"step": 147390
|
| 286893 |
+
},
|
| 286894 |
+
{
|
| 286895 |
+
"epoch": 1178.42,
|
| 286896 |
+
"learning_rate": 7.635912762520195e-06,
|
| 286897 |
+
"loss": 1.0944,
|
| 286898 |
+
"step": 147395
|
| 286899 |
+
},
|
| 286900 |
+
{
|
| 286901 |
+
"epoch": 1178.46,
|
| 286902 |
+
"learning_rate": 7.63583198707593e-06,
|
| 286903 |
+
"loss": 0.2515,
|
| 286904 |
+
"step": 147400
|
| 286905 |
+
},
|
| 286906 |
+
{
|
| 286907 |
+
"epoch": 1178.5,
|
| 286908 |
+
"learning_rate": 7.635751211631665e-06,
|
| 286909 |
+
"loss": 0.2683,
|
| 286910 |
+
"step": 147405
|
| 286911 |
+
},
|
| 286912 |
+
{
|
| 286913 |
+
"epoch": 1178.54,
|
| 286914 |
+
"learning_rate": 7.6356704361874e-06,
|
| 286915 |
+
"loss": 0.3564,
|
| 286916 |
+
"step": 147410
|
| 286917 |
+
},
|
| 286918 |
+
{
|
| 286919 |
+
"epoch": 1178.58,
|
| 286920 |
+
"learning_rate": 7.635589660743135e-06,
|
| 286921 |
+
"loss": 0.6217,
|
| 286922 |
+
"step": 147415
|
| 286923 |
+
},
|
| 286924 |
+
{
|
| 286925 |
+
"epoch": 1178.62,
|
| 286926 |
+
"learning_rate": 7.63550888529887e-06,
|
| 286927 |
+
"loss": 1.2079,
|
| 286928 |
+
"step": 147420
|
| 286929 |
+
},
|
| 286930 |
+
{
|
| 286931 |
+
"epoch": 1178.66,
|
| 286932 |
+
"learning_rate": 7.635428109854605e-06,
|
| 286933 |
+
"loss": 0.2979,
|
| 286934 |
+
"step": 147425
|
| 286935 |
+
},
|
| 286936 |
+
{
|
| 286937 |
+
"epoch": 1178.7,
|
| 286938 |
+
"learning_rate": 7.63534733441034e-06,
|
| 286939 |
+
"loss": 0.2945,
|
| 286940 |
+
"step": 147430
|
| 286941 |
+
},
|
| 286942 |
+
{
|
| 286943 |
+
"epoch": 1178.74,
|
| 286944 |
+
"learning_rate": 7.635266558966075e-06,
|
| 286945 |
+
"loss": 0.3238,
|
| 286946 |
+
"step": 147435
|
| 286947 |
+
},
|
| 286948 |
+
{
|
| 286949 |
+
"epoch": 1178.78,
|
| 286950 |
+
"learning_rate": 7.63518578352181e-06,
|
| 286951 |
+
"loss": 0.5488,
|
| 286952 |
+
"step": 147440
|
| 286953 |
+
},
|
| 286954 |
+
{
|
| 286955 |
+
"epoch": 1178.82,
|
| 286956 |
+
"learning_rate": 7.635105008077545e-06,
|
| 286957 |
+
"loss": 0.9108,
|
| 286958 |
+
"step": 147445
|
| 286959 |
+
},
|
| 286960 |
+
{
|
| 286961 |
+
"epoch": 1178.86,
|
| 286962 |
+
"learning_rate": 7.63502423263328e-06,
|
| 286963 |
+
"loss": 0.2375,
|
| 286964 |
+
"step": 147450
|
| 286965 |
+
},
|
| 286966 |
+
{
|
| 286967 |
+
"epoch": 1178.9,
|
| 286968 |
+
"learning_rate": 7.634943457189014e-06,
|
| 286969 |
+
"loss": 0.2875,
|
| 286970 |
+
"step": 147455
|
| 286971 |
+
},
|
| 286972 |
+
{
|
| 286973 |
+
"epoch": 1178.94,
|
| 286974 |
+
"learning_rate": 7.63486268174475e-06,
|
| 286975 |
+
"loss": 0.4847,
|
| 286976 |
+
"step": 147460
|
| 286977 |
+
},
|
| 286978 |
+
{
|
| 286979 |
+
"epoch": 1178.98,
|
| 286980 |
+
"learning_rate": 7.634781906300486e-06,
|
| 286981 |
+
"loss": 0.5441,
|
| 286982 |
+
"step": 147465
|
| 286983 |
+
},
|
| 286984 |
+
{
|
| 286985 |
+
"epoch": 1179.0,
|
| 286986 |
+
"eval_loss": 0.3391190469264984,
|
| 286987 |
+
"eval_runtime": 42.1037,
|
| 286988 |
+
"eval_samples_per_second": 19.975,
|
| 286989 |
+
"eval_steps_per_second": 0.641,
|
| 286990 |
+
"eval_wer": 0.16946236559139785,
|
| 286991 |
+
"step": 147468
|
| 286992 |
+
},
|
| 286993 |
+
{
|
| 286994 |
+
"epoch": 1179.02,
|
| 286995 |
+
"learning_rate": 7.63470113085622e-06,
|
| 286996 |
+
"loss": 0.3348,
|
| 286997 |
+
"step": 147470
|
| 286998 |
+
},
|
| 286999 |
+
{
|
| 287000 |
+
"epoch": 1179.06,
|
| 287001 |
+
"learning_rate": 7.634620355411956e-06,
|
| 287002 |
+
"loss": 0.2565,
|
| 287003 |
+
"step": 147475
|
| 287004 |
+
},
|
| 287005 |
+
{
|
| 287006 |
+
"epoch": 1179.1,
|
| 287007 |
+
"learning_rate": 7.63453957996769e-06,
|
| 287008 |
+
"loss": 0.2741,
|
| 287009 |
+
"step": 147480
|
| 287010 |
+
},
|
| 287011 |
+
{
|
| 287012 |
+
"epoch": 1179.14,
|
| 287013 |
+
"learning_rate": 7.634458804523426e-06,
|
| 287014 |
+
"loss": 0.3409,
|
| 287015 |
+
"step": 147485
|
| 287016 |
+
},
|
| 287017 |
+
{
|
| 287018 |
+
"epoch": 1179.18,
|
| 287019 |
+
"learning_rate": 7.63437802907916e-06,
|
| 287020 |
+
"loss": 0.5883,
|
| 287021 |
+
"step": 147490
|
| 287022 |
+
},
|
| 287023 |
+
{
|
| 287024 |
+
"epoch": 1179.22,
|
| 287025 |
+
"learning_rate": 7.634297253634896e-06,
|
| 287026 |
+
"loss": 0.9819,
|
| 287027 |
+
"step": 147495
|
| 287028 |
+
},
|
| 287029 |
+
{
|
| 287030 |
+
"epoch": 1179.26,
|
| 287031 |
+
"learning_rate": 7.63421647819063e-06,
|
| 287032 |
+
"loss": 0.2667,
|
| 287033 |
+
"step": 147500
|
| 287034 |
+
},
|
| 287035 |
+
{
|
| 287036 |
+
"epoch": 1179.3,
|
| 287037 |
+
"learning_rate": 7.634135702746366e-06,
|
| 287038 |
+
"loss": 0.2475,
|
| 287039 |
+
"step": 147505
|
| 287040 |
+
},
|
| 287041 |
+
{
|
| 287042 |
+
"epoch": 1179.34,
|
| 287043 |
+
"learning_rate": 7.6340549273021e-06,
|
| 287044 |
+
"loss": 0.3478,
|
| 287045 |
+
"step": 147510
|
| 287046 |
+
},
|
| 287047 |
+
{
|
| 287048 |
+
"epoch": 1179.38,
|
| 287049 |
+
"learning_rate": 7.633974151857836e-06,
|
| 287050 |
+
"loss": 0.5463,
|
| 287051 |
+
"step": 147515
|
| 287052 |
+
},
|
| 287053 |
+
{
|
| 287054 |
+
"epoch": 1179.42,
|
| 287055 |
+
"learning_rate": 7.633893376413572e-06,
|
| 287056 |
+
"loss": 0.9482,
|
| 287057 |
+
"step": 147520
|
| 287058 |
+
},
|
| 287059 |
+
{
|
| 287060 |
+
"epoch": 1179.46,
|
| 287061 |
+
"learning_rate": 7.633812600969306e-06,
|
| 287062 |
+
"loss": 0.2754,
|
| 287063 |
+
"step": 147525
|
| 287064 |
+
},
|
| 287065 |
+
{
|
| 287066 |
+
"epoch": 1179.5,
|
| 287067 |
+
"learning_rate": 7.633731825525042e-06,
|
| 287068 |
+
"loss": 0.2776,
|
| 287069 |
+
"step": 147530
|
| 287070 |
+
},
|
| 287071 |
+
{
|
| 287072 |
+
"epoch": 1179.54,
|
| 287073 |
+
"learning_rate": 7.633651050080776e-06,
|
| 287074 |
+
"loss": 0.3946,
|
| 287075 |
+
"step": 147535
|
| 287076 |
+
},
|
| 287077 |
+
{
|
| 287078 |
+
"epoch": 1179.58,
|
| 287079 |
+
"learning_rate": 7.633570274636512e-06,
|
| 287080 |
+
"loss": 0.5974,
|
| 287081 |
+
"step": 147540
|
| 287082 |
+
},
|
| 287083 |
+
{
|
| 287084 |
+
"epoch": 1179.62,
|
| 287085 |
+
"learning_rate": 7.633489499192246e-06,
|
| 287086 |
+
"loss": 0.96,
|
| 287087 |
+
"step": 147545
|
| 287088 |
+
},
|
| 287089 |
+
{
|
| 287090 |
+
"epoch": 1179.66,
|
| 287091 |
+
"learning_rate": 7.633408723747982e-06,
|
| 287092 |
+
"loss": 0.2411,
|
| 287093 |
+
"step": 147550
|
| 287094 |
+
},
|
| 287095 |
+
{
|
| 287096 |
+
"epoch": 1179.7,
|
| 287097 |
+
"learning_rate": 7.633327948303716e-06,
|
| 287098 |
+
"loss": 0.2866,
|
| 287099 |
+
"step": 147555
|
| 287100 |
+
},
|
| 287101 |
+
{
|
| 287102 |
+
"epoch": 1179.74,
|
| 287103 |
+
"learning_rate": 7.633247172859452e-06,
|
| 287104 |
+
"loss": 0.2981,
|
| 287105 |
+
"step": 147560
|
| 287106 |
+
},
|
| 287107 |
+
{
|
| 287108 |
+
"epoch": 1179.78,
|
| 287109 |
+
"learning_rate": 7.633166397415186e-06,
|
| 287110 |
+
"loss": 0.5826,
|
| 287111 |
+
"step": 147565
|
| 287112 |
+
},
|
| 287113 |
+
{
|
| 287114 |
+
"epoch": 1179.82,
|
| 287115 |
+
"learning_rate": 7.633085621970922e-06,
|
| 287116 |
+
"loss": 0.9468,
|
| 287117 |
+
"step": 147570
|
| 287118 |
+
},
|
| 287119 |
+
{
|
| 287120 |
+
"epoch": 1179.86,
|
| 287121 |
+
"learning_rate": 7.633004846526658e-06,
|
| 287122 |
+
"loss": 0.272,
|
| 287123 |
+
"step": 147575
|
| 287124 |
+
},
|
| 287125 |
+
{
|
| 287126 |
+
"epoch": 1179.9,
|
| 287127 |
+
"learning_rate": 7.632924071082392e-06,
|
| 287128 |
+
"loss": 0.2624,
|
| 287129 |
+
"step": 147580
|
| 287130 |
+
},
|
| 287131 |
+
{
|
| 287132 |
+
"epoch": 1179.94,
|
| 287133 |
+
"learning_rate": 7.632843295638128e-06,
|
| 287134 |
+
"loss": 0.4524,
|
| 287135 |
+
"step": 147585
|
| 287136 |
+
},
|
| 287137 |
+
{
|
| 287138 |
+
"epoch": 1179.98,
|
| 287139 |
+
"learning_rate": 7.632762520193862e-06,
|
| 287140 |
+
"loss": 0.6987,
|
| 287141 |
+
"step": 147590
|
| 287142 |
+
},
|
| 287143 |
+
{
|
| 287144 |
+
"epoch": 1180.0,
|
| 287145 |
+
"eval_loss": 0.3705912232398987,
|
| 287146 |
+
"eval_runtime": 41.2477,
|
| 287147 |
+
"eval_samples_per_second": 20.365,
|
| 287148 |
+
"eval_steps_per_second": 0.655,
|
| 287149 |
+
"eval_wer": 0.17816803905851492,
|
| 287150 |
+
"step": 147593
|
| 287151 |
+
},
|
| 287152 |
+
{
|
| 287153 |
+
"epoch": 1180.02,
|
| 287154 |
+
"learning_rate": 7.632681744749598e-06,
|
| 287155 |
+
"loss": 0.323,
|
| 287156 |
+
"step": 147595
|
| 287157 |
+
},
|
| 287158 |
+
{
|
| 287159 |
+
"epoch": 1180.06,
|
| 287160 |
+
"learning_rate": 7.632600969305332e-06,
|
| 287161 |
+
"loss": 0.2933,
|
| 287162 |
+
"step": 147600
|
| 287163 |
+
},
|
| 287164 |
+
{
|
| 287165 |
+
"epoch": 1180.1,
|
| 287166 |
+
"learning_rate": 7.632520193861068e-06,
|
| 287167 |
+
"loss": 0.2753,
|
| 287168 |
+
"step": 147605
|
| 287169 |
+
},
|
| 287170 |
+
{
|
| 287171 |
+
"epoch": 1180.14,
|
| 287172 |
+
"learning_rate": 7.632439418416802e-06,
|
| 287173 |
+
"loss": 0.3025,
|
| 287174 |
+
"step": 147610
|
| 287175 |
+
},
|
| 287176 |
+
{
|
| 287177 |
+
"epoch": 1180.18,
|
| 287178 |
+
"learning_rate": 7.632358642972537e-06,
|
| 287179 |
+
"loss": 0.5774,
|
| 287180 |
+
"step": 147615
|
| 287181 |
+
},
|
| 287182 |
+
{
|
| 287183 |
+
"epoch": 1180.22,
|
| 287184 |
+
"learning_rate": 7.632277867528272e-06,
|
| 287185 |
+
"loss": 1.0452,
|
| 287186 |
+
"step": 147620
|
| 287187 |
+
},
|
| 287188 |
+
{
|
| 287189 |
+
"epoch": 1180.26,
|
| 287190 |
+
"learning_rate": 7.632197092084007e-06,
|
| 287191 |
+
"loss": 0.2601,
|
| 287192 |
+
"step": 147625
|
| 287193 |
+
},
|
| 287194 |
+
{
|
| 287195 |
+
"epoch": 1180.3,
|
| 287196 |
+
"learning_rate": 7.632116316639742e-06,
|
| 287197 |
+
"loss": 0.258,
|
| 287198 |
+
"step": 147630
|
| 287199 |
+
},
|
| 287200 |
+
{
|
| 287201 |
+
"epoch": 1180.34,
|
| 287202 |
+
"learning_rate": 7.632035541195477e-06,
|
| 287203 |
+
"loss": 0.3367,
|
| 287204 |
+
"step": 147635
|
| 287205 |
+
},
|
| 287206 |
+
{
|
| 287207 |
+
"epoch": 1180.38,
|
| 287208 |
+
"learning_rate": 7.631954765751213e-06,
|
| 287209 |
+
"loss": 0.6126,
|
| 287210 |
+
"step": 147640
|
| 287211 |
+
},
|
| 287212 |
+
{
|
| 287213 |
+
"epoch": 1180.42,
|
| 287214 |
+
"learning_rate": 7.631873990306947e-06,
|
| 287215 |
+
"loss": 0.85,
|
| 287216 |
+
"step": 147645
|
| 287217 |
+
},
|
| 287218 |
+
{
|
| 287219 |
+
"epoch": 1180.46,
|
| 287220 |
+
"learning_rate": 7.631793214862683e-06,
|
| 287221 |
+
"loss": 0.2583,
|
| 287222 |
+
"step": 147650
|
| 287223 |
+
},
|
| 287224 |
+
{
|
| 287225 |
+
"epoch": 1180.5,
|
| 287226 |
+
"learning_rate": 7.631712439418417e-06,
|
| 287227 |
+
"loss": 0.2821,
|
| 287228 |
+
"step": 147655
|
| 287229 |
+
},
|
| 287230 |
+
{
|
| 287231 |
+
"epoch": 1180.54,
|
| 287232 |
+
"learning_rate": 7.631631663974153e-06,
|
| 287233 |
+
"loss": 0.3228,
|
| 287234 |
+
"step": 147660
|
| 287235 |
+
},
|
| 287236 |
+
{
|
| 287237 |
+
"epoch": 1180.58,
|
| 287238 |
+
"learning_rate": 7.631550888529887e-06,
|
| 287239 |
+
"loss": 0.5341,
|
| 287240 |
+
"step": 147665
|
| 287241 |
+
},
|
| 287242 |
+
{
|
| 287243 |
+
"epoch": 1180.62,
|
| 287244 |
+
"learning_rate": 7.631470113085623e-06,
|
| 287245 |
+
"loss": 1.0162,
|
| 287246 |
+
"step": 147670
|
| 287247 |
+
},
|
| 287248 |
+
{
|
| 287249 |
+
"epoch": 1180.66,
|
| 287250 |
+
"learning_rate": 7.631389337641357e-06,
|
| 287251 |
+
"loss": 0.2755,
|
| 287252 |
+
"step": 147675
|
| 287253 |
+
},
|
| 287254 |
+
{
|
| 287255 |
+
"epoch": 1180.7,
|
| 287256 |
+
"learning_rate": 7.631308562197093e-06,
|
| 287257 |
+
"loss": 0.3447,
|
| 287258 |
+
"step": 147680
|
| 287259 |
+
},
|
| 287260 |
+
{
|
| 287261 |
+
"epoch": 1180.74,
|
| 287262 |
+
"learning_rate": 7.631227786752827e-06,
|
| 287263 |
+
"loss": 0.3052,
|
| 287264 |
+
"step": 147685
|
| 287265 |
+
},
|
| 287266 |
+
{
|
| 287267 |
+
"epoch": 1180.78,
|
| 287268 |
+
"learning_rate": 7.631147011308563e-06,
|
| 287269 |
+
"loss": 0.5751,
|
| 287270 |
+
"step": 147690
|
| 287271 |
+
},
|
| 287272 |
+
{
|
| 287273 |
+
"epoch": 1180.82,
|
| 287274 |
+
"learning_rate": 7.631066235864299e-06,
|
| 287275 |
+
"loss": 1.0567,
|
| 287276 |
+
"step": 147695
|
| 287277 |
+
},
|
| 287278 |
+
{
|
| 287279 |
+
"epoch": 1180.86,
|
| 287280 |
+
"learning_rate": 7.630985460420033e-06,
|
| 287281 |
+
"loss": 0.2482,
|
| 287282 |
+
"step": 147700
|
| 287283 |
+
},
|
| 287284 |
+
{
|
| 287285 |
+
"epoch": 1180.9,
|
| 287286 |
+
"learning_rate": 7.630904684975769e-06,
|
| 287287 |
+
"loss": 0.2403,
|
| 287288 |
+
"step": 147705
|
| 287289 |
+
},
|
| 287290 |
+
{
|
| 287291 |
+
"epoch": 1180.94,
|
| 287292 |
+
"learning_rate": 7.630823909531503e-06,
|
| 287293 |
+
"loss": 0.2849,
|
| 287294 |
+
"step": 147710
|
| 287295 |
+
},
|
| 287296 |
+
{
|
| 287297 |
+
"epoch": 1180.98,
|
| 287298 |
+
"learning_rate": 7.630743134087239e-06,
|
| 287299 |
+
"loss": 0.6648,
|
| 287300 |
+
"step": 147715
|
| 287301 |
+
},
|
| 287302 |
+
{
|
| 287303 |
+
"epoch": 1181.0,
|
| 287304 |
+
"eval_loss": 0.41363415122032166,
|
| 287305 |
+
"eval_runtime": 42.0113,
|
| 287306 |
+
"eval_samples_per_second": 19.995,
|
| 287307 |
+
"eval_steps_per_second": 0.643,
|
| 287308 |
+
"eval_wer": 0.17405040041143194,
|
| 287309 |
+
"step": 147718
|
| 287310 |
}
|
| 287311 |
],
|
| 287312 |
+
"max_steps": 625000,
|
| 287313 |
"num_train_epochs": 5000,
|
| 287314 |
+
"total_flos": 4.156926088776546e+20,
|
| 287315 |
"trial_name": null,
|
| 287316 |
"trial_params": null
|
| 287317 |
}
|
model-bin/finetune/base/{checkpoint-147094 β checkpoint-147718}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1630232727.7092497/events.out.tfevents.1630232727.cc93b136ebf5.1086.129
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:780832ab08f3b77e0bc6bd7c90a53455a938679a58856ae2b0c42142a87c2140
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630233167.6375728/events.out.tfevents.1630233167.cc93b136ebf5.1086.131
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:56e6bc045d9d859c1e8b9371e289a92271072d3af78f044ca2a669a52e07c8c9
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630233719.7330806/events.out.tfevents.1630233719.cc93b136ebf5.1086.133
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5d86a1683ec2628ce5b59cc7af4cf7096857b96e3d77166012379614bf31b0b7
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630234160.490352/events.out.tfevents.1630234160.cc93b136ebf5.1086.135
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f4ff22a0f3f1a0df95a237545ce5fccf1edf9e6069dffb8fdbbb76ed56cd7005
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630234598.790891/events.out.tfevents.1630234598.cc93b136ebf5.1086.137
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1a8cc33a1b209d27fca78683101dfa55c7e9bdc0f1d50ac5de7fbf379eb64ac9
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630232727.cc93b136ebf5.1086.128
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:085aafc7a3bf03ed932af3ae988f600296460344df7d91c39e751f26e21992c2
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630233167.cc93b136ebf5.1086.130
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:47313415493040ff2021285ea26d0f27865a6c0907ab34f41895d964b76b0e9e
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630233719.cc93b136ebf5.1086.132
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7231e23afd5dadce4dd983ae4ebec40da145d93e74177c29d17840501e4af741
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630234160.cc93b136ebf5.1086.134
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b0b68f5666a97346134cf4924bea11adab7e4c0ab3c04c8e87433e96952be69d
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630234598.cc93b136ebf5.1086.136
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7544e462cd34a8da5008c743b913522e2bfa40941e7ea34ac43f9d03c4e4bf7e
|
| 3 |
+
size 8622
|