"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-70187 β checkpoint-70809}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-70187 β checkpoint-70809}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-70187 β checkpoint-70809}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-70187 β checkpoint-70809}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-70187 β checkpoint-70809}/rng_state.pth +2 -2
- model-bin/finetune/base/{checkpoint-70187 β checkpoint-70809}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-70187 β checkpoint-70809}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-70187 β checkpoint-70809}/trainer_state.json +792 -3
- model-bin/finetune/base/{checkpoint-70187 β checkpoint-70809}/training_args.bin +0 -0
- model-bin/finetune/base/log/1629855529.4768627/events.out.tfevents.1629855529.7e498afd5545.905.13 +3 -0
- model-bin/finetune/base/log/1629856180.0808132/events.out.tfevents.1629856180.7e498afd5545.905.15 +3 -0
- model-bin/finetune/base/log/1629856939.1696703/events.out.tfevents.1629856939.7e498afd5545.905.17 +3 -0
- model-bin/finetune/base/log/1629857599.5043178/events.out.tfevents.1629857599.7e498afd5545.905.19 +3 -0
- model-bin/finetune/base/log/1629858276.3230329/events.out.tfevents.1629858276.7e498afd5545.905.21 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629855529.7e498afd5545.905.12 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629856180.7e498afd5545.905.14 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629856939.7e498afd5545.905.16 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629857599.7e498afd5545.905.18 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629858276.7e498afd5545.905.20 +3 -0
model-bin/finetune/base/{checkpoint-70187 β checkpoint-70809}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-70187 β checkpoint-70809}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1238d3e0407c1b5eab603ddadea5d697c591b6ed7191dded1195fc755b757e8e
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-70187 β checkpoint-70809}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-70187 β checkpoint-70809}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:254d4cf0f3862cae7773484aa9cfabaf779e738f449d472541eb9a091e3fb36a
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-70187 β checkpoint-70809}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9c9688cedf4af3de5be5cd9e8d216e7a95ddb6ceceb3ddf456a28ff4bf1c92c9
|
| 3 |
+
size 14503
|
model-bin/finetune/base/{checkpoint-70187 β checkpoint-70809}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:70f0230abe0d9e7f4ba777c6d4ca204523033a231565d6e066189da086fa610e
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-70187 β checkpoint-70809}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e207074215dc1be4bcfdc4da83cf51e498c21731dc990103b72674b1b112cc7a
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-70187 β checkpoint-70809}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.18412114350410416,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-69565",
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -188664,11 +188664,800 @@
|
|
| 188664 |
"eval_steps_per_second": 0.67,
|
| 188665 |
"eval_wer": 0.19290399522957663,
|
| 188666 |
"step": 70187
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 188667 |
}
|
| 188668 |
],
|
| 188669 |
"max_steps": 620000,
|
| 188670 |
"num_train_epochs": 5000,
|
| 188671 |
-
"total_flos": 1.
|
| 188672 |
"trial_name": null,
|
| 188673 |
"trial_params": null
|
| 188674 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.18412114350410416,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-69565",
|
| 4 |
+
"epoch": 570.995983935743,
|
| 5 |
+
"global_step": 70809,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 188664 |
"eval_steps_per_second": 0.67,
|
| 188665 |
"eval_wer": 0.19290399522957663,
|
| 188666 |
"step": 70187
|
| 188667 |
+
},
|
| 188668 |
+
{
|
| 188669 |
+
"epoch": 561.02,
|
| 188670 |
+
"learning_rate": 8.891602564102566e-06,
|
| 188671 |
+
"loss": 0.3236,
|
| 188672 |
+
"step": 70190
|
| 188673 |
+
},
|
| 188674 |
+
{
|
| 188675 |
+
"epoch": 561.06,
|
| 188676 |
+
"learning_rate": 8.891522435897435e-06,
|
| 188677 |
+
"loss": 0.3191,
|
| 188678 |
+
"step": 70195
|
| 188679 |
+
},
|
| 188680 |
+
{
|
| 188681 |
+
"epoch": 561.1,
|
| 188682 |
+
"learning_rate": 8.891442307692308e-06,
|
| 188683 |
+
"loss": 0.3566,
|
| 188684 |
+
"step": 70200
|
| 188685 |
+
},
|
| 188686 |
+
{
|
| 188687 |
+
"epoch": 561.14,
|
| 188688 |
+
"learning_rate": 8.891362179487181e-06,
|
| 188689 |
+
"loss": 0.3839,
|
| 188690 |
+
"step": 70205
|
| 188691 |
+
},
|
| 188692 |
+
{
|
| 188693 |
+
"epoch": 561.18,
|
| 188694 |
+
"learning_rate": 8.891282051282051e-06,
|
| 188695 |
+
"loss": 0.7562,
|
| 188696 |
+
"step": 70210
|
| 188697 |
+
},
|
| 188698 |
+
{
|
| 188699 |
+
"epoch": 561.22,
|
| 188700 |
+
"learning_rate": 8.891201923076924e-06,
|
| 188701 |
+
"loss": 0.8426,
|
| 188702 |
+
"step": 70215
|
| 188703 |
+
},
|
| 188704 |
+
{
|
| 188705 |
+
"epoch": 561.26,
|
| 188706 |
+
"learning_rate": 8.891121794871795e-06,
|
| 188707 |
+
"loss": 0.2785,
|
| 188708 |
+
"step": 70220
|
| 188709 |
+
},
|
| 188710 |
+
{
|
| 188711 |
+
"epoch": 561.3,
|
| 188712 |
+
"learning_rate": 8.891041666666667e-06,
|
| 188713 |
+
"loss": 0.2801,
|
| 188714 |
+
"step": 70225
|
| 188715 |
+
},
|
| 188716 |
+
{
|
| 188717 |
+
"epoch": 561.34,
|
| 188718 |
+
"learning_rate": 8.890961538461538e-06,
|
| 188719 |
+
"loss": 0.442,
|
| 188720 |
+
"step": 70230
|
| 188721 |
+
},
|
| 188722 |
+
{
|
| 188723 |
+
"epoch": 561.38,
|
| 188724 |
+
"learning_rate": 8.890881410256411e-06,
|
| 188725 |
+
"loss": 0.7326,
|
| 188726 |
+
"step": 70235
|
| 188727 |
+
},
|
| 188728 |
+
{
|
| 188729 |
+
"epoch": 561.42,
|
| 188730 |
+
"learning_rate": 8.890801282051283e-06,
|
| 188731 |
+
"loss": 1.0664,
|
| 188732 |
+
"step": 70240
|
| 188733 |
+
},
|
| 188734 |
+
{
|
| 188735 |
+
"epoch": 561.46,
|
| 188736 |
+
"learning_rate": 8.890721153846154e-06,
|
| 188737 |
+
"loss": 0.2819,
|
| 188738 |
+
"step": 70245
|
| 188739 |
+
},
|
| 188740 |
+
{
|
| 188741 |
+
"epoch": 561.5,
|
| 188742 |
+
"learning_rate": 8.890641025641025e-06,
|
| 188743 |
+
"loss": 0.3275,
|
| 188744 |
+
"step": 70250
|
| 188745 |
+
},
|
| 188746 |
+
{
|
| 188747 |
+
"epoch": 561.54,
|
| 188748 |
+
"learning_rate": 8.890560897435898e-06,
|
| 188749 |
+
"loss": 0.4213,
|
| 188750 |
+
"step": 70255
|
| 188751 |
+
},
|
| 188752 |
+
{
|
| 188753 |
+
"epoch": 561.58,
|
| 188754 |
+
"learning_rate": 8.89048076923077e-06,
|
| 188755 |
+
"loss": 0.908,
|
| 188756 |
+
"step": 70260
|
| 188757 |
+
},
|
| 188758 |
+
{
|
| 188759 |
+
"epoch": 561.62,
|
| 188760 |
+
"learning_rate": 8.890400641025641e-06,
|
| 188761 |
+
"loss": 0.8696,
|
| 188762 |
+
"step": 70265
|
| 188763 |
+
},
|
| 188764 |
+
{
|
| 188765 |
+
"epoch": 561.66,
|
| 188766 |
+
"learning_rate": 8.890320512820514e-06,
|
| 188767 |
+
"loss": 0.3694,
|
| 188768 |
+
"step": 70270
|
| 188769 |
+
},
|
| 188770 |
+
{
|
| 188771 |
+
"epoch": 561.7,
|
| 188772 |
+
"learning_rate": 8.890240384615386e-06,
|
| 188773 |
+
"loss": 0.2928,
|
| 188774 |
+
"step": 70275
|
| 188775 |
+
},
|
| 188776 |
+
{
|
| 188777 |
+
"epoch": 561.74,
|
| 188778 |
+
"learning_rate": 8.890160256410257e-06,
|
| 188779 |
+
"loss": 0.3682,
|
| 188780 |
+
"step": 70280
|
| 188781 |
+
},
|
| 188782 |
+
{
|
| 188783 |
+
"epoch": 561.78,
|
| 188784 |
+
"learning_rate": 8.890080128205128e-06,
|
| 188785 |
+
"loss": 0.8639,
|
| 188786 |
+
"step": 70285
|
| 188787 |
+
},
|
| 188788 |
+
{
|
| 188789 |
+
"epoch": 561.82,
|
| 188790 |
+
"learning_rate": 8.890000000000001e-06,
|
| 188791 |
+
"loss": 0.9488,
|
| 188792 |
+
"step": 70290
|
| 188793 |
+
},
|
| 188794 |
+
{
|
| 188795 |
+
"epoch": 561.86,
|
| 188796 |
+
"learning_rate": 8.889919871794873e-06,
|
| 188797 |
+
"loss": 0.3665,
|
| 188798 |
+
"step": 70295
|
| 188799 |
+
},
|
| 188800 |
+
{
|
| 188801 |
+
"epoch": 561.9,
|
| 188802 |
+
"learning_rate": 8.889839743589744e-06,
|
| 188803 |
+
"loss": 0.3282,
|
| 188804 |
+
"step": 70300
|
| 188805 |
+
},
|
| 188806 |
+
{
|
| 188807 |
+
"epoch": 561.94,
|
| 188808 |
+
"learning_rate": 8.889759615384617e-06,
|
| 188809 |
+
"loss": 0.4235,
|
| 188810 |
+
"step": 70305
|
| 188811 |
+
},
|
| 188812 |
+
{
|
| 188813 |
+
"epoch": 561.98,
|
| 188814 |
+
"learning_rate": 8.889679487179488e-06,
|
| 188815 |
+
"loss": 0.8138,
|
| 188816 |
+
"step": 70310
|
| 188817 |
+
},
|
| 188818 |
+
{
|
| 188819 |
+
"epoch": 562.0,
|
| 188820 |
+
"eval_loss": 0.3786197006702423,
|
| 188821 |
+
"eval_runtime": 39.7295,
|
| 188822 |
+
"eval_samples_per_second": 21.118,
|
| 188823 |
+
"eval_steps_per_second": 0.68,
|
| 188824 |
+
"eval_wer": 0.197922893006575,
|
| 188825 |
+
"step": 70312
|
| 188826 |
+
},
|
| 188827 |
+
{
|
| 188828 |
+
"epoch": 562.02,
|
| 188829 |
+
"learning_rate": 8.88959935897436e-06,
|
| 188830 |
+
"loss": 0.3524,
|
| 188831 |
+
"step": 70315
|
| 188832 |
+
},
|
| 188833 |
+
{
|
| 188834 |
+
"epoch": 562.06,
|
| 188835 |
+
"learning_rate": 8.889519230769231e-06,
|
| 188836 |
+
"loss": 0.3185,
|
| 188837 |
+
"step": 70320
|
| 188838 |
+
},
|
| 188839 |
+
{
|
| 188840 |
+
"epoch": 562.1,
|
| 188841 |
+
"learning_rate": 8.889439102564104e-06,
|
| 188842 |
+
"loss": 0.3284,
|
| 188843 |
+
"step": 70325
|
| 188844 |
+
},
|
| 188845 |
+
{
|
| 188846 |
+
"epoch": 562.14,
|
| 188847 |
+
"learning_rate": 8.889358974358976e-06,
|
| 188848 |
+
"loss": 0.388,
|
| 188849 |
+
"step": 70330
|
| 188850 |
+
},
|
| 188851 |
+
{
|
| 188852 |
+
"epoch": 562.18,
|
| 188853 |
+
"learning_rate": 8.889278846153847e-06,
|
| 188854 |
+
"loss": 0.8204,
|
| 188855 |
+
"step": 70335
|
| 188856 |
+
},
|
| 188857 |
+
{
|
| 188858 |
+
"epoch": 562.22,
|
| 188859 |
+
"learning_rate": 8.889198717948718e-06,
|
| 188860 |
+
"loss": 0.8117,
|
| 188861 |
+
"step": 70340
|
| 188862 |
+
},
|
| 188863 |
+
{
|
| 188864 |
+
"epoch": 562.26,
|
| 188865 |
+
"learning_rate": 8.889118589743591e-06,
|
| 188866 |
+
"loss": 0.3556,
|
| 188867 |
+
"step": 70345
|
| 188868 |
+
},
|
| 188869 |
+
{
|
| 188870 |
+
"epoch": 562.3,
|
| 188871 |
+
"learning_rate": 8.889038461538461e-06,
|
| 188872 |
+
"loss": 0.3245,
|
| 188873 |
+
"step": 70350
|
| 188874 |
+
},
|
| 188875 |
+
{
|
| 188876 |
+
"epoch": 562.34,
|
| 188877 |
+
"learning_rate": 8.888958333333334e-06,
|
| 188878 |
+
"loss": 0.3523,
|
| 188879 |
+
"step": 70355
|
| 188880 |
+
},
|
| 188881 |
+
{
|
| 188882 |
+
"epoch": 562.38,
|
| 188883 |
+
"learning_rate": 8.888878205128207e-06,
|
| 188884 |
+
"loss": 0.7863,
|
| 188885 |
+
"step": 70360
|
| 188886 |
+
},
|
| 188887 |
+
{
|
| 188888 |
+
"epoch": 562.42,
|
| 188889 |
+
"learning_rate": 8.888798076923077e-06,
|
| 188890 |
+
"loss": 0.8844,
|
| 188891 |
+
"step": 70365
|
| 188892 |
+
},
|
| 188893 |
+
{
|
| 188894 |
+
"epoch": 562.46,
|
| 188895 |
+
"learning_rate": 8.88871794871795e-06,
|
| 188896 |
+
"loss": 0.362,
|
| 188897 |
+
"step": 70370
|
| 188898 |
+
},
|
| 188899 |
+
{
|
| 188900 |
+
"epoch": 562.5,
|
| 188901 |
+
"learning_rate": 8.888637820512821e-06,
|
| 188902 |
+
"loss": 0.3011,
|
| 188903 |
+
"step": 70375
|
| 188904 |
+
},
|
| 188905 |
+
{
|
| 188906 |
+
"epoch": 562.54,
|
| 188907 |
+
"learning_rate": 8.888557692307693e-06,
|
| 188908 |
+
"loss": 0.37,
|
| 188909 |
+
"step": 70380
|
| 188910 |
+
},
|
| 188911 |
+
{
|
| 188912 |
+
"epoch": 562.58,
|
| 188913 |
+
"learning_rate": 8.888477564102564e-06,
|
| 188914 |
+
"loss": 0.6502,
|
| 188915 |
+
"step": 70385
|
| 188916 |
+
},
|
| 188917 |
+
{
|
| 188918 |
+
"epoch": 562.62,
|
| 188919 |
+
"learning_rate": 8.888397435897437e-06,
|
| 188920 |
+
"loss": 0.8631,
|
| 188921 |
+
"step": 70390
|
| 188922 |
+
},
|
| 188923 |
+
{
|
| 188924 |
+
"epoch": 562.66,
|
| 188925 |
+
"learning_rate": 8.888317307692308e-06,
|
| 188926 |
+
"loss": 0.4869,
|
| 188927 |
+
"step": 70395
|
| 188928 |
+
},
|
| 188929 |
+
{
|
| 188930 |
+
"epoch": 562.7,
|
| 188931 |
+
"learning_rate": 8.88823717948718e-06,
|
| 188932 |
+
"loss": 0.3648,
|
| 188933 |
+
"step": 70400
|
| 188934 |
+
},
|
| 188935 |
+
{
|
| 188936 |
+
"epoch": 562.74,
|
| 188937 |
+
"learning_rate": 8.888157051282053e-06,
|
| 188938 |
+
"loss": 0.3954,
|
| 188939 |
+
"step": 70405
|
| 188940 |
+
},
|
| 188941 |
+
{
|
| 188942 |
+
"epoch": 562.78,
|
| 188943 |
+
"learning_rate": 8.888076923076924e-06,
|
| 188944 |
+
"loss": 0.7799,
|
| 188945 |
+
"step": 70410
|
| 188946 |
+
},
|
| 188947 |
+
{
|
| 188948 |
+
"epoch": 562.82,
|
| 188949 |
+
"learning_rate": 8.887996794871795e-06,
|
| 188950 |
+
"loss": 0.7664,
|
| 188951 |
+
"step": 70415
|
| 188952 |
+
},
|
| 188953 |
+
{
|
| 188954 |
+
"epoch": 562.86,
|
| 188955 |
+
"learning_rate": 8.887916666666667e-06,
|
| 188956 |
+
"loss": 0.316,
|
| 188957 |
+
"step": 70420
|
| 188958 |
+
},
|
| 188959 |
+
{
|
| 188960 |
+
"epoch": 562.9,
|
| 188961 |
+
"learning_rate": 8.88783653846154e-06,
|
| 188962 |
+
"loss": 0.3008,
|
| 188963 |
+
"step": 70425
|
| 188964 |
+
},
|
| 188965 |
+
{
|
| 188966 |
+
"epoch": 562.94,
|
| 188967 |
+
"learning_rate": 8.887756410256411e-06,
|
| 188968 |
+
"loss": 0.4025,
|
| 188969 |
+
"step": 70430
|
| 188970 |
+
},
|
| 188971 |
+
{
|
| 188972 |
+
"epoch": 562.98,
|
| 188973 |
+
"learning_rate": 8.887676282051283e-06,
|
| 188974 |
+
"loss": 0.7471,
|
| 188975 |
+
"step": 70435
|
| 188976 |
+
},
|
| 188977 |
+
{
|
| 188978 |
+
"epoch": 563.0,
|
| 188979 |
+
"eval_loss": 0.46410617232322693,
|
| 188980 |
+
"eval_runtime": 42.712,
|
| 188981 |
+
"eval_samples_per_second": 19.643,
|
| 188982 |
+
"eval_steps_per_second": 0.632,
|
| 188983 |
+
"eval_wer": 0.19668026042784575,
|
| 188984 |
+
"step": 70437
|
| 188985 |
+
},
|
| 188986 |
+
{
|
| 188987 |
+
"epoch": 568.02,
|
| 188988 |
+
"learning_rate": 8.887596153846154e-06,
|
| 188989 |
+
"loss": 0.3198,
|
| 188990 |
+
"step": 70440
|
| 188991 |
+
},
|
| 188992 |
+
{
|
| 188993 |
+
"epoch": 568.06,
|
| 188994 |
+
"learning_rate": 8.887516025641027e-06,
|
| 188995 |
+
"loss": 0.3777,
|
| 188996 |
+
"step": 70445
|
| 188997 |
+
},
|
| 188998 |
+
{
|
| 188999 |
+
"epoch": 568.1,
|
| 189000 |
+
"learning_rate": 8.887435897435898e-06,
|
| 189001 |
+
"loss": 0.365,
|
| 189002 |
+
"step": 70450
|
| 189003 |
+
},
|
| 189004 |
+
{
|
| 189005 |
+
"epoch": 568.14,
|
| 189006 |
+
"learning_rate": 8.88735576923077e-06,
|
| 189007 |
+
"loss": 0.3966,
|
| 189008 |
+
"step": 70455
|
| 189009 |
+
},
|
| 189010 |
+
{
|
| 189011 |
+
"epoch": 568.18,
|
| 189012 |
+
"learning_rate": 8.887275641025643e-06,
|
| 189013 |
+
"loss": 0.8013,
|
| 189014 |
+
"step": 70460
|
| 189015 |
+
},
|
| 189016 |
+
{
|
| 189017 |
+
"epoch": 568.22,
|
| 189018 |
+
"learning_rate": 8.887195512820514e-06,
|
| 189019 |
+
"loss": 0.8617,
|
| 189020 |
+
"step": 70465
|
| 189021 |
+
},
|
| 189022 |
+
{
|
| 189023 |
+
"epoch": 568.27,
|
| 189024 |
+
"learning_rate": 8.887115384615385e-06,
|
| 189025 |
+
"loss": 0.3124,
|
| 189026 |
+
"step": 70470
|
| 189027 |
+
},
|
| 189028 |
+
{
|
| 189029 |
+
"epoch": 568.31,
|
| 189030 |
+
"learning_rate": 8.887035256410257e-06,
|
| 189031 |
+
"loss": 0.3463,
|
| 189032 |
+
"step": 70475
|
| 189033 |
+
},
|
| 189034 |
+
{
|
| 189035 |
+
"epoch": 568.35,
|
| 189036 |
+
"learning_rate": 8.88695512820513e-06,
|
| 189037 |
+
"loss": 0.3506,
|
| 189038 |
+
"step": 70480
|
| 189039 |
+
},
|
| 189040 |
+
{
|
| 189041 |
+
"epoch": 568.39,
|
| 189042 |
+
"learning_rate": 8.886875000000001e-06,
|
| 189043 |
+
"loss": 0.8531,
|
| 189044 |
+
"step": 70485
|
| 189045 |
+
},
|
| 189046 |
+
{
|
| 189047 |
+
"epoch": 568.43,
|
| 189048 |
+
"learning_rate": 8.886794871794873e-06,
|
| 189049 |
+
"loss": 0.9379,
|
| 189050 |
+
"step": 70490
|
| 189051 |
+
},
|
| 189052 |
+
{
|
| 189053 |
+
"epoch": 568.47,
|
| 189054 |
+
"learning_rate": 8.886714743589744e-06,
|
| 189055 |
+
"loss": 0.3052,
|
| 189056 |
+
"step": 70495
|
| 189057 |
+
},
|
| 189058 |
+
{
|
| 189059 |
+
"epoch": 568.51,
|
| 189060 |
+
"learning_rate": 8.886634615384617e-06,
|
| 189061 |
+
"loss": 0.2875,
|
| 189062 |
+
"step": 70500
|
| 189063 |
+
},
|
| 189064 |
+
{
|
| 189065 |
+
"epoch": 568.55,
|
| 189066 |
+
"learning_rate": 8.886554487179488e-06,
|
| 189067 |
+
"loss": 0.3442,
|
| 189068 |
+
"step": 70505
|
| 189069 |
+
},
|
| 189070 |
+
{
|
| 189071 |
+
"epoch": 568.59,
|
| 189072 |
+
"learning_rate": 8.88647435897436e-06,
|
| 189073 |
+
"loss": 0.7231,
|
| 189074 |
+
"step": 70510
|
| 189075 |
+
},
|
| 189076 |
+
{
|
| 189077 |
+
"epoch": 568.63,
|
| 189078 |
+
"learning_rate": 8.886394230769233e-06,
|
| 189079 |
+
"loss": 0.8665,
|
| 189080 |
+
"step": 70515
|
| 189081 |
+
},
|
| 189082 |
+
{
|
| 189083 |
+
"epoch": 568.67,
|
| 189084 |
+
"learning_rate": 8.886314102564102e-06,
|
| 189085 |
+
"loss": 0.3137,
|
| 189086 |
+
"step": 70520
|
| 189087 |
+
},
|
| 189088 |
+
{
|
| 189089 |
+
"epoch": 568.71,
|
| 189090 |
+
"learning_rate": 8.886233974358975e-06,
|
| 189091 |
+
"loss": 0.3238,
|
| 189092 |
+
"step": 70525
|
| 189093 |
+
},
|
| 189094 |
+
{
|
| 189095 |
+
"epoch": 568.75,
|
| 189096 |
+
"learning_rate": 8.886153846153847e-06,
|
| 189097 |
+
"loss": 0.3674,
|
| 189098 |
+
"step": 70530
|
| 189099 |
+
},
|
| 189100 |
+
{
|
| 189101 |
+
"epoch": 568.79,
|
| 189102 |
+
"learning_rate": 8.886073717948718e-06,
|
| 189103 |
+
"loss": 0.7046,
|
| 189104 |
+
"step": 70535
|
| 189105 |
+
},
|
| 189106 |
+
{
|
| 189107 |
+
"epoch": 568.83,
|
| 189108 |
+
"learning_rate": 8.88599358974359e-06,
|
| 189109 |
+
"loss": 1.0743,
|
| 189110 |
+
"step": 70540
|
| 189111 |
+
},
|
| 189112 |
+
{
|
| 189113 |
+
"epoch": 568.87,
|
| 189114 |
+
"learning_rate": 8.885913461538463e-06,
|
| 189115 |
+
"loss": 0.2614,
|
| 189116 |
+
"step": 70545
|
| 189117 |
+
},
|
| 189118 |
+
{
|
| 189119 |
+
"epoch": 568.91,
|
| 189120 |
+
"learning_rate": 8.885833333333334e-06,
|
| 189121 |
+
"loss": 0.4457,
|
| 189122 |
+
"step": 70550
|
| 189123 |
+
},
|
| 189124 |
+
{
|
| 189125 |
+
"epoch": 568.95,
|
| 189126 |
+
"learning_rate": 8.885753205128205e-06,
|
| 189127 |
+
"loss": 0.399,
|
| 189128 |
+
"step": 70555
|
| 189129 |
+
},
|
| 189130 |
+
{
|
| 189131 |
+
"epoch": 568.99,
|
| 189132 |
+
"learning_rate": 8.885673076923078e-06,
|
| 189133 |
+
"loss": 0.9402,
|
| 189134 |
+
"step": 70560
|
| 189135 |
+
},
|
| 189136 |
+
{
|
| 189137 |
+
"epoch": 569.0,
|
| 189138 |
+
"eval_loss": 0.4022397994995117,
|
| 189139 |
+
"eval_runtime": 37.522,
|
| 189140 |
+
"eval_samples_per_second": 22.387,
|
| 189141 |
+
"eval_steps_per_second": 0.72,
|
| 189142 |
+
"eval_wer": 0.1899042004421518,
|
| 189143 |
+
"step": 70561
|
| 189144 |
+
},
|
| 189145 |
+
{
|
| 189146 |
+
"epoch": 569.03,
|
| 189147 |
+
"learning_rate": 8.88559294871795e-06,
|
| 189148 |
+
"loss": 0.4009,
|
| 189149 |
+
"step": 70565
|
| 189150 |
+
},
|
| 189151 |
+
{
|
| 189152 |
+
"epoch": 569.07,
|
| 189153 |
+
"learning_rate": 8.885512820512821e-06,
|
| 189154 |
+
"loss": 0.3283,
|
| 189155 |
+
"step": 70570
|
| 189156 |
+
},
|
| 189157 |
+
{
|
| 189158 |
+
"epoch": 569.11,
|
| 189159 |
+
"learning_rate": 8.885432692307692e-06,
|
| 189160 |
+
"loss": 0.3124,
|
| 189161 |
+
"step": 70575
|
| 189162 |
+
},
|
| 189163 |
+
{
|
| 189164 |
+
"epoch": 569.15,
|
| 189165 |
+
"learning_rate": 8.885352564102565e-06,
|
| 189166 |
+
"loss": 0.4535,
|
| 189167 |
+
"step": 70580
|
| 189168 |
+
},
|
| 189169 |
+
{
|
| 189170 |
+
"epoch": 569.19,
|
| 189171 |
+
"learning_rate": 8.885272435897437e-06,
|
| 189172 |
+
"loss": 0.9139,
|
| 189173 |
+
"step": 70585
|
| 189174 |
+
},
|
| 189175 |
+
{
|
| 189176 |
+
"epoch": 569.23,
|
| 189177 |
+
"learning_rate": 8.885192307692308e-06,
|
| 189178 |
+
"loss": 0.6242,
|
| 189179 |
+
"step": 70590
|
| 189180 |
+
},
|
| 189181 |
+
{
|
| 189182 |
+
"epoch": 569.27,
|
| 189183 |
+
"learning_rate": 8.88511217948718e-06,
|
| 189184 |
+
"loss": 0.3918,
|
| 189185 |
+
"step": 70595
|
| 189186 |
+
},
|
| 189187 |
+
{
|
| 189188 |
+
"epoch": 569.31,
|
| 189189 |
+
"learning_rate": 8.885032051282053e-06,
|
| 189190 |
+
"loss": 0.3409,
|
| 189191 |
+
"step": 70600
|
| 189192 |
+
},
|
| 189193 |
+
{
|
| 189194 |
+
"epoch": 569.35,
|
| 189195 |
+
"learning_rate": 8.884951923076924e-06,
|
| 189196 |
+
"loss": 0.427,
|
| 189197 |
+
"step": 70605
|
| 189198 |
+
},
|
| 189199 |
+
{
|
| 189200 |
+
"epoch": 569.39,
|
| 189201 |
+
"learning_rate": 8.884871794871795e-06,
|
| 189202 |
+
"loss": 0.9461,
|
| 189203 |
+
"step": 70610
|
| 189204 |
+
},
|
| 189205 |
+
{
|
| 189206 |
+
"epoch": 569.43,
|
| 189207 |
+
"learning_rate": 8.884791666666668e-06,
|
| 189208 |
+
"loss": 0.6715,
|
| 189209 |
+
"step": 70615
|
| 189210 |
+
},
|
| 189211 |
+
{
|
| 189212 |
+
"epoch": 569.47,
|
| 189213 |
+
"learning_rate": 8.88471153846154e-06,
|
| 189214 |
+
"loss": 0.3049,
|
| 189215 |
+
"step": 70620
|
| 189216 |
+
},
|
| 189217 |
+
{
|
| 189218 |
+
"epoch": 569.51,
|
| 189219 |
+
"learning_rate": 8.884631410256411e-06,
|
| 189220 |
+
"loss": 0.3306,
|
| 189221 |
+
"step": 70625
|
| 189222 |
+
},
|
| 189223 |
+
{
|
| 189224 |
+
"epoch": 569.55,
|
| 189225 |
+
"learning_rate": 8.884551282051282e-06,
|
| 189226 |
+
"loss": 0.3854,
|
| 189227 |
+
"step": 70630
|
| 189228 |
+
},
|
| 189229 |
+
{
|
| 189230 |
+
"epoch": 569.59,
|
| 189231 |
+
"learning_rate": 8.884471153846156e-06,
|
| 189232 |
+
"loss": 0.9237,
|
| 189233 |
+
"step": 70635
|
| 189234 |
+
},
|
| 189235 |
+
{
|
| 189236 |
+
"epoch": 569.63,
|
| 189237 |
+
"learning_rate": 8.884391025641025e-06,
|
| 189238 |
+
"loss": 0.6815,
|
| 189239 |
+
"step": 70640
|
| 189240 |
+
},
|
| 189241 |
+
{
|
| 189242 |
+
"epoch": 569.67,
|
| 189243 |
+
"learning_rate": 8.884310897435898e-06,
|
| 189244 |
+
"loss": 0.341,
|
| 189245 |
+
"step": 70645
|
| 189246 |
+
},
|
| 189247 |
+
{
|
| 189248 |
+
"epoch": 569.71,
|
| 189249 |
+
"learning_rate": 8.88423076923077e-06,
|
| 189250 |
+
"loss": 0.301,
|
| 189251 |
+
"step": 70650
|
| 189252 |
+
},
|
| 189253 |
+
{
|
| 189254 |
+
"epoch": 569.76,
|
| 189255 |
+
"learning_rate": 8.884150641025641e-06,
|
| 189256 |
+
"loss": 0.4376,
|
| 189257 |
+
"step": 70655
|
| 189258 |
+
},
|
| 189259 |
+
{
|
| 189260 |
+
"epoch": 569.8,
|
| 189261 |
+
"learning_rate": 8.884070512820514e-06,
|
| 189262 |
+
"loss": 0.9414,
|
| 189263 |
+
"step": 70660
|
| 189264 |
+
},
|
| 189265 |
+
{
|
| 189266 |
+
"epoch": 569.84,
|
| 189267 |
+
"learning_rate": 8.883990384615385e-06,
|
| 189268 |
+
"loss": 0.7538,
|
| 189269 |
+
"step": 70665
|
| 189270 |
+
},
|
| 189271 |
+
{
|
| 189272 |
+
"epoch": 569.88,
|
| 189273 |
+
"learning_rate": 8.883910256410257e-06,
|
| 189274 |
+
"loss": 0.2806,
|
| 189275 |
+
"step": 70670
|
| 189276 |
+
},
|
| 189277 |
+
{
|
| 189278 |
+
"epoch": 569.92,
|
| 189279 |
+
"learning_rate": 8.883830128205128e-06,
|
| 189280 |
+
"loss": 0.3797,
|
| 189281 |
+
"step": 70675
|
| 189282 |
+
},
|
| 189283 |
+
{
|
| 189284 |
+
"epoch": 569.96,
|
| 189285 |
+
"learning_rate": 8.883750000000001e-06,
|
| 189286 |
+
"loss": 0.4361,
|
| 189287 |
+
"step": 70680
|
| 189288 |
+
},
|
| 189289 |
+
{
|
| 189290 |
+
"epoch": 570.0,
|
| 189291 |
+
"learning_rate": 8.883669871794872e-06,
|
| 189292 |
+
"loss": 1.1359,
|
| 189293 |
+
"step": 70685
|
| 189294 |
+
},
|
| 189295 |
+
{
|
| 189296 |
+
"epoch": 570.0,
|
| 189297 |
+
"eval_loss": 0.4082900285720825,
|
| 189298 |
+
"eval_runtime": 39.7838,
|
| 189299 |
+
"eval_samples_per_second": 21.114,
|
| 189300 |
+
"eval_steps_per_second": 0.679,
|
| 189301 |
+
"eval_wer": 0.19334857311407713,
|
| 189302 |
+
"step": 70685
|
| 189303 |
+
},
|
| 189304 |
+
{
|
| 189305 |
+
"epoch": 570.04,
|
| 189306 |
+
"learning_rate": 8.883589743589744e-06,
|
| 189307 |
+
"loss": 0.3446,
|
| 189308 |
+
"step": 70690
|
| 189309 |
+
},
|
| 189310 |
+
{
|
| 189311 |
+
"epoch": 570.08,
|
| 189312 |
+
"learning_rate": 8.883509615384615e-06,
|
| 189313 |
+
"loss": 0.3753,
|
| 189314 |
+
"step": 70695
|
| 189315 |
+
},
|
| 189316 |
+
{
|
| 189317 |
+
"epoch": 570.12,
|
| 189318 |
+
"learning_rate": 8.883429487179488e-06,
|
| 189319 |
+
"loss": 0.304,
|
| 189320 |
+
"step": 70700
|
| 189321 |
+
},
|
| 189322 |
+
{
|
| 189323 |
+
"epoch": 570.16,
|
| 189324 |
+
"learning_rate": 8.88334935897436e-06,
|
| 189325 |
+
"loss": 0.4448,
|
| 189326 |
+
"step": 70705
|
| 189327 |
+
},
|
| 189328 |
+
{
|
| 189329 |
+
"epoch": 570.2,
|
| 189330 |
+
"learning_rate": 8.883269230769231e-06,
|
| 189331 |
+
"loss": 1.4043,
|
| 189332 |
+
"step": 70710
|
| 189333 |
+
},
|
| 189334 |
+
{
|
| 189335 |
+
"epoch": 570.24,
|
| 189336 |
+
"learning_rate": 8.883189102564104e-06,
|
| 189337 |
+
"loss": 0.3952,
|
| 189338 |
+
"step": 70715
|
| 189339 |
+
},
|
| 189340 |
+
{
|
| 189341 |
+
"epoch": 570.28,
|
| 189342 |
+
"learning_rate": 8.883108974358975e-06,
|
| 189343 |
+
"loss": 0.3139,
|
| 189344 |
+
"step": 70720
|
| 189345 |
+
},
|
| 189346 |
+
{
|
| 189347 |
+
"epoch": 570.32,
|
| 189348 |
+
"learning_rate": 8.883028846153847e-06,
|
| 189349 |
+
"loss": 0.3671,
|
| 189350 |
+
"step": 70725
|
| 189351 |
+
},
|
| 189352 |
+
{
|
| 189353 |
+
"epoch": 570.36,
|
| 189354 |
+
"learning_rate": 8.882948717948718e-06,
|
| 189355 |
+
"loss": 0.4336,
|
| 189356 |
+
"step": 70730
|
| 189357 |
+
},
|
| 189358 |
+
{
|
| 189359 |
+
"epoch": 570.4,
|
| 189360 |
+
"learning_rate": 8.882868589743591e-06,
|
| 189361 |
+
"loss": 1.2766,
|
| 189362 |
+
"step": 70735
|
| 189363 |
+
},
|
| 189364 |
+
{
|
| 189365 |
+
"epoch": 570.44,
|
| 189366 |
+
"learning_rate": 8.882788461538463e-06,
|
| 189367 |
+
"loss": 0.4221,
|
| 189368 |
+
"step": 70740
|
| 189369 |
+
},
|
| 189370 |
+
{
|
| 189371 |
+
"epoch": 570.48,
|
| 189372 |
+
"learning_rate": 8.882708333333334e-06,
|
| 189373 |
+
"loss": 0.3531,
|
| 189374 |
+
"step": 70745
|
| 189375 |
+
},
|
| 189376 |
+
{
|
| 189377 |
+
"epoch": 570.52,
|
| 189378 |
+
"learning_rate": 8.882628205128205e-06,
|
| 189379 |
+
"loss": 0.3561,
|
| 189380 |
+
"step": 70750
|
| 189381 |
+
},
|
| 189382 |
+
{
|
| 189383 |
+
"epoch": 570.56,
|
| 189384 |
+
"learning_rate": 8.882548076923078e-06,
|
| 189385 |
+
"loss": 0.4438,
|
| 189386 |
+
"step": 70755
|
| 189387 |
+
},
|
| 189388 |
+
{
|
| 189389 |
+
"epoch": 570.6,
|
| 189390 |
+
"learning_rate": 8.88246794871795e-06,
|
| 189391 |
+
"loss": 1.1606,
|
| 189392 |
+
"step": 70760
|
| 189393 |
+
},
|
| 189394 |
+
{
|
| 189395 |
+
"epoch": 570.64,
|
| 189396 |
+
"learning_rate": 8.882387820512821e-06,
|
| 189397 |
+
"loss": 0.3717,
|
| 189398 |
+
"step": 70765
|
| 189399 |
+
},
|
| 189400 |
+
{
|
| 189401 |
+
"epoch": 570.68,
|
| 189402 |
+
"learning_rate": 8.882307692307694e-06,
|
| 189403 |
+
"loss": 0.3317,
|
| 189404 |
+
"step": 70770
|
| 189405 |
+
},
|
| 189406 |
+
{
|
| 189407 |
+
"epoch": 570.72,
|
| 189408 |
+
"learning_rate": 8.882227564102565e-06,
|
| 189409 |
+
"loss": 0.4233,
|
| 189410 |
+
"step": 70775
|
| 189411 |
+
},
|
| 189412 |
+
{
|
| 189413 |
+
"epoch": 570.76,
|
| 189414 |
+
"learning_rate": 8.882147435897437e-06,
|
| 189415 |
+
"loss": 0.4874,
|
| 189416 |
+
"step": 70780
|
| 189417 |
+
},
|
| 189418 |
+
{
|
| 189419 |
+
"epoch": 570.8,
|
| 189420 |
+
"learning_rate": 8.882067307692308e-06,
|
| 189421 |
+
"loss": 1.2514,
|
| 189422 |
+
"step": 70785
|
| 189423 |
+
},
|
| 189424 |
+
{
|
| 189425 |
+
"epoch": 570.84,
|
| 189426 |
+
"learning_rate": 8.881987179487181e-06,
|
| 189427 |
+
"loss": 0.308,
|
| 189428 |
+
"step": 70790
|
| 189429 |
+
},
|
| 189430 |
+
{
|
| 189431 |
+
"epoch": 570.88,
|
| 189432 |
+
"learning_rate": 8.88190705128205e-06,
|
| 189433 |
+
"loss": 0.2971,
|
| 189434 |
+
"step": 70795
|
| 189435 |
+
},
|
| 189436 |
+
{
|
| 189437 |
+
"epoch": 570.92,
|
| 189438 |
+
"learning_rate": 8.881826923076924e-06,
|
| 189439 |
+
"loss": 0.3294,
|
| 189440 |
+
"step": 70800
|
| 189441 |
+
},
|
| 189442 |
+
{
|
| 189443 |
+
"epoch": 570.96,
|
| 189444 |
+
"learning_rate": 8.881746794871797e-06,
|
| 189445 |
+
"loss": 0.5041,
|
| 189446 |
+
"step": 70805
|
| 189447 |
+
},
|
| 189448 |
+
{
|
| 189449 |
+
"epoch": 571.0,
|
| 189450 |
+
"eval_loss": 0.3579244613647461,
|
| 189451 |
+
"eval_runtime": 40.1364,
|
| 189452 |
+
"eval_samples_per_second": 20.929,
|
| 189453 |
+
"eval_steps_per_second": 0.673,
|
| 189454 |
+
"eval_wer": 0.18833597008485545,
|
| 189455 |
+
"step": 70809
|
| 189456 |
}
|
| 189457 |
],
|
| 189458 |
"max_steps": 620000,
|
| 189459 |
"num_train_epochs": 5000,
|
| 189460 |
+
"total_flos": 1.992660946951832e+20,
|
| 189461 |
"trial_name": null,
|
| 189462 |
"trial_params": null
|
| 189463 |
}
|
model-bin/finetune/base/{checkpoint-70187 β checkpoint-70809}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1629855529.4768627/events.out.tfevents.1629855529.7e498afd5545.905.13
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6a691631c88f84295fd9ab41c5956717343a6f9dec943932d6a3ad8f46030e3b
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629856180.0808132/events.out.tfevents.1629856180.7e498afd5545.905.15
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4d7a9169e28b1bce5e031d1aa0d74d07354c7bd91a9be4845b40d52cb4bbdf1d
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629856939.1696703/events.out.tfevents.1629856939.7e498afd5545.905.17
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8adcac9c4c255d936902703e352e0e045101ded6951f4111567f54ce342c2a35
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629857599.5043178/events.out.tfevents.1629857599.7e498afd5545.905.19
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:145fddd07a3f4491f7a746dace0ef93d228c1fec0640cb33769649f55c299987
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629858276.3230329/events.out.tfevents.1629858276.7e498afd5545.905.21
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ef04fe4fbfefdcc0b6de0a1a16d7a2ca942190d8f97b43d3d147ad41760d4c44
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1629855529.7e498afd5545.905.12
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:34507e83b75bf4bbc54a8bd105bd143753b51cf27592bb1d69b985c9f4189e64
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629856180.7e498afd5545.905.14
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c92cf6f8487beed64bbf54ef7187b87e4064683edf6121913f4cbb80ea2744c7
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629856939.7e498afd5545.905.16
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0b5035b69e2011893044a066b819d4a21bdb64d19d3e1b29bdbac313a76640b5
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629857599.7e498afd5545.905.18
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:89e235409744eeaf21e0dc20dead8539999f0032993c80fb08d09a71a54e26ff
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629858276.7e498afd5545.905.20
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e8490d6025e42e5a0dbdd012564ebe0f57e0d403757bc257d5d1c103539ebcc5
|
| 3 |
+
size 8462
|