"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-138512 β checkpoint-139135}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-138512 β checkpoint-139135}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-138512 β checkpoint-139135}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-138512 β checkpoint-139135}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-138512 β checkpoint-139135}/rng_state.pth +2 -2
- model-bin/finetune/base/{checkpoint-138512 β checkpoint-139135}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-138512 β checkpoint-139135}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-138512 β checkpoint-139135}/trainer_state.json +799 -4
- model-bin/finetune/base/{checkpoint-138512 β checkpoint-139135}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630189707.037633/events.out.tfevents.1630189707.86bb0ddabf9b.4092.331 +3 -0
- model-bin/finetune/base/log/1630190103.5307064/events.out.tfevents.1630190103.86bb0ddabf9b.4092.333 +3 -0
- model-bin/finetune/base/log/1630190493.9352865/events.out.tfevents.1630190493.86bb0ddabf9b.4092.335 +3 -0
- model-bin/finetune/base/log/1630190885.1745086/events.out.tfevents.1630190885.86bb0ddabf9b.4092.337 +3 -0
- model-bin/finetune/base/log/1630191280.9197931/events.out.tfevents.1630191280.86bb0ddabf9b.4092.339 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630189707.86bb0ddabf9b.4092.330 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630190103.86bb0ddabf9b.4092.332 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630190493.86bb0ddabf9b.4092.334 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630190885.86bb0ddabf9b.4092.336 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630191280.86bb0ddabf9b.4092.338 +3 -0
model-bin/finetune/base/{checkpoint-138512 β checkpoint-139135}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-138512 β checkpoint-139135}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c37a6c10720a2f9c5a311eeb46c5ff3fd878b61fdb3d52018c8a201bf14df5c1
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-138512 β checkpoint-139135}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-138512 β checkpoint-139135}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0f3315de3d2b3896b78e3b43a47cecce273b45bcf96541a4ab3dd8799683e303
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-138512 β checkpoint-139135}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b86b810236303b881bbf5ca5ba5ed6f0c8e236a139fc8382041637bb625e4446
|
| 3 |
+
size 14503
|
model-bin/finetune/base/{checkpoint-138512 β checkpoint-139135}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0fbaa9c93c377bbc85cc6181cc1a8b53221928929faaa4dcbecfe2654907c58c
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-138512 β checkpoint-139135}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b27ebbc04d558d0510257f4a9ccd450abe095bc98ee9bdb09bb4fa64e7a21824
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-138512 β checkpoint-139135}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1689111747851003,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-132910",
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -275595,11 +275595,806 @@
|
|
| 275595 |
"eval_steps_per_second": 0.748,
|
| 275596 |
"eval_wer": 0.18466898954703834,
|
| 275597 |
"step": 138512
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 275598 |
}
|
| 275599 |
],
|
| 275600 |
-
"max_steps":
|
| 275601 |
"num_train_epochs": 5000,
|
| 275602 |
-
"total_flos": 3.
|
| 275603 |
"trial_name": null,
|
| 275604 |
"trial_params": null
|
| 275605 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1689111747851003,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-132910",
|
| 4 |
+
"epoch": 1122.0,
|
| 5 |
+
"global_step": 139135,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 275595 |
"eval_steps_per_second": 0.748,
|
| 275596 |
"eval_wer": 0.18466898954703834,
|
| 275597 |
"step": 138512
|
| 275598 |
+
},
|
| 275599 |
+
{
|
| 275600 |
+
"epoch": 1108.02,
|
| 275601 |
+
"learning_rate": 7.797115384615386e-06,
|
| 275602 |
+
"loss": 0.4021,
|
| 275603 |
+
"step": 138515
|
| 275604 |
+
},
|
| 275605 |
+
{
|
| 275606 |
+
"epoch": 1108.06,
|
| 275607 |
+
"learning_rate": 7.797035256410257e-06,
|
| 275608 |
+
"loss": 0.3586,
|
| 275609 |
+
"step": 138520
|
| 275610 |
+
},
|
| 275611 |
+
{
|
| 275612 |
+
"epoch": 1108.1,
|
| 275613 |
+
"learning_rate": 7.796955128205128e-06,
|
| 275614 |
+
"loss": 0.2991,
|
| 275615 |
+
"step": 138525
|
| 275616 |
+
},
|
| 275617 |
+
{
|
| 275618 |
+
"epoch": 1108.14,
|
| 275619 |
+
"learning_rate": 7.796875e-06,
|
| 275620 |
+
"loss": 0.3653,
|
| 275621 |
+
"step": 138530
|
| 275622 |
+
},
|
| 275623 |
+
{
|
| 275624 |
+
"epoch": 1108.18,
|
| 275625 |
+
"learning_rate": 7.796794871794873e-06,
|
| 275626 |
+
"loss": 0.7166,
|
| 275627 |
+
"step": 138535
|
| 275628 |
+
},
|
| 275629 |
+
{
|
| 275630 |
+
"epoch": 1108.22,
|
| 275631 |
+
"learning_rate": 7.796714743589744e-06,
|
| 275632 |
+
"loss": 0.7254,
|
| 275633 |
+
"step": 138540
|
| 275634 |
+
},
|
| 275635 |
+
{
|
| 275636 |
+
"epoch": 1108.26,
|
| 275637 |
+
"learning_rate": 7.796634615384616e-06,
|
| 275638 |
+
"loss": 0.3057,
|
| 275639 |
+
"step": 138545
|
| 275640 |
+
},
|
| 275641 |
+
{
|
| 275642 |
+
"epoch": 1108.3,
|
| 275643 |
+
"learning_rate": 7.796554487179487e-06,
|
| 275644 |
+
"loss": 0.3161,
|
| 275645 |
+
"step": 138550
|
| 275646 |
+
},
|
| 275647 |
+
{
|
| 275648 |
+
"epoch": 1108.34,
|
| 275649 |
+
"learning_rate": 7.79647435897436e-06,
|
| 275650 |
+
"loss": 0.3378,
|
| 275651 |
+
"step": 138555
|
| 275652 |
+
},
|
| 275653 |
+
{
|
| 275654 |
+
"epoch": 1108.38,
|
| 275655 |
+
"learning_rate": 7.796394230769231e-06,
|
| 275656 |
+
"loss": 0.7019,
|
| 275657 |
+
"step": 138560
|
| 275658 |
+
},
|
| 275659 |
+
{
|
| 275660 |
+
"epoch": 1108.42,
|
| 275661 |
+
"learning_rate": 7.796314102564103e-06,
|
| 275662 |
+
"loss": 0.848,
|
| 275663 |
+
"step": 138565
|
| 275664 |
+
},
|
| 275665 |
+
{
|
| 275666 |
+
"epoch": 1108.46,
|
| 275667 |
+
"learning_rate": 7.796233974358976e-06,
|
| 275668 |
+
"loss": 0.2843,
|
| 275669 |
+
"step": 138570
|
| 275670 |
+
},
|
| 275671 |
+
{
|
| 275672 |
+
"epoch": 1108.5,
|
| 275673 |
+
"learning_rate": 7.796153846153847e-06,
|
| 275674 |
+
"loss": 0.2654,
|
| 275675 |
+
"step": 138575
|
| 275676 |
+
},
|
| 275677 |
+
{
|
| 275678 |
+
"epoch": 1108.54,
|
| 275679 |
+
"learning_rate": 7.796073717948718e-06,
|
| 275680 |
+
"loss": 0.3823,
|
| 275681 |
+
"step": 138580
|
| 275682 |
+
},
|
| 275683 |
+
{
|
| 275684 |
+
"epoch": 1108.58,
|
| 275685 |
+
"learning_rate": 7.79599358974359e-06,
|
| 275686 |
+
"loss": 0.6801,
|
| 275687 |
+
"step": 138585
|
| 275688 |
+
},
|
| 275689 |
+
{
|
| 275690 |
+
"epoch": 1108.62,
|
| 275691 |
+
"learning_rate": 7.795913461538463e-06,
|
| 275692 |
+
"loss": 0.8877,
|
| 275693 |
+
"step": 138590
|
| 275694 |
+
},
|
| 275695 |
+
{
|
| 275696 |
+
"epoch": 1108.66,
|
| 275697 |
+
"learning_rate": 7.795833333333334e-06,
|
| 275698 |
+
"loss": 0.2561,
|
| 275699 |
+
"step": 138595
|
| 275700 |
+
},
|
| 275701 |
+
{
|
| 275702 |
+
"epoch": 1108.7,
|
| 275703 |
+
"learning_rate": 7.795753205128206e-06,
|
| 275704 |
+
"loss": 0.456,
|
| 275705 |
+
"step": 138600
|
| 275706 |
+
},
|
| 275707 |
+
{
|
| 275708 |
+
"epoch": 1108.74,
|
| 275709 |
+
"learning_rate": 7.795673076923077e-06,
|
| 275710 |
+
"loss": 0.2958,
|
| 275711 |
+
"step": 138605
|
| 275712 |
+
},
|
| 275713 |
+
{
|
| 275714 |
+
"epoch": 1108.78,
|
| 275715 |
+
"learning_rate": 7.79559294871795e-06,
|
| 275716 |
+
"loss": 0.8138,
|
| 275717 |
+
"step": 138610
|
| 275718 |
+
},
|
| 275719 |
+
{
|
| 275720 |
+
"epoch": 1108.82,
|
| 275721 |
+
"learning_rate": 7.795512820512821e-06,
|
| 275722 |
+
"loss": 0.7981,
|
| 275723 |
+
"step": 138615
|
| 275724 |
+
},
|
| 275725 |
+
{
|
| 275726 |
+
"epoch": 1108.86,
|
| 275727 |
+
"learning_rate": 7.795432692307693e-06,
|
| 275728 |
+
"loss": 0.2522,
|
| 275729 |
+
"step": 138620
|
| 275730 |
+
},
|
| 275731 |
+
{
|
| 275732 |
+
"epoch": 1108.9,
|
| 275733 |
+
"learning_rate": 7.795352564102566e-06,
|
| 275734 |
+
"loss": 0.3746,
|
| 275735 |
+
"step": 138625
|
| 275736 |
+
},
|
| 275737 |
+
{
|
| 275738 |
+
"epoch": 1108.94,
|
| 275739 |
+
"learning_rate": 7.795272435897437e-06,
|
| 275740 |
+
"loss": 0.3806,
|
| 275741 |
+
"step": 138630
|
| 275742 |
+
},
|
| 275743 |
+
{
|
| 275744 |
+
"epoch": 1108.98,
|
| 275745 |
+
"learning_rate": 7.795192307692308e-06,
|
| 275746 |
+
"loss": 0.7995,
|
| 275747 |
+
"step": 138635
|
| 275748 |
+
},
|
| 275749 |
+
{
|
| 275750 |
+
"epoch": 1109.0,
|
| 275751 |
+
"eval_loss": 0.3630291223526001,
|
| 275752 |
+
"eval_runtime": 35.9528,
|
| 275753 |
+
"eval_samples_per_second": 23.253,
|
| 275754 |
+
"eval_steps_per_second": 0.751,
|
| 275755 |
+
"eval_wer": 0.18319528393403478,
|
| 275756 |
+
"step": 138637
|
| 275757 |
+
},
|
| 275758 |
+
{
|
| 275759 |
+
"epoch": 1109.02,
|
| 275760 |
+
"learning_rate": 7.79511217948718e-06,
|
| 275761 |
+
"loss": 0.337,
|
| 275762 |
+
"step": 138640
|
| 275763 |
+
},
|
| 275764 |
+
{
|
| 275765 |
+
"epoch": 1109.06,
|
| 275766 |
+
"learning_rate": 7.795032051282053e-06,
|
| 275767 |
+
"loss": 0.2926,
|
| 275768 |
+
"step": 138645
|
| 275769 |
+
},
|
| 275770 |
+
{
|
| 275771 |
+
"epoch": 1109.1,
|
| 275772 |
+
"learning_rate": 7.794951923076923e-06,
|
| 275773 |
+
"loss": 0.3021,
|
| 275774 |
+
"step": 138650
|
| 275775 |
+
},
|
| 275776 |
+
{
|
| 275777 |
+
"epoch": 1109.14,
|
| 275778 |
+
"learning_rate": 7.794871794871796e-06,
|
| 275779 |
+
"loss": 0.3388,
|
| 275780 |
+
"step": 138655
|
| 275781 |
+
},
|
| 275782 |
+
{
|
| 275783 |
+
"epoch": 1109.18,
|
| 275784 |
+
"learning_rate": 7.794791666666667e-06,
|
| 275785 |
+
"loss": 0.8173,
|
| 275786 |
+
"step": 138660
|
| 275787 |
+
},
|
| 275788 |
+
{
|
| 275789 |
+
"epoch": 1109.22,
|
| 275790 |
+
"learning_rate": 7.794711538461538e-06,
|
| 275791 |
+
"loss": 1.0279,
|
| 275792 |
+
"step": 138665
|
| 275793 |
+
},
|
| 275794 |
+
{
|
| 275795 |
+
"epoch": 1109.26,
|
| 275796 |
+
"learning_rate": 7.794631410256411e-06,
|
| 275797 |
+
"loss": 0.265,
|
| 275798 |
+
"step": 138670
|
| 275799 |
+
},
|
| 275800 |
+
{
|
| 275801 |
+
"epoch": 1109.3,
|
| 275802 |
+
"learning_rate": 7.794551282051283e-06,
|
| 275803 |
+
"loss": 0.2509,
|
| 275804 |
+
"step": 138675
|
| 275805 |
+
},
|
| 275806 |
+
{
|
| 275807 |
+
"epoch": 1109.34,
|
| 275808 |
+
"learning_rate": 7.794471153846154e-06,
|
| 275809 |
+
"loss": 0.2781,
|
| 275810 |
+
"step": 138680
|
| 275811 |
+
},
|
| 275812 |
+
{
|
| 275813 |
+
"epoch": 1109.38,
|
| 275814 |
+
"learning_rate": 7.794391025641025e-06,
|
| 275815 |
+
"loss": 0.6717,
|
| 275816 |
+
"step": 138685
|
| 275817 |
+
},
|
| 275818 |
+
{
|
| 275819 |
+
"epoch": 1109.42,
|
| 275820 |
+
"learning_rate": 7.794310897435899e-06,
|
| 275821 |
+
"loss": 0.9528,
|
| 275822 |
+
"step": 138690
|
| 275823 |
+
},
|
| 275824 |
+
{
|
| 275825 |
+
"epoch": 1109.46,
|
| 275826 |
+
"learning_rate": 7.79423076923077e-06,
|
| 275827 |
+
"loss": 0.2734,
|
| 275828 |
+
"step": 138695
|
| 275829 |
+
},
|
| 275830 |
+
{
|
| 275831 |
+
"epoch": 1109.5,
|
| 275832 |
+
"learning_rate": 7.794150641025641e-06,
|
| 275833 |
+
"loss": 0.2718,
|
| 275834 |
+
"step": 138700
|
| 275835 |
+
},
|
| 275836 |
+
{
|
| 275837 |
+
"epoch": 1109.54,
|
| 275838 |
+
"learning_rate": 7.794070512820513e-06,
|
| 275839 |
+
"loss": 0.3879,
|
| 275840 |
+
"step": 138705
|
| 275841 |
+
},
|
| 275842 |
+
{
|
| 275843 |
+
"epoch": 1109.58,
|
| 275844 |
+
"learning_rate": 7.793990384615386e-06,
|
| 275845 |
+
"loss": 0.7201,
|
| 275846 |
+
"step": 138710
|
| 275847 |
+
},
|
| 275848 |
+
{
|
| 275849 |
+
"epoch": 1109.62,
|
| 275850 |
+
"learning_rate": 7.793910256410257e-06,
|
| 275851 |
+
"loss": 0.8282,
|
| 275852 |
+
"step": 138715
|
| 275853 |
+
},
|
| 275854 |
+
{
|
| 275855 |
+
"epoch": 1109.66,
|
| 275856 |
+
"learning_rate": 7.793830128205128e-06,
|
| 275857 |
+
"loss": 0.2921,
|
| 275858 |
+
"step": 138720
|
| 275859 |
+
},
|
| 275860 |
+
{
|
| 275861 |
+
"epoch": 1109.7,
|
| 275862 |
+
"learning_rate": 7.793750000000001e-06,
|
| 275863 |
+
"loss": 0.3045,
|
| 275864 |
+
"step": 138725
|
| 275865 |
+
},
|
| 275866 |
+
{
|
| 275867 |
+
"epoch": 1109.74,
|
| 275868 |
+
"learning_rate": 7.793669871794873e-06,
|
| 275869 |
+
"loss": 0.3144,
|
| 275870 |
+
"step": 138730
|
| 275871 |
+
},
|
| 275872 |
+
{
|
| 275873 |
+
"epoch": 1109.78,
|
| 275874 |
+
"learning_rate": 7.793589743589744e-06,
|
| 275875 |
+
"loss": 0.839,
|
| 275876 |
+
"step": 138735
|
| 275877 |
+
},
|
| 275878 |
+
{
|
| 275879 |
+
"epoch": 1109.82,
|
| 275880 |
+
"learning_rate": 7.793509615384615e-06,
|
| 275881 |
+
"loss": 0.8708,
|
| 275882 |
+
"step": 138740
|
| 275883 |
+
},
|
| 275884 |
+
{
|
| 275885 |
+
"epoch": 1109.86,
|
| 275886 |
+
"learning_rate": 7.793429487179489e-06,
|
| 275887 |
+
"loss": 0.2566,
|
| 275888 |
+
"step": 138745
|
| 275889 |
+
},
|
| 275890 |
+
{
|
| 275891 |
+
"epoch": 1109.9,
|
| 275892 |
+
"learning_rate": 7.79334935897436e-06,
|
| 275893 |
+
"loss": 0.2887,
|
| 275894 |
+
"step": 138750
|
| 275895 |
+
},
|
| 275896 |
+
{
|
| 275897 |
+
"epoch": 1109.94,
|
| 275898 |
+
"learning_rate": 7.793269230769231e-06,
|
| 275899 |
+
"loss": 0.3972,
|
| 275900 |
+
"step": 138755
|
| 275901 |
+
},
|
| 275902 |
+
{
|
| 275903 |
+
"epoch": 1109.98,
|
| 275904 |
+
"learning_rate": 7.793189102564103e-06,
|
| 275905 |
+
"loss": 0.7939,
|
| 275906 |
+
"step": 138760
|
| 275907 |
+
},
|
| 275908 |
+
{
|
| 275909 |
+
"epoch": 1110.0,
|
| 275910 |
+
"eval_loss": 0.4037955105304718,
|
| 275911 |
+
"eval_runtime": 36.9443,
|
| 275912 |
+
"eval_samples_per_second": 22.629,
|
| 275913 |
+
"eval_steps_per_second": 0.731,
|
| 275914 |
+
"eval_wer": 0.1821542476080023,
|
| 275915 |
+
"step": 138762
|
| 275916 |
+
},
|
| 275917 |
+
{
|
| 275918 |
+
"epoch": 1119.02,
|
| 275919 |
+
"learning_rate": 7.793108974358976e-06,
|
| 275920 |
+
"loss": 0.3205,
|
| 275921 |
+
"step": 138765
|
| 275922 |
+
},
|
| 275923 |
+
{
|
| 275924 |
+
"epoch": 1119.06,
|
| 275925 |
+
"learning_rate": 7.793028846153847e-06,
|
| 275926 |
+
"loss": 0.3026,
|
| 275927 |
+
"step": 138770
|
| 275928 |
+
},
|
| 275929 |
+
{
|
| 275930 |
+
"epoch": 1119.1,
|
| 275931 |
+
"learning_rate": 7.792948717948718e-06,
|
| 275932 |
+
"loss": 0.2726,
|
| 275933 |
+
"step": 138775
|
| 275934 |
+
},
|
| 275935 |
+
{
|
| 275936 |
+
"epoch": 1119.14,
|
| 275937 |
+
"learning_rate": 7.792868589743591e-06,
|
| 275938 |
+
"loss": 0.3822,
|
| 275939 |
+
"step": 138780
|
| 275940 |
+
},
|
| 275941 |
+
{
|
| 275942 |
+
"epoch": 1119.18,
|
| 275943 |
+
"learning_rate": 7.792788461538461e-06,
|
| 275944 |
+
"loss": 0.6922,
|
| 275945 |
+
"step": 138785
|
| 275946 |
+
},
|
| 275947 |
+
{
|
| 275948 |
+
"epoch": 1119.22,
|
| 275949 |
+
"learning_rate": 7.792708333333334e-06,
|
| 275950 |
+
"loss": 0.7749,
|
| 275951 |
+
"step": 138790
|
| 275952 |
+
},
|
| 275953 |
+
{
|
| 275954 |
+
"epoch": 1119.27,
|
| 275955 |
+
"learning_rate": 7.792628205128206e-06,
|
| 275956 |
+
"loss": 0.2817,
|
| 275957 |
+
"step": 138795
|
| 275958 |
+
},
|
| 275959 |
+
{
|
| 275960 |
+
"epoch": 1119.31,
|
| 275961 |
+
"learning_rate": 7.792548076923077e-06,
|
| 275962 |
+
"loss": 0.2985,
|
| 275963 |
+
"step": 138800
|
| 275964 |
+
},
|
| 275965 |
+
{
|
| 275966 |
+
"epoch": 1119.35,
|
| 275967 |
+
"learning_rate": 7.792467948717948e-06,
|
| 275968 |
+
"loss": 0.3379,
|
| 275969 |
+
"step": 138805
|
| 275970 |
+
},
|
| 275971 |
+
{
|
| 275972 |
+
"epoch": 1119.39,
|
| 275973 |
+
"learning_rate": 7.792387820512821e-06,
|
| 275974 |
+
"loss": 0.6886,
|
| 275975 |
+
"step": 138810
|
| 275976 |
+
},
|
| 275977 |
+
{
|
| 275978 |
+
"epoch": 1119.43,
|
| 275979 |
+
"learning_rate": 7.792307692307693e-06,
|
| 275980 |
+
"loss": 0.8331,
|
| 275981 |
+
"step": 138815
|
| 275982 |
+
},
|
| 275983 |
+
{
|
| 275984 |
+
"epoch": 1119.47,
|
| 275985 |
+
"learning_rate": 7.792227564102564e-06,
|
| 275986 |
+
"loss": 0.2971,
|
| 275987 |
+
"step": 138820
|
| 275988 |
+
},
|
| 275989 |
+
{
|
| 275990 |
+
"epoch": 1119.51,
|
| 275991 |
+
"learning_rate": 7.792147435897437e-06,
|
| 275992 |
+
"loss": 0.3145,
|
| 275993 |
+
"step": 138825
|
| 275994 |
+
},
|
| 275995 |
+
{
|
| 275996 |
+
"epoch": 1119.55,
|
| 275997 |
+
"learning_rate": 7.792067307692308e-06,
|
| 275998 |
+
"loss": 0.3495,
|
| 275999 |
+
"step": 138830
|
| 276000 |
+
},
|
| 276001 |
+
{
|
| 276002 |
+
"epoch": 1119.59,
|
| 276003 |
+
"learning_rate": 7.79198717948718e-06,
|
| 276004 |
+
"loss": 0.7801,
|
| 276005 |
+
"step": 138835
|
| 276006 |
+
},
|
| 276007 |
+
{
|
| 276008 |
+
"epoch": 1119.63,
|
| 276009 |
+
"learning_rate": 7.791907051282051e-06,
|
| 276010 |
+
"loss": 0.8077,
|
| 276011 |
+
"step": 138840
|
| 276012 |
+
},
|
| 276013 |
+
{
|
| 276014 |
+
"epoch": 1119.67,
|
| 276015 |
+
"learning_rate": 7.791826923076924e-06,
|
| 276016 |
+
"loss": 0.2616,
|
| 276017 |
+
"step": 138845
|
| 276018 |
+
},
|
| 276019 |
+
{
|
| 276020 |
+
"epoch": 1119.71,
|
| 276021 |
+
"learning_rate": 7.791746794871796e-06,
|
| 276022 |
+
"loss": 0.2896,
|
| 276023 |
+
"step": 138850
|
| 276024 |
+
},
|
| 276025 |
+
{
|
| 276026 |
+
"epoch": 1119.75,
|
| 276027 |
+
"learning_rate": 7.791666666666667e-06,
|
| 276028 |
+
"loss": 0.2968,
|
| 276029 |
+
"step": 138855
|
| 276030 |
+
},
|
| 276031 |
+
{
|
| 276032 |
+
"epoch": 1119.79,
|
| 276033 |
+
"learning_rate": 7.791586538461538e-06,
|
| 276034 |
+
"loss": 0.7159,
|
| 276035 |
+
"step": 138860
|
| 276036 |
+
},
|
| 276037 |
+
{
|
| 276038 |
+
"epoch": 1119.83,
|
| 276039 |
+
"learning_rate": 7.791506410256411e-06,
|
| 276040 |
+
"loss": 0.8684,
|
| 276041 |
+
"step": 138865
|
| 276042 |
+
},
|
| 276043 |
+
{
|
| 276044 |
+
"epoch": 1119.87,
|
| 276045 |
+
"learning_rate": 7.791426282051283e-06,
|
| 276046 |
+
"loss": 0.2864,
|
| 276047 |
+
"step": 138870
|
| 276048 |
+
},
|
| 276049 |
+
{
|
| 276050 |
+
"epoch": 1119.91,
|
| 276051 |
+
"learning_rate": 7.791346153846154e-06,
|
| 276052 |
+
"loss": 0.3273,
|
| 276053 |
+
"step": 138875
|
| 276054 |
+
},
|
| 276055 |
+
{
|
| 276056 |
+
"epoch": 1119.95,
|
| 276057 |
+
"learning_rate": 7.791266025641027e-06,
|
| 276058 |
+
"loss": 0.3125,
|
| 276059 |
+
"step": 138880
|
| 276060 |
+
},
|
| 276061 |
+
{
|
| 276062 |
+
"epoch": 1119.99,
|
| 276063 |
+
"learning_rate": 7.791185897435898e-06,
|
| 276064 |
+
"loss": 0.9501,
|
| 276065 |
+
"step": 138885
|
| 276066 |
+
},
|
| 276067 |
+
{
|
| 276068 |
+
"epoch": 1120.0,
|
| 276069 |
+
"eval_loss": 0.3200376629829407,
|
| 276070 |
+
"eval_runtime": 37.1182,
|
| 276071 |
+
"eval_samples_per_second": 22.496,
|
| 276072 |
+
"eval_steps_per_second": 0.727,
|
| 276073 |
+
"eval_wer": 0.17168076498712762,
|
| 276074 |
+
"step": 138886
|
| 276075 |
+
},
|
| 276076 |
+
{
|
| 276077 |
+
"epoch": 1111.03,
|
| 276078 |
+
"learning_rate": 7.79110576923077e-06,
|
| 276079 |
+
"loss": 0.4418,
|
| 276080 |
+
"step": 138890
|
| 276081 |
+
},
|
| 276082 |
+
{
|
| 276083 |
+
"epoch": 1111.07,
|
| 276084 |
+
"learning_rate": 7.791025641025641e-06,
|
| 276085 |
+
"loss": 0.2828,
|
| 276086 |
+
"step": 138895
|
| 276087 |
+
},
|
| 276088 |
+
{
|
| 276089 |
+
"epoch": 1111.11,
|
| 276090 |
+
"learning_rate": 7.790945512820514e-06,
|
| 276091 |
+
"loss": 0.2359,
|
| 276092 |
+
"step": 138900
|
| 276093 |
+
},
|
| 276094 |
+
{
|
| 276095 |
+
"epoch": 1111.15,
|
| 276096 |
+
"learning_rate": 7.790865384615386e-06,
|
| 276097 |
+
"loss": 0.3876,
|
| 276098 |
+
"step": 138905
|
| 276099 |
+
},
|
| 276100 |
+
{
|
| 276101 |
+
"epoch": 1111.19,
|
| 276102 |
+
"learning_rate": 7.790785256410257e-06,
|
| 276103 |
+
"loss": 0.9399,
|
| 276104 |
+
"step": 138910
|
| 276105 |
+
},
|
| 276106 |
+
{
|
| 276107 |
+
"epoch": 1111.23,
|
| 276108 |
+
"learning_rate": 7.79070512820513e-06,
|
| 276109 |
+
"loss": 0.6252,
|
| 276110 |
+
"step": 138915
|
| 276111 |
+
},
|
| 276112 |
+
{
|
| 276113 |
+
"epoch": 1111.27,
|
| 276114 |
+
"learning_rate": 7.790625000000001e-06,
|
| 276115 |
+
"loss": 0.265,
|
| 276116 |
+
"step": 138920
|
| 276117 |
+
},
|
| 276118 |
+
{
|
| 276119 |
+
"epoch": 1111.31,
|
| 276120 |
+
"learning_rate": 7.790544871794873e-06,
|
| 276121 |
+
"loss": 0.2534,
|
| 276122 |
+
"step": 138925
|
| 276123 |
+
},
|
| 276124 |
+
{
|
| 276125 |
+
"epoch": 1111.35,
|
| 276126 |
+
"learning_rate": 7.790464743589744e-06,
|
| 276127 |
+
"loss": 0.4191,
|
| 276128 |
+
"step": 138930
|
| 276129 |
+
},
|
| 276130 |
+
{
|
| 276131 |
+
"epoch": 1111.39,
|
| 276132 |
+
"learning_rate": 7.790384615384617e-06,
|
| 276133 |
+
"loss": 0.8346,
|
| 276134 |
+
"step": 138935
|
| 276135 |
+
},
|
| 276136 |
+
{
|
| 276137 |
+
"epoch": 1111.43,
|
| 276138 |
+
"learning_rate": 7.790304487179487e-06,
|
| 276139 |
+
"loss": 0.5682,
|
| 276140 |
+
"step": 138940
|
| 276141 |
+
},
|
| 276142 |
+
{
|
| 276143 |
+
"epoch": 1111.47,
|
| 276144 |
+
"learning_rate": 7.79022435897436e-06,
|
| 276145 |
+
"loss": 0.2528,
|
| 276146 |
+
"step": 138945
|
| 276147 |
+
},
|
| 276148 |
+
{
|
| 276149 |
+
"epoch": 1111.51,
|
| 276150 |
+
"learning_rate": 7.790144230769231e-06,
|
| 276151 |
+
"loss": 0.2672,
|
| 276152 |
+
"step": 138950
|
| 276153 |
+
},
|
| 276154 |
+
{
|
| 276155 |
+
"epoch": 1111.55,
|
| 276156 |
+
"learning_rate": 7.790064102564103e-06,
|
| 276157 |
+
"loss": 0.3749,
|
| 276158 |
+
"step": 138955
|
| 276159 |
+
},
|
| 276160 |
+
{
|
| 276161 |
+
"epoch": 1111.59,
|
| 276162 |
+
"learning_rate": 7.789983974358974e-06,
|
| 276163 |
+
"loss": 0.8947,
|
| 276164 |
+
"step": 138960
|
| 276165 |
+
},
|
| 276166 |
+
{
|
| 276167 |
+
"epoch": 1111.63,
|
| 276168 |
+
"learning_rate": 7.789903846153847e-06,
|
| 276169 |
+
"loss": 0.7404,
|
| 276170 |
+
"step": 138965
|
| 276171 |
+
},
|
| 276172 |
+
{
|
| 276173 |
+
"epoch": 1111.67,
|
| 276174 |
+
"learning_rate": 7.789823717948718e-06,
|
| 276175 |
+
"loss": 0.3066,
|
| 276176 |
+
"step": 138970
|
| 276177 |
+
},
|
| 276178 |
+
{
|
| 276179 |
+
"epoch": 1111.71,
|
| 276180 |
+
"learning_rate": 7.78974358974359e-06,
|
| 276181 |
+
"loss": 0.2967,
|
| 276182 |
+
"step": 138975
|
| 276183 |
+
},
|
| 276184 |
+
{
|
| 276185 |
+
"epoch": 1111.75,
|
| 276186 |
+
"learning_rate": 7.789663461538463e-06,
|
| 276187 |
+
"loss": 0.4045,
|
| 276188 |
+
"step": 138980
|
| 276189 |
+
},
|
| 276190 |
+
{
|
| 276191 |
+
"epoch": 1111.79,
|
| 276192 |
+
"learning_rate": 7.789583333333334e-06,
|
| 276193 |
+
"loss": 1.014,
|
| 276194 |
+
"step": 138985
|
| 276195 |
+
},
|
| 276196 |
+
{
|
| 276197 |
+
"epoch": 1111.83,
|
| 276198 |
+
"learning_rate": 7.789503205128205e-06,
|
| 276199 |
+
"loss": 0.6553,
|
| 276200 |
+
"step": 138990
|
| 276201 |
+
},
|
| 276202 |
+
{
|
| 276203 |
+
"epoch": 1111.87,
|
| 276204 |
+
"learning_rate": 7.789423076923077e-06,
|
| 276205 |
+
"loss": 0.2991,
|
| 276206 |
+
"step": 138995
|
| 276207 |
+
},
|
| 276208 |
+
{
|
| 276209 |
+
"epoch": 1111.91,
|
| 276210 |
+
"learning_rate": 7.78934294871795e-06,
|
| 276211 |
+
"loss": 0.3082,
|
| 276212 |
+
"step": 139000
|
| 276213 |
+
},
|
| 276214 |
+
{
|
| 276215 |
+
"epoch": 1111.95,
|
| 276216 |
+
"learning_rate": 7.789262820512821e-06,
|
| 276217 |
+
"loss": 0.3834,
|
| 276218 |
+
"step": 139005
|
| 276219 |
+
},
|
| 276220 |
+
{
|
| 276221 |
+
"epoch": 1111.99,
|
| 276222 |
+
"learning_rate": 7.789182692307693e-06,
|
| 276223 |
+
"loss": 1.0895,
|
| 276224 |
+
"step": 139010
|
| 276225 |
+
},
|
| 276226 |
+
{
|
| 276227 |
+
"epoch": 1112.0,
|
| 276228 |
+
"eval_loss": 0.31876030564308167,
|
| 276229 |
+
"eval_runtime": 37.3414,
|
| 276230 |
+
"eval_samples_per_second": 22.361,
|
| 276231 |
+
"eval_steps_per_second": 0.723,
|
| 276232 |
+
"eval_wer": 0.17835901110713007,
|
| 276233 |
+
"step": 139011
|
| 276234 |
+
},
|
| 276235 |
+
{
|
| 276236 |
+
"epoch": 1121.03,
|
| 276237 |
+
"learning_rate": 7.789102564102566e-06,
|
| 276238 |
+
"loss": 0.4076,
|
| 276239 |
+
"step": 139015
|
| 276240 |
+
},
|
| 276241 |
+
{
|
| 276242 |
+
"epoch": 1121.07,
|
| 276243 |
+
"learning_rate": 7.789022435897437e-06,
|
| 276244 |
+
"loss": 0.2467,
|
| 276245 |
+
"step": 139020
|
| 276246 |
+
},
|
| 276247 |
+
{
|
| 276248 |
+
"epoch": 1121.11,
|
| 276249 |
+
"learning_rate": 7.788942307692308e-06,
|
| 276250 |
+
"loss": 0.2756,
|
| 276251 |
+
"step": 139025
|
| 276252 |
+
},
|
| 276253 |
+
{
|
| 276254 |
+
"epoch": 1121.15,
|
| 276255 |
+
"learning_rate": 7.78886217948718e-06,
|
| 276256 |
+
"loss": 0.4109,
|
| 276257 |
+
"step": 139030
|
| 276258 |
+
},
|
| 276259 |
+
{
|
| 276260 |
+
"epoch": 1121.19,
|
| 276261 |
+
"learning_rate": 7.788782051282053e-06,
|
| 276262 |
+
"loss": 0.801,
|
| 276263 |
+
"step": 139035
|
| 276264 |
+
},
|
| 276265 |
+
{
|
| 276266 |
+
"epoch": 1121.23,
|
| 276267 |
+
"learning_rate": 7.788701923076924e-06,
|
| 276268 |
+
"loss": 0.6533,
|
| 276269 |
+
"step": 139040
|
| 276270 |
+
},
|
| 276271 |
+
{
|
| 276272 |
+
"epoch": 1121.27,
|
| 276273 |
+
"learning_rate": 7.788621794871795e-06,
|
| 276274 |
+
"loss": 0.279,
|
| 276275 |
+
"step": 139045
|
| 276276 |
+
},
|
| 276277 |
+
{
|
| 276278 |
+
"epoch": 1121.31,
|
| 276279 |
+
"learning_rate": 7.788541666666667e-06,
|
| 276280 |
+
"loss": 0.2622,
|
| 276281 |
+
"step": 139050
|
| 276282 |
+
},
|
| 276283 |
+
{
|
| 276284 |
+
"epoch": 1121.35,
|
| 276285 |
+
"learning_rate": 7.78846153846154e-06,
|
| 276286 |
+
"loss": 0.3736,
|
| 276287 |
+
"step": 139055
|
| 276288 |
+
},
|
| 276289 |
+
{
|
| 276290 |
+
"epoch": 1121.4,
|
| 276291 |
+
"learning_rate": 7.78838141025641e-06,
|
| 276292 |
+
"loss": 0.8162,
|
| 276293 |
+
"step": 139060
|
| 276294 |
+
},
|
| 276295 |
+
{
|
| 276296 |
+
"epoch": 1121.44,
|
| 276297 |
+
"learning_rate": 7.788301282051283e-06,
|
| 276298 |
+
"loss": 0.6164,
|
| 276299 |
+
"step": 139065
|
| 276300 |
+
},
|
| 276301 |
+
{
|
| 276302 |
+
"epoch": 1121.48,
|
| 276303 |
+
"learning_rate": 7.788221153846156e-06,
|
| 276304 |
+
"loss": 0.2711,
|
| 276305 |
+
"step": 139070
|
| 276306 |
+
},
|
| 276307 |
+
{
|
| 276308 |
+
"epoch": 1121.52,
|
| 276309 |
+
"learning_rate": 7.788141025641025e-06,
|
| 276310 |
+
"loss": 0.3306,
|
| 276311 |
+
"step": 139075
|
| 276312 |
+
},
|
| 276313 |
+
{
|
| 276314 |
+
"epoch": 1121.56,
|
| 276315 |
+
"learning_rate": 7.788060897435898e-06,
|
| 276316 |
+
"loss": 0.3728,
|
| 276317 |
+
"step": 139080
|
| 276318 |
+
},
|
| 276319 |
+
{
|
| 276320 |
+
"epoch": 1121.6,
|
| 276321 |
+
"learning_rate": 7.78798076923077e-06,
|
| 276322 |
+
"loss": 0.8783,
|
| 276323 |
+
"step": 139085
|
| 276324 |
+
},
|
| 276325 |
+
{
|
| 276326 |
+
"epoch": 1121.64,
|
| 276327 |
+
"learning_rate": 7.787900641025643e-06,
|
| 276328 |
+
"loss": 0.6683,
|
| 276329 |
+
"step": 139090
|
| 276330 |
+
},
|
| 276331 |
+
{
|
| 276332 |
+
"epoch": 1121.68,
|
| 276333 |
+
"learning_rate": 7.787820512820512e-06,
|
| 276334 |
+
"loss": 0.2538,
|
| 276335 |
+
"step": 139095
|
| 276336 |
+
},
|
| 276337 |
+
{
|
| 276338 |
+
"epoch": 1121.72,
|
| 276339 |
+
"learning_rate": 7.787740384615385e-06,
|
| 276340 |
+
"loss": 0.2686,
|
| 276341 |
+
"step": 139100
|
| 276342 |
+
},
|
| 276343 |
+
{
|
| 276344 |
+
"epoch": 1121.76,
|
| 276345 |
+
"learning_rate": 7.787660256410257e-06,
|
| 276346 |
+
"loss": 0.3903,
|
| 276347 |
+
"step": 139105
|
| 276348 |
+
},
|
| 276349 |
+
{
|
| 276350 |
+
"epoch": 1121.8,
|
| 276351 |
+
"learning_rate": 7.787580128205128e-06,
|
| 276352 |
+
"loss": 0.8713,
|
| 276353 |
+
"step": 139110
|
| 276354 |
+
},
|
| 276355 |
+
{
|
| 276356 |
+
"epoch": 1121.84,
|
| 276357 |
+
"learning_rate": 7.787500000000001e-06,
|
| 276358 |
+
"loss": 0.6841,
|
| 276359 |
+
"step": 139115
|
| 276360 |
+
},
|
| 276361 |
+
{
|
| 276362 |
+
"epoch": 1121.88,
|
| 276363 |
+
"learning_rate": 7.787419871794873e-06,
|
| 276364 |
+
"loss": 0.2623,
|
| 276365 |
+
"step": 139120
|
| 276366 |
+
},
|
| 276367 |
+
{
|
| 276368 |
+
"epoch": 1121.92,
|
| 276369 |
+
"learning_rate": 7.787339743589744e-06,
|
| 276370 |
+
"loss": 0.3016,
|
| 276371 |
+
"step": 139125
|
| 276372 |
+
},
|
| 276373 |
+
{
|
| 276374 |
+
"epoch": 1121.96,
|
| 276375 |
+
"learning_rate": 7.787259615384615e-06,
|
| 276376 |
+
"loss": 0.4188,
|
| 276377 |
+
"step": 139130
|
| 276378 |
+
},
|
| 276379 |
+
{
|
| 276380 |
+
"epoch": 1122.0,
|
| 276381 |
+
"learning_rate": 7.787179487179488e-06,
|
| 276382 |
+
"loss": 1.0256,
|
| 276383 |
+
"step": 139135
|
| 276384 |
+
},
|
| 276385 |
+
{
|
| 276386 |
+
"epoch": 1122.0,
|
| 276387 |
+
"eval_loss": 0.32184985280036926,
|
| 276388 |
+
"eval_runtime": 36.7062,
|
| 276389 |
+
"eval_samples_per_second": 22.775,
|
| 276390 |
+
"eval_steps_per_second": 0.736,
|
| 276391 |
+
"eval_wer": 0.17295327903986285,
|
| 276392 |
+
"step": 139135
|
| 276393 |
}
|
| 276394 |
],
|
| 276395 |
+
"max_steps": 620000,
|
| 276396 |
"num_train_epochs": 5000,
|
| 276397 |
+
"total_flos": 3.9151223111877696e+20,
|
| 276398 |
"trial_name": null,
|
| 276399 |
"trial_params": null
|
| 276400 |
}
|
model-bin/finetune/base/{checkpoint-138512 β checkpoint-139135}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1630189707.037633/events.out.tfevents.1630189707.86bb0ddabf9b.4092.331
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8105ba58ee49b2c56559d1735ef306fc11b3412d5aa62ace6eb4f7af781421d2
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630190103.5307064/events.out.tfevents.1630190103.86bb0ddabf9b.4092.333
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:200693c1140fe685b8e8d45edea9aa82f1728c63bdffbe00eadc2bd92917e971
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630190493.9352865/events.out.tfevents.1630190493.86bb0ddabf9b.4092.335
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:edaaa85003ae3661a4851bda98d346f4930d8a296393bdff574ead503c4f03b1
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630190885.1745086/events.out.tfevents.1630190885.86bb0ddabf9b.4092.337
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9b3b10de24853954f64ab646b1981760ac05579539eb207b2b0c89ffd131fed7
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630191280.9197931/events.out.tfevents.1630191280.86bb0ddabf9b.4092.339
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9ee95b301322cd1e205ffd09157857bdb545d31e980c4b20ec409309d672c0a9
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630189707.86bb0ddabf9b.4092.330
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c129546bf9928c795a684a2ce19fe177f9d615baf7f4d4b737511d43228c7efb
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630190103.86bb0ddabf9b.4092.332
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:23f1b1dd1249d1daa8ae4ae1823686c079346c976b897974a51ffcecce515e32
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630190493.86bb0ddabf9b.4092.334
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f042417cb824e599bc0221bf72238bacc13ca2ca8a3f35acc5ff60e8425bf0fb
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630190885.86bb0ddabf9b.4092.336
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:033c568364aa1cdf6e8112392edfb47829c42688caa01db22e892f71b2dec91b
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630191280.86bb0ddabf9b.4092.338
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ee2cf2dfdea2640274ecea5c578ef011e4294a8f0600836968ebac190e5ff70e
|
| 3 |
+
size 8622
|