"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-137888 β checkpoint-138512}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-137888 β checkpoint-138512}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-137888 β checkpoint-138512}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-137888 β checkpoint-138512}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-137888 β checkpoint-138512}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-137888 β checkpoint-138512}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-137888 β checkpoint-138512}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-137888 β checkpoint-138512}/trainer_state.json +798 -3
- model-bin/finetune/base/{checkpoint-137888 β checkpoint-138512}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630187518.4754748/events.out.tfevents.1630187518.86bb0ddabf9b.4092.321 +3 -0
- model-bin/finetune/base/log/1630187909.429926/events.out.tfevents.1630187909.86bb0ddabf9b.4092.323 +3 -0
- model-bin/finetune/base/log/1630188491.401316/events.out.tfevents.1630188491.86bb0ddabf9b.4092.325 +3 -0
- model-bin/finetune/base/log/1630188882.7400389/events.out.tfevents.1630188882.86bb0ddabf9b.4092.327 +3 -0
- model-bin/finetune/base/log/1630189275.6364982/events.out.tfevents.1630189275.86bb0ddabf9b.4092.329 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630187518.86bb0ddabf9b.4092.320 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630187909.86bb0ddabf9b.4092.322 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630188491.86bb0ddabf9b.4092.324 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630188882.86bb0ddabf9b.4092.326 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630189275.86bb0ddabf9b.4092.328 +3 -0
model-bin/finetune/base/{checkpoint-137888 β checkpoint-138512}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-137888 β checkpoint-138512}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:094ab79c0a676f26f006c9c635d691bf4bc34b3dc01b5b7926c0d9aab570700f
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-137888 β checkpoint-138512}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-137888 β checkpoint-138512}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f484fb1ae55345d9050588dd8093e0ae7735254644e7b3de1c21e031e2960f9b
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-137888 β checkpoint-138512}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14439
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:711aa21e4cd6a8f74e849c91658cece80afaa5f93c29945b86162a395eca0b19
|
| 3 |
size 14439
|
model-bin/finetune/base/{checkpoint-137888 β checkpoint-138512}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e557c50e42563d0842cba713dd1805ff147dbdcb2bd772e420247a4984b117cb
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-137888 β checkpoint-138512}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d10d085c3768e26bac8f8d1cefd6dd8978f641ae9735849c3524411405d4e5c0
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-137888 β checkpoint-138512}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1689111747851003,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-132910",
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -274800,11 +274800,806 @@
|
|
| 274800 |
"eval_steps_per_second": 0.749,
|
| 274801 |
"eval_wer": 0.19015407741450582,
|
| 274802 |
"step": 137888
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 274803 |
}
|
| 274804 |
],
|
| 274805 |
"max_steps": 625000,
|
| 274806 |
"num_train_epochs": 5000,
|
| 274807 |
-
"total_flos": 3.
|
| 274808 |
"trial_name": null,
|
| 274809 |
"trial_params": null
|
| 274810 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1689111747851003,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-132910",
|
| 4 |
+
"epoch": 1108.0,
|
| 5 |
+
"global_step": 138512,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 274800 |
"eval_steps_per_second": 0.749,
|
| 274801 |
"eval_wer": 0.19015407741450582,
|
| 274802 |
"step": 137888
|
| 274803 |
+
},
|
| 274804 |
+
{
|
| 274805 |
+
"epoch": 1103.02,
|
| 274806 |
+
"learning_rate": 7.80709935897436e-06,
|
| 274807 |
+
"loss": 0.3958,
|
| 274808 |
+
"step": 137890
|
| 274809 |
+
},
|
| 274810 |
+
{
|
| 274811 |
+
"epoch": 1103.06,
|
| 274812 |
+
"learning_rate": 7.807019230769231e-06,
|
| 274813 |
+
"loss": 0.3402,
|
| 274814 |
+
"step": 137895
|
| 274815 |
+
},
|
| 274816 |
+
{
|
| 274817 |
+
"epoch": 1103.1,
|
| 274818 |
+
"learning_rate": 7.806939102564102e-06,
|
| 274819 |
+
"loss": 0.2809,
|
| 274820 |
+
"step": 137900
|
| 274821 |
+
},
|
| 274822 |
+
{
|
| 274823 |
+
"epoch": 1103.14,
|
| 274824 |
+
"learning_rate": 7.806858974358975e-06,
|
| 274825 |
+
"loss": 0.3216,
|
| 274826 |
+
"step": 137905
|
| 274827 |
+
},
|
| 274828 |
+
{
|
| 274829 |
+
"epoch": 1103.18,
|
| 274830 |
+
"learning_rate": 7.806778846153847e-06,
|
| 274831 |
+
"loss": 0.581,
|
| 274832 |
+
"step": 137910
|
| 274833 |
+
},
|
| 274834 |
+
{
|
| 274835 |
+
"epoch": 1103.22,
|
| 274836 |
+
"learning_rate": 7.806698717948718e-06,
|
| 274837 |
+
"loss": 0.9586,
|
| 274838 |
+
"step": 137915
|
| 274839 |
+
},
|
| 274840 |
+
{
|
| 274841 |
+
"epoch": 1103.26,
|
| 274842 |
+
"learning_rate": 7.806618589743591e-06,
|
| 274843 |
+
"loss": 0.2833,
|
| 274844 |
+
"step": 137920
|
| 274845 |
+
},
|
| 274846 |
+
{
|
| 274847 |
+
"epoch": 1103.3,
|
| 274848 |
+
"learning_rate": 7.806538461538462e-06,
|
| 274849 |
+
"loss": 0.2956,
|
| 274850 |
+
"step": 137925
|
| 274851 |
+
},
|
| 274852 |
+
{
|
| 274853 |
+
"epoch": 1103.34,
|
| 274854 |
+
"learning_rate": 7.806458333333334e-06,
|
| 274855 |
+
"loss": 0.3089,
|
| 274856 |
+
"step": 137930
|
| 274857 |
+
},
|
| 274858 |
+
{
|
| 274859 |
+
"epoch": 1103.38,
|
| 274860 |
+
"learning_rate": 7.806378205128205e-06,
|
| 274861 |
+
"loss": 0.6143,
|
| 274862 |
+
"step": 137935
|
| 274863 |
+
},
|
| 274864 |
+
{
|
| 274865 |
+
"epoch": 1103.42,
|
| 274866 |
+
"learning_rate": 7.806298076923078e-06,
|
| 274867 |
+
"loss": 1.0034,
|
| 274868 |
+
"step": 137940
|
| 274869 |
+
},
|
| 274870 |
+
{
|
| 274871 |
+
"epoch": 1103.46,
|
| 274872 |
+
"learning_rate": 7.80621794871795e-06,
|
| 274873 |
+
"loss": 0.2756,
|
| 274874 |
+
"step": 137945
|
| 274875 |
+
},
|
| 274876 |
+
{
|
| 274877 |
+
"epoch": 1103.5,
|
| 274878 |
+
"learning_rate": 7.806137820512821e-06,
|
| 274879 |
+
"loss": 0.2822,
|
| 274880 |
+
"step": 137950
|
| 274881 |
+
},
|
| 274882 |
+
{
|
| 274883 |
+
"epoch": 1103.54,
|
| 274884 |
+
"learning_rate": 7.806057692307692e-06,
|
| 274885 |
+
"loss": 0.3213,
|
| 274886 |
+
"step": 137955
|
| 274887 |
+
},
|
| 274888 |
+
{
|
| 274889 |
+
"epoch": 1103.58,
|
| 274890 |
+
"learning_rate": 7.805977564102565e-06,
|
| 274891 |
+
"loss": 0.5596,
|
| 274892 |
+
"step": 137960
|
| 274893 |
+
},
|
| 274894 |
+
{
|
| 274895 |
+
"epoch": 1103.62,
|
| 274896 |
+
"learning_rate": 7.805897435897437e-06,
|
| 274897 |
+
"loss": 0.9454,
|
| 274898 |
+
"step": 137965
|
| 274899 |
+
},
|
| 274900 |
+
{
|
| 274901 |
+
"epoch": 1103.66,
|
| 274902 |
+
"learning_rate": 7.805817307692308e-06,
|
| 274903 |
+
"loss": 0.2599,
|
| 274904 |
+
"step": 137970
|
| 274905 |
+
},
|
| 274906 |
+
{
|
| 274907 |
+
"epoch": 1103.7,
|
| 274908 |
+
"learning_rate": 7.805737179487181e-06,
|
| 274909 |
+
"loss": 0.2692,
|
| 274910 |
+
"step": 137975
|
| 274911 |
+
},
|
| 274912 |
+
{
|
| 274913 |
+
"epoch": 1103.74,
|
| 274914 |
+
"learning_rate": 7.805657051282052e-06,
|
| 274915 |
+
"loss": 0.2986,
|
| 274916 |
+
"step": 137980
|
| 274917 |
+
},
|
| 274918 |
+
{
|
| 274919 |
+
"epoch": 1103.78,
|
| 274920 |
+
"learning_rate": 7.805576923076924e-06,
|
| 274921 |
+
"loss": 0.4933,
|
| 274922 |
+
"step": 137985
|
| 274923 |
+
},
|
| 274924 |
+
{
|
| 274925 |
+
"epoch": 1103.82,
|
| 274926 |
+
"learning_rate": 7.805496794871795e-06,
|
| 274927 |
+
"loss": 0.943,
|
| 274928 |
+
"step": 137990
|
| 274929 |
+
},
|
| 274930 |
+
{
|
| 274931 |
+
"epoch": 1103.86,
|
| 274932 |
+
"learning_rate": 7.805416666666668e-06,
|
| 274933 |
+
"loss": 0.2714,
|
| 274934 |
+
"step": 137995
|
| 274935 |
+
},
|
| 274936 |
+
{
|
| 274937 |
+
"epoch": 1103.9,
|
| 274938 |
+
"learning_rate": 7.805336538461538e-06,
|
| 274939 |
+
"loss": 0.2898,
|
| 274940 |
+
"step": 138000
|
| 274941 |
+
},
|
| 274942 |
+
{
|
| 274943 |
+
"epoch": 1103.94,
|
| 274944 |
+
"learning_rate": 7.805256410256411e-06,
|
| 274945 |
+
"loss": 0.294,
|
| 274946 |
+
"step": 138005
|
| 274947 |
+
},
|
| 274948 |
+
{
|
| 274949 |
+
"epoch": 1103.98,
|
| 274950 |
+
"learning_rate": 7.805176282051282e-06,
|
| 274951 |
+
"loss": 0.5794,
|
| 274952 |
+
"step": 138010
|
| 274953 |
+
},
|
| 274954 |
+
{
|
| 274955 |
+
"epoch": 1104.0,
|
| 274956 |
+
"eval_loss": 0.3462262451648712,
|
| 274957 |
+
"eval_runtime": 37.3606,
|
| 274958 |
+
"eval_samples_per_second": 22.564,
|
| 274959 |
+
"eval_steps_per_second": 0.723,
|
| 274960 |
+
"eval_wer": 0.17421728466310776,
|
| 274961 |
+
"step": 138013
|
| 274962 |
+
},
|
| 274963 |
+
{
|
| 274964 |
+
"epoch": 1113.02,
|
| 274965 |
+
"learning_rate": 7.805096153846154e-06,
|
| 274966 |
+
"loss": 0.3059,
|
| 274967 |
+
"step": 138015
|
| 274968 |
+
},
|
| 274969 |
+
{
|
| 274970 |
+
"epoch": 1113.06,
|
| 274971 |
+
"learning_rate": 7.805016025641027e-06,
|
| 274972 |
+
"loss": 0.2681,
|
| 274973 |
+
"step": 138020
|
| 274974 |
+
},
|
| 274975 |
+
{
|
| 274976 |
+
"epoch": 1113.1,
|
| 274977 |
+
"learning_rate": 7.804935897435898e-06,
|
| 274978 |
+
"loss": 0.2699,
|
| 274979 |
+
"step": 138025
|
| 274980 |
+
},
|
| 274981 |
+
{
|
| 274982 |
+
"epoch": 1113.14,
|
| 274983 |
+
"learning_rate": 7.80485576923077e-06,
|
| 274984 |
+
"loss": 0.3977,
|
| 274985 |
+
"step": 138030
|
| 274986 |
+
},
|
| 274987 |
+
{
|
| 274988 |
+
"epoch": 1113.18,
|
| 274989 |
+
"learning_rate": 7.80477564102564e-06,
|
| 274990 |
+
"loss": 0.7177,
|
| 274991 |
+
"step": 138035
|
| 274992 |
+
},
|
| 274993 |
+
{
|
| 274994 |
+
"epoch": 1113.22,
|
| 274995 |
+
"learning_rate": 7.804711538461538e-06,
|
| 274996 |
+
"loss": 0.94,
|
| 274997 |
+
"step": 138040
|
| 274998 |
+
},
|
| 274999 |
+
{
|
| 275000 |
+
"epoch": 1113.26,
|
| 275001 |
+
"learning_rate": 7.804631410256411e-06,
|
| 275002 |
+
"loss": 0.3183,
|
| 275003 |
+
"step": 138045
|
| 275004 |
+
},
|
| 275005 |
+
{
|
| 275006 |
+
"epoch": 1113.3,
|
| 275007 |
+
"learning_rate": 7.804551282051283e-06,
|
| 275008 |
+
"loss": 0.2608,
|
| 275009 |
+
"step": 138050
|
| 275010 |
+
},
|
| 275011 |
+
{
|
| 275012 |
+
"epoch": 1113.34,
|
| 275013 |
+
"learning_rate": 7.804471153846154e-06,
|
| 275014 |
+
"loss": 0.3397,
|
| 275015 |
+
"step": 138055
|
| 275016 |
+
},
|
| 275017 |
+
{
|
| 275018 |
+
"epoch": 1113.38,
|
| 275019 |
+
"learning_rate": 7.804391025641025e-06,
|
| 275020 |
+
"loss": 0.5871,
|
| 275021 |
+
"step": 138060
|
| 275022 |
+
},
|
| 275023 |
+
{
|
| 275024 |
+
"epoch": 1113.42,
|
| 275025 |
+
"learning_rate": 7.804310897435898e-06,
|
| 275026 |
+
"loss": 1.0535,
|
| 275027 |
+
"step": 138065
|
| 275028 |
+
},
|
| 275029 |
+
{
|
| 275030 |
+
"epoch": 1113.46,
|
| 275031 |
+
"learning_rate": 7.80423076923077e-06,
|
| 275032 |
+
"loss": 0.2598,
|
| 275033 |
+
"step": 138070
|
| 275034 |
+
},
|
| 275035 |
+
{
|
| 275036 |
+
"epoch": 1113.5,
|
| 275037 |
+
"learning_rate": 7.804150641025641e-06,
|
| 275038 |
+
"loss": 0.2877,
|
| 275039 |
+
"step": 138075
|
| 275040 |
+
},
|
| 275041 |
+
{
|
| 275042 |
+
"epoch": 1113.54,
|
| 275043 |
+
"learning_rate": 7.804070512820512e-06,
|
| 275044 |
+
"loss": 0.3107,
|
| 275045 |
+
"step": 138080
|
| 275046 |
+
},
|
| 275047 |
+
{
|
| 275048 |
+
"epoch": 1113.58,
|
| 275049 |
+
"learning_rate": 7.803990384615386e-06,
|
| 275050 |
+
"loss": 0.56,
|
| 275051 |
+
"step": 138085
|
| 275052 |
+
},
|
| 275053 |
+
{
|
| 275054 |
+
"epoch": 1113.62,
|
| 275055 |
+
"learning_rate": 7.803910256410257e-06,
|
| 275056 |
+
"loss": 1.0354,
|
| 275057 |
+
"step": 138090
|
| 275058 |
+
},
|
| 275059 |
+
{
|
| 275060 |
+
"epoch": 1113.66,
|
| 275061 |
+
"learning_rate": 7.803830128205128e-06,
|
| 275062 |
+
"loss": 0.315,
|
| 275063 |
+
"step": 138095
|
| 275064 |
+
},
|
| 275065 |
+
{
|
| 275066 |
+
"epoch": 1113.7,
|
| 275067 |
+
"learning_rate": 7.803750000000001e-06,
|
| 275068 |
+
"loss": 0.323,
|
| 275069 |
+
"step": 138100
|
| 275070 |
+
},
|
| 275071 |
+
{
|
| 275072 |
+
"epoch": 1113.74,
|
| 275073 |
+
"learning_rate": 7.803669871794873e-06,
|
| 275074 |
+
"loss": 0.3117,
|
| 275075 |
+
"step": 138105
|
| 275076 |
+
},
|
| 275077 |
+
{
|
| 275078 |
+
"epoch": 1113.78,
|
| 275079 |
+
"learning_rate": 7.803589743589744e-06,
|
| 275080 |
+
"loss": 0.6508,
|
| 275081 |
+
"step": 138110
|
| 275082 |
+
},
|
| 275083 |
+
{
|
| 275084 |
+
"epoch": 1113.82,
|
| 275085 |
+
"learning_rate": 7.803509615384615e-06,
|
| 275086 |
+
"loss": 1.0869,
|
| 275087 |
+
"step": 138115
|
| 275088 |
+
},
|
| 275089 |
+
{
|
| 275090 |
+
"epoch": 1113.86,
|
| 275091 |
+
"learning_rate": 7.803429487179488e-06,
|
| 275092 |
+
"loss": 0.2849,
|
| 275093 |
+
"step": 138120
|
| 275094 |
+
},
|
| 275095 |
+
{
|
| 275096 |
+
"epoch": 1113.9,
|
| 275097 |
+
"learning_rate": 7.80334935897436e-06,
|
| 275098 |
+
"loss": 0.2898,
|
| 275099 |
+
"step": 138125
|
| 275100 |
+
},
|
| 275101 |
+
{
|
| 275102 |
+
"epoch": 1113.94,
|
| 275103 |
+
"learning_rate": 7.803269230769231e-06,
|
| 275104 |
+
"loss": 0.4061,
|
| 275105 |
+
"step": 138130
|
| 275106 |
+
},
|
| 275107 |
+
{
|
| 275108 |
+
"epoch": 1113.98,
|
| 275109 |
+
"learning_rate": 7.803189102564103e-06,
|
| 275110 |
+
"loss": 0.7345,
|
| 275111 |
+
"step": 138135
|
| 275112 |
+
},
|
| 275113 |
+
{
|
| 275114 |
+
"epoch": 1114.0,
|
| 275115 |
+
"eval_loss": 0.31746795773506165,
|
| 275116 |
+
"eval_runtime": 51.8192,
|
| 275117 |
+
"eval_samples_per_second": 16.268,
|
| 275118 |
+
"eval_steps_per_second": 0.521,
|
| 275119 |
+
"eval_wer": 0.17246036823771949,
|
| 275120 |
+
"step": 138137
|
| 275121 |
+
},
|
| 275122 |
+
{
|
| 275123 |
+
"epoch": 1105.02,
|
| 275124 |
+
"learning_rate": 7.803108974358976e-06,
|
| 275125 |
+
"loss": 0.3401,
|
| 275126 |
+
"step": 138140
|
| 275127 |
+
},
|
| 275128 |
+
{
|
| 275129 |
+
"epoch": 1105.06,
|
| 275130 |
+
"learning_rate": 7.803028846153847e-06,
|
| 275131 |
+
"loss": 0.2625,
|
| 275132 |
+
"step": 138145
|
| 275133 |
+
},
|
| 275134 |
+
{
|
| 275135 |
+
"epoch": 1105.1,
|
| 275136 |
+
"learning_rate": 7.802948717948718e-06,
|
| 275137 |
+
"loss": 0.3134,
|
| 275138 |
+
"step": 138150
|
| 275139 |
+
},
|
| 275140 |
+
{
|
| 275141 |
+
"epoch": 1105.14,
|
| 275142 |
+
"learning_rate": 7.802868589743591e-06,
|
| 275143 |
+
"loss": 0.2992,
|
| 275144 |
+
"step": 138155
|
| 275145 |
+
},
|
| 275146 |
+
{
|
| 275147 |
+
"epoch": 1105.18,
|
| 275148 |
+
"learning_rate": 7.802788461538463e-06,
|
| 275149 |
+
"loss": 0.6537,
|
| 275150 |
+
"step": 138160
|
| 275151 |
+
},
|
| 275152 |
+
{
|
| 275153 |
+
"epoch": 1105.22,
|
| 275154 |
+
"learning_rate": 7.80272435897436e-06,
|
| 275155 |
+
"loss": 0.8494,
|
| 275156 |
+
"step": 138165
|
| 275157 |
+
},
|
| 275158 |
+
{
|
| 275159 |
+
"epoch": 1105.26,
|
| 275160 |
+
"learning_rate": 7.802644230769231e-06,
|
| 275161 |
+
"loss": 0.2404,
|
| 275162 |
+
"step": 138170
|
| 275163 |
+
},
|
| 275164 |
+
{
|
| 275165 |
+
"epoch": 1105.3,
|
| 275166 |
+
"learning_rate": 7.802564102564103e-06,
|
| 275167 |
+
"loss": 0.2829,
|
| 275168 |
+
"step": 138175
|
| 275169 |
+
},
|
| 275170 |
+
{
|
| 275171 |
+
"epoch": 1105.34,
|
| 275172 |
+
"learning_rate": 7.802483974358976e-06,
|
| 275173 |
+
"loss": 0.3426,
|
| 275174 |
+
"step": 138180
|
| 275175 |
+
},
|
| 275176 |
+
{
|
| 275177 |
+
"epoch": 1105.38,
|
| 275178 |
+
"learning_rate": 7.802403846153847e-06,
|
| 275179 |
+
"loss": 0.6896,
|
| 275180 |
+
"step": 138185
|
| 275181 |
+
},
|
| 275182 |
+
{
|
| 275183 |
+
"epoch": 1105.42,
|
| 275184 |
+
"learning_rate": 7.802323717948719e-06,
|
| 275185 |
+
"loss": 0.7599,
|
| 275186 |
+
"step": 138190
|
| 275187 |
+
},
|
| 275188 |
+
{
|
| 275189 |
+
"epoch": 1105.46,
|
| 275190 |
+
"learning_rate": 7.80224358974359e-06,
|
| 275191 |
+
"loss": 0.2685,
|
| 275192 |
+
"step": 138195
|
| 275193 |
+
},
|
| 275194 |
+
{
|
| 275195 |
+
"epoch": 1105.5,
|
| 275196 |
+
"learning_rate": 7.802163461538463e-06,
|
| 275197 |
+
"loss": 0.258,
|
| 275198 |
+
"step": 138200
|
| 275199 |
+
},
|
| 275200 |
+
{
|
| 275201 |
+
"epoch": 1105.54,
|
| 275202 |
+
"learning_rate": 7.802083333333333e-06,
|
| 275203 |
+
"loss": 0.3245,
|
| 275204 |
+
"step": 138205
|
| 275205 |
+
},
|
| 275206 |
+
{
|
| 275207 |
+
"epoch": 1105.58,
|
| 275208 |
+
"learning_rate": 7.802003205128206e-06,
|
| 275209 |
+
"loss": 0.836,
|
| 275210 |
+
"step": 138210
|
| 275211 |
+
},
|
| 275212 |
+
{
|
| 275213 |
+
"epoch": 1105.62,
|
| 275214 |
+
"learning_rate": 7.801923076923079e-06,
|
| 275215 |
+
"loss": 0.9533,
|
| 275216 |
+
"step": 138215
|
| 275217 |
+
},
|
| 275218 |
+
{
|
| 275219 |
+
"epoch": 1105.66,
|
| 275220 |
+
"learning_rate": 7.801842948717948e-06,
|
| 275221 |
+
"loss": 0.256,
|
| 275222 |
+
"step": 138220
|
| 275223 |
+
},
|
| 275224 |
+
{
|
| 275225 |
+
"epoch": 1105.7,
|
| 275226 |
+
"learning_rate": 7.801762820512822e-06,
|
| 275227 |
+
"loss": 0.2897,
|
| 275228 |
+
"step": 138225
|
| 275229 |
+
},
|
| 275230 |
+
{
|
| 275231 |
+
"epoch": 1105.74,
|
| 275232 |
+
"learning_rate": 7.801682692307693e-06,
|
| 275233 |
+
"loss": 0.3314,
|
| 275234 |
+
"step": 138230
|
| 275235 |
+
},
|
| 275236 |
+
{
|
| 275237 |
+
"epoch": 1105.78,
|
| 275238 |
+
"learning_rate": 7.801602564102564e-06,
|
| 275239 |
+
"loss": 0.7014,
|
| 275240 |
+
"step": 138235
|
| 275241 |
+
},
|
| 275242 |
+
{
|
| 275243 |
+
"epoch": 1105.82,
|
| 275244 |
+
"learning_rate": 7.801522435897436e-06,
|
| 275245 |
+
"loss": 0.9438,
|
| 275246 |
+
"step": 138240
|
| 275247 |
+
},
|
| 275248 |
+
{
|
| 275249 |
+
"epoch": 1105.86,
|
| 275250 |
+
"learning_rate": 7.801442307692309e-06,
|
| 275251 |
+
"loss": 0.2693,
|
| 275252 |
+
"step": 138245
|
| 275253 |
+
},
|
| 275254 |
+
{
|
| 275255 |
+
"epoch": 1105.9,
|
| 275256 |
+
"learning_rate": 7.80136217948718e-06,
|
| 275257 |
+
"loss": 0.2833,
|
| 275258 |
+
"step": 138250
|
| 275259 |
+
},
|
| 275260 |
+
{
|
| 275261 |
+
"epoch": 1105.94,
|
| 275262 |
+
"learning_rate": 7.801282051282051e-06,
|
| 275263 |
+
"loss": 0.3501,
|
| 275264 |
+
"step": 138255
|
| 275265 |
+
},
|
| 275266 |
+
{
|
| 275267 |
+
"epoch": 1105.98,
|
| 275268 |
+
"learning_rate": 7.801201923076924e-06,
|
| 275269 |
+
"loss": 0.7127,
|
| 275270 |
+
"step": 138260
|
| 275271 |
+
},
|
| 275272 |
+
{
|
| 275273 |
+
"epoch": 1106.0,
|
| 275274 |
+
"eval_loss": 0.38287559151649475,
|
| 275275 |
+
"eval_runtime": 35.5403,
|
| 275276 |
+
"eval_samples_per_second": 23.523,
|
| 275277 |
+
"eval_steps_per_second": 0.76,
|
| 275278 |
+
"eval_wer": 0.17257330448407074,
|
| 275279 |
+
"step": 138262
|
| 275280 |
+
},
|
| 275281 |
+
{
|
| 275282 |
+
"epoch": 1106.02,
|
| 275283 |
+
"learning_rate": 7.801121794871796e-06,
|
| 275284 |
+
"loss": 0.3405,
|
| 275285 |
+
"step": 138265
|
| 275286 |
+
},
|
| 275287 |
+
{
|
| 275288 |
+
"epoch": 1106.06,
|
| 275289 |
+
"learning_rate": 7.801041666666667e-06,
|
| 275290 |
+
"loss": 0.2896,
|
| 275291 |
+
"step": 138270
|
| 275292 |
+
},
|
| 275293 |
+
{
|
| 275294 |
+
"epoch": 1106.1,
|
| 275295 |
+
"learning_rate": 7.800961538461538e-06,
|
| 275296 |
+
"loss": 0.2766,
|
| 275297 |
+
"step": 138275
|
| 275298 |
+
},
|
| 275299 |
+
{
|
| 275300 |
+
"epoch": 1106.14,
|
| 275301 |
+
"learning_rate": 7.800881410256412e-06,
|
| 275302 |
+
"loss": 0.3587,
|
| 275303 |
+
"step": 138280
|
| 275304 |
+
},
|
| 275305 |
+
{
|
| 275306 |
+
"epoch": 1106.18,
|
| 275307 |
+
"learning_rate": 7.800801282051283e-06,
|
| 275308 |
+
"loss": 0.6423,
|
| 275309 |
+
"step": 138285
|
| 275310 |
+
},
|
| 275311 |
+
{
|
| 275312 |
+
"epoch": 1106.22,
|
| 275313 |
+
"learning_rate": 7.800721153846154e-06,
|
| 275314 |
+
"loss": 0.7485,
|
| 275315 |
+
"step": 138290
|
| 275316 |
+
},
|
| 275317 |
+
{
|
| 275318 |
+
"epoch": 1106.26,
|
| 275319 |
+
"learning_rate": 7.800641025641026e-06,
|
| 275320 |
+
"loss": 0.264,
|
| 275321 |
+
"step": 138295
|
| 275322 |
+
},
|
| 275323 |
+
{
|
| 275324 |
+
"epoch": 1106.3,
|
| 275325 |
+
"learning_rate": 7.800560897435899e-06,
|
| 275326 |
+
"loss": 0.2421,
|
| 275327 |
+
"step": 138300
|
| 275328 |
+
},
|
| 275329 |
+
{
|
| 275330 |
+
"epoch": 1106.34,
|
| 275331 |
+
"learning_rate": 7.80048076923077e-06,
|
| 275332 |
+
"loss": 0.401,
|
| 275333 |
+
"step": 138305
|
| 275334 |
+
},
|
| 275335 |
+
{
|
| 275336 |
+
"epoch": 1106.38,
|
| 275337 |
+
"learning_rate": 7.800400641025641e-06,
|
| 275338 |
+
"loss": 0.7231,
|
| 275339 |
+
"step": 138310
|
| 275340 |
+
},
|
| 275341 |
+
{
|
| 275342 |
+
"epoch": 1106.42,
|
| 275343 |
+
"learning_rate": 7.800320512820514e-06,
|
| 275344 |
+
"loss": 0.8366,
|
| 275345 |
+
"step": 138315
|
| 275346 |
+
},
|
| 275347 |
+
{
|
| 275348 |
+
"epoch": 1106.46,
|
| 275349 |
+
"learning_rate": 7.800240384615386e-06,
|
| 275350 |
+
"loss": 0.2772,
|
| 275351 |
+
"step": 138320
|
| 275352 |
+
},
|
| 275353 |
+
{
|
| 275354 |
+
"epoch": 1106.5,
|
| 275355 |
+
"learning_rate": 7.800160256410257e-06,
|
| 275356 |
+
"loss": 0.2557,
|
| 275357 |
+
"step": 138325
|
| 275358 |
+
},
|
| 275359 |
+
{
|
| 275360 |
+
"epoch": 1106.54,
|
| 275361 |
+
"learning_rate": 7.800080128205129e-06,
|
| 275362 |
+
"loss": 0.3251,
|
| 275363 |
+
"step": 138330
|
| 275364 |
+
},
|
| 275365 |
+
{
|
| 275366 |
+
"epoch": 1106.58,
|
| 275367 |
+
"learning_rate": 7.800000000000002e-06,
|
| 275368 |
+
"loss": 0.6557,
|
| 275369 |
+
"step": 138335
|
| 275370 |
+
},
|
| 275371 |
+
{
|
| 275372 |
+
"epoch": 1106.62,
|
| 275373 |
+
"learning_rate": 7.799919871794873e-06,
|
| 275374 |
+
"loss": 0.7998,
|
| 275375 |
+
"step": 138340
|
| 275376 |
+
},
|
| 275377 |
+
{
|
| 275378 |
+
"epoch": 1106.66,
|
| 275379 |
+
"learning_rate": 7.799839743589744e-06,
|
| 275380 |
+
"loss": 0.2524,
|
| 275381 |
+
"step": 138345
|
| 275382 |
+
},
|
| 275383 |
+
{
|
| 275384 |
+
"epoch": 1106.7,
|
| 275385 |
+
"learning_rate": 7.799759615384616e-06,
|
| 275386 |
+
"loss": 0.2873,
|
| 275387 |
+
"step": 138350
|
| 275388 |
+
},
|
| 275389 |
+
{
|
| 275390 |
+
"epoch": 1106.74,
|
| 275391 |
+
"learning_rate": 7.799679487179489e-06,
|
| 275392 |
+
"loss": 0.3847,
|
| 275393 |
+
"step": 138355
|
| 275394 |
+
},
|
| 275395 |
+
{
|
| 275396 |
+
"epoch": 1106.78,
|
| 275397 |
+
"learning_rate": 7.79959935897436e-06,
|
| 275398 |
+
"loss": 0.7309,
|
| 275399 |
+
"step": 138360
|
| 275400 |
+
},
|
| 275401 |
+
{
|
| 275402 |
+
"epoch": 1106.82,
|
| 275403 |
+
"learning_rate": 7.799519230769231e-06,
|
| 275404 |
+
"loss": 0.7499,
|
| 275405 |
+
"step": 138365
|
| 275406 |
+
},
|
| 275407 |
+
{
|
| 275408 |
+
"epoch": 1106.86,
|
| 275409 |
+
"learning_rate": 7.799439102564104e-06,
|
| 275410 |
+
"loss": 0.3012,
|
| 275411 |
+
"step": 138370
|
| 275412 |
+
},
|
| 275413 |
+
{
|
| 275414 |
+
"epoch": 1106.9,
|
| 275415 |
+
"learning_rate": 7.799358974358974e-06,
|
| 275416 |
+
"loss": 0.2795,
|
| 275417 |
+
"step": 138375
|
| 275418 |
+
},
|
| 275419 |
+
{
|
| 275420 |
+
"epoch": 1106.94,
|
| 275421 |
+
"learning_rate": 7.799278846153847e-06,
|
| 275422 |
+
"loss": 0.3877,
|
| 275423 |
+
"step": 138380
|
| 275424 |
+
},
|
| 275425 |
+
{
|
| 275426 |
+
"epoch": 1106.98,
|
| 275427 |
+
"learning_rate": 7.799198717948719e-06,
|
| 275428 |
+
"loss": 0.6934,
|
| 275429 |
+
"step": 138385
|
| 275430 |
+
},
|
| 275431 |
+
{
|
| 275432 |
+
"epoch": 1107.0,
|
| 275433 |
+
"eval_loss": 0.37267568707466125,
|
| 275434 |
+
"eval_runtime": 36.1145,
|
| 275435 |
+
"eval_samples_per_second": 23.149,
|
| 275436 |
+
"eval_steps_per_second": 0.748,
|
| 275437 |
+
"eval_wer": 0.17828487302171514,
|
| 275438 |
+
"step": 138387
|
| 275439 |
+
},
|
| 275440 |
+
{
|
| 275441 |
+
"epoch": 1107.02,
|
| 275442 |
+
"learning_rate": 7.79911858974359e-06,
|
| 275443 |
+
"loss": 0.3236,
|
| 275444 |
+
"step": 138390
|
| 275445 |
+
},
|
| 275446 |
+
{
|
| 275447 |
+
"epoch": 1107.06,
|
| 275448 |
+
"learning_rate": 7.799038461538461e-06,
|
| 275449 |
+
"loss": 0.2492,
|
| 275450 |
+
"step": 138395
|
| 275451 |
+
},
|
| 275452 |
+
{
|
| 275453 |
+
"epoch": 1107.1,
|
| 275454 |
+
"learning_rate": 7.798958333333334e-06,
|
| 275455 |
+
"loss": 0.2898,
|
| 275456 |
+
"step": 138400
|
| 275457 |
+
},
|
| 275458 |
+
{
|
| 275459 |
+
"epoch": 1107.14,
|
| 275460 |
+
"learning_rate": 7.798878205128206e-06,
|
| 275461 |
+
"loss": 0.3389,
|
| 275462 |
+
"step": 138405
|
| 275463 |
+
},
|
| 275464 |
+
{
|
| 275465 |
+
"epoch": 1107.18,
|
| 275466 |
+
"learning_rate": 7.798798076923077e-06,
|
| 275467 |
+
"loss": 0.7556,
|
| 275468 |
+
"step": 138410
|
| 275469 |
+
},
|
| 275470 |
+
{
|
| 275471 |
+
"epoch": 1107.22,
|
| 275472 |
+
"learning_rate": 7.79871794871795e-06,
|
| 275473 |
+
"loss": 0.8205,
|
| 275474 |
+
"step": 138415
|
| 275475 |
+
},
|
| 275476 |
+
{
|
| 275477 |
+
"epoch": 1107.26,
|
| 275478 |
+
"learning_rate": 7.798637820512821e-06,
|
| 275479 |
+
"loss": 0.2496,
|
| 275480 |
+
"step": 138420
|
| 275481 |
+
},
|
| 275482 |
+
{
|
| 275483 |
+
"epoch": 1107.3,
|
| 275484 |
+
"learning_rate": 7.798557692307693e-06,
|
| 275485 |
+
"loss": 0.2758,
|
| 275486 |
+
"step": 138425
|
| 275487 |
+
},
|
| 275488 |
+
{
|
| 275489 |
+
"epoch": 1107.34,
|
| 275490 |
+
"learning_rate": 7.798477564102564e-06,
|
| 275491 |
+
"loss": 0.3497,
|
| 275492 |
+
"step": 138430
|
| 275493 |
+
},
|
| 275494 |
+
{
|
| 275495 |
+
"epoch": 1107.38,
|
| 275496 |
+
"learning_rate": 7.798397435897437e-06,
|
| 275497 |
+
"loss": 0.6569,
|
| 275498 |
+
"step": 138435
|
| 275499 |
+
},
|
| 275500 |
+
{
|
| 275501 |
+
"epoch": 1107.42,
|
| 275502 |
+
"learning_rate": 7.798317307692309e-06,
|
| 275503 |
+
"loss": 0.7951,
|
| 275504 |
+
"step": 138440
|
| 275505 |
+
},
|
| 275506 |
+
{
|
| 275507 |
+
"epoch": 1107.46,
|
| 275508 |
+
"learning_rate": 7.79823717948718e-06,
|
| 275509 |
+
"loss": 0.3205,
|
| 275510 |
+
"step": 138445
|
| 275511 |
+
},
|
| 275512 |
+
{
|
| 275513 |
+
"epoch": 1107.5,
|
| 275514 |
+
"learning_rate": 7.798157051282051e-06,
|
| 275515 |
+
"loss": 0.278,
|
| 275516 |
+
"step": 138450
|
| 275517 |
+
},
|
| 275518 |
+
{
|
| 275519 |
+
"epoch": 1107.54,
|
| 275520 |
+
"learning_rate": 7.798076923076924e-06,
|
| 275521 |
+
"loss": 0.4093,
|
| 275522 |
+
"step": 138455
|
| 275523 |
+
},
|
| 275524 |
+
{
|
| 275525 |
+
"epoch": 1107.58,
|
| 275526 |
+
"learning_rate": 7.797996794871796e-06,
|
| 275527 |
+
"loss": 0.6617,
|
| 275528 |
+
"step": 138460
|
| 275529 |
+
},
|
| 275530 |
+
{
|
| 275531 |
+
"epoch": 1107.62,
|
| 275532 |
+
"learning_rate": 7.797916666666667e-06,
|
| 275533 |
+
"loss": 0.8341,
|
| 275534 |
+
"step": 138465
|
| 275535 |
+
},
|
| 275536 |
+
{
|
| 275537 |
+
"epoch": 1107.66,
|
| 275538 |
+
"learning_rate": 7.79783653846154e-06,
|
| 275539 |
+
"loss": 0.2369,
|
| 275540 |
+
"step": 138470
|
| 275541 |
+
},
|
| 275542 |
+
{
|
| 275543 |
+
"epoch": 1107.7,
|
| 275544 |
+
"learning_rate": 7.797756410256411e-06,
|
| 275545 |
+
"loss": 0.2439,
|
| 275546 |
+
"step": 138475
|
| 275547 |
+
},
|
| 275548 |
+
{
|
| 275549 |
+
"epoch": 1107.74,
|
| 275550 |
+
"learning_rate": 7.797676282051283e-06,
|
| 275551 |
+
"loss": 0.3456,
|
| 275552 |
+
"step": 138480
|
| 275553 |
+
},
|
| 275554 |
+
{
|
| 275555 |
+
"epoch": 1107.78,
|
| 275556 |
+
"learning_rate": 7.797596153846154e-06,
|
| 275557 |
+
"loss": 0.7806,
|
| 275558 |
+
"step": 138485
|
| 275559 |
+
},
|
| 275560 |
+
{
|
| 275561 |
+
"epoch": 1107.82,
|
| 275562 |
+
"learning_rate": 7.797516025641027e-06,
|
| 275563 |
+
"loss": 0.7638,
|
| 275564 |
+
"step": 138490
|
| 275565 |
+
},
|
| 275566 |
+
{
|
| 275567 |
+
"epoch": 1107.86,
|
| 275568 |
+
"learning_rate": 7.797435897435897e-06,
|
| 275569 |
+
"loss": 0.2938,
|
| 275570 |
+
"step": 138495
|
| 275571 |
+
},
|
| 275572 |
+
{
|
| 275573 |
+
"epoch": 1107.9,
|
| 275574 |
+
"learning_rate": 7.79735576923077e-06,
|
| 275575 |
+
"loss": 0.3411,
|
| 275576 |
+
"step": 138500
|
| 275577 |
+
},
|
| 275578 |
+
{
|
| 275579 |
+
"epoch": 1107.94,
|
| 275580 |
+
"learning_rate": 7.797275641025641e-06,
|
| 275581 |
+
"loss": 0.3161,
|
| 275582 |
+
"step": 138505
|
| 275583 |
+
},
|
| 275584 |
+
{
|
| 275585 |
+
"epoch": 1107.98,
|
| 275586 |
+
"learning_rate": 7.797195512820513e-06,
|
| 275587 |
+
"loss": 0.7295,
|
| 275588 |
+
"step": 138510
|
| 275589 |
+
},
|
| 275590 |
+
{
|
| 275591 |
+
"epoch": 1108.0,
|
| 275592 |
+
"eval_loss": 0.3931129574775696,
|
| 275593 |
+
"eval_runtime": 36.088,
|
| 275594 |
+
"eval_samples_per_second": 23.166,
|
| 275595 |
+
"eval_steps_per_second": 0.748,
|
| 275596 |
+
"eval_wer": 0.18466898954703834,
|
| 275597 |
+
"step": 138512
|
| 275598 |
}
|
| 275599 |
],
|
| 275600 |
"max_steps": 625000,
|
| 275601 |
"num_train_epochs": 5000,
|
| 275602 |
+
"total_flos": 3.897633134684353e+20,
|
| 275603 |
"trial_name": null,
|
| 275604 |
"trial_params": null
|
| 275605 |
}
|
model-bin/finetune/base/{checkpoint-137888 β checkpoint-138512}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1630187518.4754748/events.out.tfevents.1630187518.86bb0ddabf9b.4092.321
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:54dc410a9a3fecaf7a92c972ba99677d0a42240633dfb1c015b59793d60ef9f1
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630187909.429926/events.out.tfevents.1630187909.86bb0ddabf9b.4092.323
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1404d827ccfb8c5e68a6b4990910bc44837b071e0771053f40ca026cda8cc707
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630188491.401316/events.out.tfevents.1630188491.86bb0ddabf9b.4092.325
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3558c3f6c3ed2d6a0f7284da15776df1b48b0d5f8a8c23ee3a0fdbc5f63594ac
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630188882.7400389/events.out.tfevents.1630188882.86bb0ddabf9b.4092.327
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:94ac868c21910420bf191c3c5006f624d3708185f9cf8a4be4ae9080e0739a4f
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630189275.6364982/events.out.tfevents.1630189275.86bb0ddabf9b.4092.329
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5ff3a9f53fc8f497ee8e231391129eb678601aed10d1f46bd2d8124ee727166f
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630187518.86bb0ddabf9b.4092.320
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1b318e619b200054bbd0c93ac0f28337fb8916db0753e444957cef08f84065e8
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630187909.86bb0ddabf9b.4092.322
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4ff1b4874a0347a993a59548c643dc101db9e5c41c1bbd343f380af66f54d006
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630188491.86bb0ddabf9b.4092.324
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:526205bd347a62339a04014e18cb2712fdad26a9998baa3d0a55101ebb2a29a8
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630188882.86bb0ddabf9b.4092.326
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:591bb54395771f7d05a9a290ba8779def3be495405010b31284a4ab646d30318
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630189275.86bb0ddabf9b.4092.328
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3e5fbce309cfc86b1e1cff57bd192069f11466881a4145d14e54e5b001283b6c
|
| 3 |
+
size 8622
|