"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-137263 β checkpoint-137888}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-137263 β checkpoint-137888}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-137263 β checkpoint-137888}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-137263 β checkpoint-137888}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-137263 β checkpoint-137888}/rng_state.pth +2 -2
- model-bin/finetune/base/{checkpoint-137263 β checkpoint-137888}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-137263 β checkpoint-137888}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-137263 β checkpoint-137888}/trainer_state.json +798 -3
- model-bin/finetune/base/{checkpoint-137263 β checkpoint-137888}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630185503.3679285/events.out.tfevents.1630185503.86bb0ddabf9b.4092.311 +3 -0
- model-bin/finetune/base/log/1630185910.66528/events.out.tfevents.1630185910.86bb0ddabf9b.4092.313 +3 -0
- model-bin/finetune/base/log/1630186302.3971124/events.out.tfevents.1630186302.86bb0ddabf9b.4092.315 +3 -0
- model-bin/finetune/base/log/1630186692.7030358/events.out.tfevents.1630186692.86bb0ddabf9b.4092.317 +3 -0
- model-bin/finetune/base/log/1630187079.2727177/events.out.tfevents.1630187079.86bb0ddabf9b.4092.319 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630185503.86bb0ddabf9b.4092.310 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630185910.86bb0ddabf9b.4092.312 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630186302.86bb0ddabf9b.4092.314 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630186692.86bb0ddabf9b.4092.316 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630187079.86bb0ddabf9b.4092.318 +3 -0
model-bin/finetune/base/{checkpoint-137263 β checkpoint-137888}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-137263 β checkpoint-137888}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4a083b1e6ce1522088d8c2f308f3bd9afa4908f9ff8ccbf4ddba99c717a5fa92
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-137263 β checkpoint-137888}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-137263 β checkpoint-137888}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bea9e8f5afb74d2679a4a1d50adebf015eefa70deea4c48290ec98f12b84afd3
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-137263 β checkpoint-137888}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3d42dbb415c82e3775c72ee20b380024ed68e439d62b0305c873bfe20dac54ef
|
| 3 |
+
size 14439
|
model-bin/finetune/base/{checkpoint-137263 β checkpoint-137888}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:266665f53539a177c01e29ab36757780591034e2c9aadb0520efb947194c5ceb
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-137263 β checkpoint-137888}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:53b6f7ab2a0a36127d4275c2473bda90560022891b15f8d6c9d8fef41711fc6f
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-137263 β checkpoint-137888}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1689111747851003,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-132910",
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -274005,11 +274005,806 @@
|
|
| 274005 |
"eval_steps_per_second": 0.732,
|
| 274006 |
"eval_wer": 0.1787237226004339,
|
| 274007 |
"step": 137263
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 274008 |
}
|
| 274009 |
],
|
| 274010 |
"max_steps": 625000,
|
| 274011 |
"num_train_epochs": 5000,
|
| 274012 |
-
"total_flos": 3.
|
| 274013 |
"trial_name": null,
|
| 274014 |
"trial_params": null
|
| 274015 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1689111747851003,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-132910",
|
| 4 |
+
"epoch": 1103.0,
|
| 5 |
+
"global_step": 137888,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 274005 |
"eval_steps_per_second": 0.732,
|
| 274006 |
"eval_wer": 0.1787237226004339,
|
| 274007 |
"step": 137263
|
| 274008 |
+
},
|
| 274009 |
+
{
|
| 274010 |
+
"epoch": 1098.02,
|
| 274011 |
+
"learning_rate": 7.817115384615386e-06,
|
| 274012 |
+
"loss": 0.3239,
|
| 274013 |
+
"step": 137265
|
| 274014 |
+
},
|
| 274015 |
+
{
|
| 274016 |
+
"epoch": 1098.06,
|
| 274017 |
+
"learning_rate": 7.817035256410257e-06,
|
| 274018 |
+
"loss": 0.3262,
|
| 274019 |
+
"step": 137270
|
| 274020 |
+
},
|
| 274021 |
+
{
|
| 274022 |
+
"epoch": 1098.1,
|
| 274023 |
+
"learning_rate": 7.81695512820513e-06,
|
| 274024 |
+
"loss": 0.2122,
|
| 274025 |
+
"step": 137275
|
| 274026 |
+
},
|
| 274027 |
+
{
|
| 274028 |
+
"epoch": 1098.14,
|
| 274029 |
+
"learning_rate": 7.816875e-06,
|
| 274030 |
+
"loss": 0.2913,
|
| 274031 |
+
"step": 137280
|
| 274032 |
+
},
|
| 274033 |
+
{
|
| 274034 |
+
"epoch": 1098.18,
|
| 274035 |
+
"learning_rate": 7.816794871794873e-06,
|
| 274036 |
+
"loss": 0.6546,
|
| 274037 |
+
"step": 137285
|
| 274038 |
+
},
|
| 274039 |
+
{
|
| 274040 |
+
"epoch": 1098.22,
|
| 274041 |
+
"learning_rate": 7.816714743589744e-06,
|
| 274042 |
+
"loss": 0.9316,
|
| 274043 |
+
"step": 137290
|
| 274044 |
+
},
|
| 274045 |
+
{
|
| 274046 |
+
"epoch": 1098.26,
|
| 274047 |
+
"learning_rate": 7.816634615384617e-06,
|
| 274048 |
+
"loss": 0.2835,
|
| 274049 |
+
"step": 137295
|
| 274050 |
+
},
|
| 274051 |
+
{
|
| 274052 |
+
"epoch": 1098.3,
|
| 274053 |
+
"learning_rate": 7.816554487179487e-06,
|
| 274054 |
+
"loss": 0.2565,
|
| 274055 |
+
"step": 137300
|
| 274056 |
+
},
|
| 274057 |
+
{
|
| 274058 |
+
"epoch": 1098.34,
|
| 274059 |
+
"learning_rate": 7.81647435897436e-06,
|
| 274060 |
+
"loss": 0.3442,
|
| 274061 |
+
"step": 137305
|
| 274062 |
+
},
|
| 274063 |
+
{
|
| 274064 |
+
"epoch": 1098.38,
|
| 274065 |
+
"learning_rate": 7.816394230769231e-06,
|
| 274066 |
+
"loss": 0.5282,
|
| 274067 |
+
"step": 137310
|
| 274068 |
+
},
|
| 274069 |
+
{
|
| 274070 |
+
"epoch": 1098.42,
|
| 274071 |
+
"learning_rate": 7.816314102564102e-06,
|
| 274072 |
+
"loss": 0.9422,
|
| 274073 |
+
"step": 137315
|
| 274074 |
+
},
|
| 274075 |
+
{
|
| 274076 |
+
"epoch": 1098.46,
|
| 274077 |
+
"learning_rate": 7.816233974358976e-06,
|
| 274078 |
+
"loss": 0.2689,
|
| 274079 |
+
"step": 137320
|
| 274080 |
+
},
|
| 274081 |
+
{
|
| 274082 |
+
"epoch": 1098.5,
|
| 274083 |
+
"learning_rate": 7.816153846153847e-06,
|
| 274084 |
+
"loss": 0.2509,
|
| 274085 |
+
"step": 137325
|
| 274086 |
+
},
|
| 274087 |
+
{
|
| 274088 |
+
"epoch": 1098.54,
|
| 274089 |
+
"learning_rate": 7.816073717948718e-06,
|
| 274090 |
+
"loss": 0.2991,
|
| 274091 |
+
"step": 137330
|
| 274092 |
+
},
|
| 274093 |
+
{
|
| 274094 |
+
"epoch": 1098.58,
|
| 274095 |
+
"learning_rate": 7.81599358974359e-06,
|
| 274096 |
+
"loss": 0.715,
|
| 274097 |
+
"step": 137335
|
| 274098 |
+
},
|
| 274099 |
+
{
|
| 274100 |
+
"epoch": 1098.62,
|
| 274101 |
+
"learning_rate": 7.815913461538463e-06,
|
| 274102 |
+
"loss": 1.002,
|
| 274103 |
+
"step": 137340
|
| 274104 |
+
},
|
| 274105 |
+
{
|
| 274106 |
+
"epoch": 1098.66,
|
| 274107 |
+
"learning_rate": 7.815833333333334e-06,
|
| 274108 |
+
"loss": 0.2357,
|
| 274109 |
+
"step": 137345
|
| 274110 |
+
},
|
| 274111 |
+
{
|
| 274112 |
+
"epoch": 1098.7,
|
| 274113 |
+
"learning_rate": 7.815753205128205e-06,
|
| 274114 |
+
"loss": 0.2285,
|
| 274115 |
+
"step": 137350
|
| 274116 |
+
},
|
| 274117 |
+
{
|
| 274118 |
+
"epoch": 1098.74,
|
| 274119 |
+
"learning_rate": 7.815673076923077e-06,
|
| 274120 |
+
"loss": 0.288,
|
| 274121 |
+
"step": 137355
|
| 274122 |
+
},
|
| 274123 |
+
{
|
| 274124 |
+
"epoch": 1098.78,
|
| 274125 |
+
"learning_rate": 7.81559294871795e-06,
|
| 274126 |
+
"loss": 0.5729,
|
| 274127 |
+
"step": 137360
|
| 274128 |
+
},
|
| 274129 |
+
{
|
| 274130 |
+
"epoch": 1098.82,
|
| 274131 |
+
"learning_rate": 7.815512820512821e-06,
|
| 274132 |
+
"loss": 0.9696,
|
| 274133 |
+
"step": 137365
|
| 274134 |
+
},
|
| 274135 |
+
{
|
| 274136 |
+
"epoch": 1098.86,
|
| 274137 |
+
"learning_rate": 7.815432692307693e-06,
|
| 274138 |
+
"loss": 0.2812,
|
| 274139 |
+
"step": 137370
|
| 274140 |
+
},
|
| 274141 |
+
{
|
| 274142 |
+
"epoch": 1098.9,
|
| 274143 |
+
"learning_rate": 7.815352564102566e-06,
|
| 274144 |
+
"loss": 0.2379,
|
| 274145 |
+
"step": 137375
|
| 274146 |
+
},
|
| 274147 |
+
{
|
| 274148 |
+
"epoch": 1098.94,
|
| 274149 |
+
"learning_rate": 7.815272435897437e-06,
|
| 274150 |
+
"loss": 0.3379,
|
| 274151 |
+
"step": 137380
|
| 274152 |
+
},
|
| 274153 |
+
{
|
| 274154 |
+
"epoch": 1098.98,
|
| 274155 |
+
"learning_rate": 7.815192307692308e-06,
|
| 274156 |
+
"loss": 0.6368,
|
| 274157 |
+
"step": 137385
|
| 274158 |
+
},
|
| 274159 |
+
{
|
| 274160 |
+
"epoch": 1099.0,
|
| 274161 |
+
"eval_loss": 0.343011736869812,
|
| 274162 |
+
"eval_runtime": 37.5693,
|
| 274163 |
+
"eval_samples_per_second": 22.439,
|
| 274164 |
+
"eval_steps_per_second": 0.719,
|
| 274165 |
+
"eval_wer": 0.1803230543318649,
|
| 274166 |
+
"step": 137388
|
| 274167 |
+
},
|
| 274168 |
+
{
|
| 274169 |
+
"epoch": 1099.02,
|
| 274170 |
+
"learning_rate": 7.81511217948718e-06,
|
| 274171 |
+
"loss": 0.3496,
|
| 274172 |
+
"step": 137390
|
| 274173 |
+
},
|
| 274174 |
+
{
|
| 274175 |
+
"epoch": 1099.06,
|
| 274176 |
+
"learning_rate": 7.815032051282053e-06,
|
| 274177 |
+
"loss": 0.2633,
|
| 274178 |
+
"step": 137395
|
| 274179 |
+
},
|
| 274180 |
+
{
|
| 274181 |
+
"epoch": 1099.1,
|
| 274182 |
+
"learning_rate": 7.814951923076924e-06,
|
| 274183 |
+
"loss": 0.2516,
|
| 274184 |
+
"step": 137400
|
| 274185 |
+
},
|
| 274186 |
+
{
|
| 274187 |
+
"epoch": 1099.14,
|
| 274188 |
+
"learning_rate": 7.814871794871795e-06,
|
| 274189 |
+
"loss": 0.3469,
|
| 274190 |
+
"step": 137405
|
| 274191 |
+
},
|
| 274192 |
+
{
|
| 274193 |
+
"epoch": 1099.18,
|
| 274194 |
+
"learning_rate": 7.814791666666667e-06,
|
| 274195 |
+
"loss": 0.7097,
|
| 274196 |
+
"step": 137410
|
| 274197 |
+
},
|
| 274198 |
+
{
|
| 274199 |
+
"epoch": 1099.22,
|
| 274200 |
+
"learning_rate": 7.81471153846154e-06,
|
| 274201 |
+
"loss": 1.0595,
|
| 274202 |
+
"step": 137415
|
| 274203 |
+
},
|
| 274204 |
+
{
|
| 274205 |
+
"epoch": 1099.26,
|
| 274206 |
+
"learning_rate": 7.814631410256411e-06,
|
| 274207 |
+
"loss": 0.2628,
|
| 274208 |
+
"step": 137420
|
| 274209 |
+
},
|
| 274210 |
+
{
|
| 274211 |
+
"epoch": 1099.3,
|
| 274212 |
+
"learning_rate": 7.814551282051283e-06,
|
| 274213 |
+
"loss": 0.2798,
|
| 274214 |
+
"step": 137425
|
| 274215 |
+
},
|
| 274216 |
+
{
|
| 274217 |
+
"epoch": 1099.34,
|
| 274218 |
+
"learning_rate": 7.814471153846156e-06,
|
| 274219 |
+
"loss": 0.334,
|
| 274220 |
+
"step": 137430
|
| 274221 |
+
},
|
| 274222 |
+
{
|
| 274223 |
+
"epoch": 1099.38,
|
| 274224 |
+
"learning_rate": 7.814391025641025e-06,
|
| 274225 |
+
"loss": 0.605,
|
| 274226 |
+
"step": 137435
|
| 274227 |
+
},
|
| 274228 |
+
{
|
| 274229 |
+
"epoch": 1099.42,
|
| 274230 |
+
"learning_rate": 7.814310897435898e-06,
|
| 274231 |
+
"loss": 1.0427,
|
| 274232 |
+
"step": 137440
|
| 274233 |
+
},
|
| 274234 |
+
{
|
| 274235 |
+
"epoch": 1099.46,
|
| 274236 |
+
"learning_rate": 7.81423076923077e-06,
|
| 274237 |
+
"loss": 0.3009,
|
| 274238 |
+
"step": 137445
|
| 274239 |
+
},
|
| 274240 |
+
{
|
| 274241 |
+
"epoch": 1099.5,
|
| 274242 |
+
"learning_rate": 7.814150641025641e-06,
|
| 274243 |
+
"loss": 0.268,
|
| 274244 |
+
"step": 137450
|
| 274245 |
+
},
|
| 274246 |
+
{
|
| 274247 |
+
"epoch": 1099.54,
|
| 274248 |
+
"learning_rate": 7.814070512820512e-06,
|
| 274249 |
+
"loss": 0.3541,
|
| 274250 |
+
"step": 137455
|
| 274251 |
+
},
|
| 274252 |
+
{
|
| 274253 |
+
"epoch": 1099.58,
|
| 274254 |
+
"learning_rate": 7.813990384615385e-06,
|
| 274255 |
+
"loss": 0.5783,
|
| 274256 |
+
"step": 137460
|
| 274257 |
+
},
|
| 274258 |
+
{
|
| 274259 |
+
"epoch": 1099.62,
|
| 274260 |
+
"learning_rate": 7.813910256410257e-06,
|
| 274261 |
+
"loss": 1.0424,
|
| 274262 |
+
"step": 137465
|
| 274263 |
+
},
|
| 274264 |
+
{
|
| 274265 |
+
"epoch": 1099.66,
|
| 274266 |
+
"learning_rate": 7.813830128205128e-06,
|
| 274267 |
+
"loss": 0.3697,
|
| 274268 |
+
"step": 137470
|
| 274269 |
+
},
|
| 274270 |
+
{
|
| 274271 |
+
"epoch": 1099.7,
|
| 274272 |
+
"learning_rate": 7.813750000000001e-06,
|
| 274273 |
+
"loss": 0.2716,
|
| 274274 |
+
"step": 137475
|
| 274275 |
+
},
|
| 274276 |
+
{
|
| 274277 |
+
"epoch": 1099.74,
|
| 274278 |
+
"learning_rate": 7.813669871794873e-06,
|
| 274279 |
+
"loss": 0.3446,
|
| 274280 |
+
"step": 137480
|
| 274281 |
+
},
|
| 274282 |
+
{
|
| 274283 |
+
"epoch": 1099.78,
|
| 274284 |
+
"learning_rate": 7.813589743589744e-06,
|
| 274285 |
+
"loss": 0.5443,
|
| 274286 |
+
"step": 137485
|
| 274287 |
+
},
|
| 274288 |
+
{
|
| 274289 |
+
"epoch": 1099.82,
|
| 274290 |
+
"learning_rate": 7.813509615384615e-06,
|
| 274291 |
+
"loss": 0.922,
|
| 274292 |
+
"step": 137490
|
| 274293 |
+
},
|
| 274294 |
+
{
|
| 274295 |
+
"epoch": 1099.86,
|
| 274296 |
+
"learning_rate": 7.813429487179488e-06,
|
| 274297 |
+
"loss": 0.2658,
|
| 274298 |
+
"step": 137495
|
| 274299 |
+
},
|
| 274300 |
+
{
|
| 274301 |
+
"epoch": 1099.9,
|
| 274302 |
+
"learning_rate": 7.81334935897436e-06,
|
| 274303 |
+
"loss": 0.2983,
|
| 274304 |
+
"step": 137500
|
| 274305 |
+
},
|
| 274306 |
+
{
|
| 274307 |
+
"epoch": 1099.94,
|
| 274308 |
+
"learning_rate": 7.813269230769231e-06,
|
| 274309 |
+
"loss": 0.3811,
|
| 274310 |
+
"step": 137505
|
| 274311 |
+
},
|
| 274312 |
+
{
|
| 274313 |
+
"epoch": 1099.98,
|
| 274314 |
+
"learning_rate": 7.813189102564102e-06,
|
| 274315 |
+
"loss": 0.653,
|
| 274316 |
+
"step": 137510
|
| 274317 |
+
},
|
| 274318 |
+
{
|
| 274319 |
+
"epoch": 1100.0,
|
| 274320 |
+
"eval_loss": 0.3978674113750458,
|
| 274321 |
+
"eval_runtime": 36.6864,
|
| 274322 |
+
"eval_samples_per_second": 22.979,
|
| 274323 |
+
"eval_steps_per_second": 0.736,
|
| 274324 |
+
"eval_wer": 0.17095375722543352,
|
| 274325 |
+
"step": 137513
|
| 274326 |
+
},
|
| 274327 |
+
{
|
| 274328 |
+
"epoch": 1100.02,
|
| 274329 |
+
"learning_rate": 7.813108974358975e-06,
|
| 274330 |
+
"loss": 0.3937,
|
| 274331 |
+
"step": 137515
|
| 274332 |
+
},
|
| 274333 |
+
{
|
| 274334 |
+
"epoch": 1100.06,
|
| 274335 |
+
"learning_rate": 7.813028846153847e-06,
|
| 274336 |
+
"loss": 0.3149,
|
| 274337 |
+
"step": 137520
|
| 274338 |
+
},
|
| 274339 |
+
{
|
| 274340 |
+
"epoch": 1100.1,
|
| 274341 |
+
"learning_rate": 7.812948717948718e-06,
|
| 274342 |
+
"loss": 0.287,
|
| 274343 |
+
"step": 137525
|
| 274344 |
+
},
|
| 274345 |
+
{
|
| 274346 |
+
"epoch": 1100.14,
|
| 274347 |
+
"learning_rate": 7.812868589743591e-06,
|
| 274348 |
+
"loss": 0.2861,
|
| 274349 |
+
"step": 137530
|
| 274350 |
+
},
|
| 274351 |
+
{
|
| 274352 |
+
"epoch": 1100.18,
|
| 274353 |
+
"learning_rate": 7.812788461538463e-06,
|
| 274354 |
+
"loss": 0.56,
|
| 274355 |
+
"step": 137535
|
| 274356 |
+
},
|
| 274357 |
+
{
|
| 274358 |
+
"epoch": 1100.22,
|
| 274359 |
+
"learning_rate": 7.812708333333334e-06,
|
| 274360 |
+
"loss": 1.0073,
|
| 274361 |
+
"step": 137540
|
| 274362 |
+
},
|
| 274363 |
+
{
|
| 274364 |
+
"epoch": 1100.26,
|
| 274365 |
+
"learning_rate": 7.812628205128205e-06,
|
| 274366 |
+
"loss": 0.4221,
|
| 274367 |
+
"step": 137545
|
| 274368 |
+
},
|
| 274369 |
+
{
|
| 274370 |
+
"epoch": 1100.3,
|
| 274371 |
+
"learning_rate": 7.812548076923078e-06,
|
| 274372 |
+
"loss": 0.3091,
|
| 274373 |
+
"step": 137550
|
| 274374 |
+
},
|
| 274375 |
+
{
|
| 274376 |
+
"epoch": 1100.34,
|
| 274377 |
+
"learning_rate": 7.81246794871795e-06,
|
| 274378 |
+
"loss": 0.3689,
|
| 274379 |
+
"step": 137555
|
| 274380 |
+
},
|
| 274381 |
+
{
|
| 274382 |
+
"epoch": 1100.38,
|
| 274383 |
+
"learning_rate": 7.812387820512821e-06,
|
| 274384 |
+
"loss": 0.5658,
|
| 274385 |
+
"step": 137560
|
| 274386 |
+
},
|
| 274387 |
+
{
|
| 274388 |
+
"epoch": 1100.42,
|
| 274389 |
+
"learning_rate": 7.812307692307694e-06,
|
| 274390 |
+
"loss": 0.9342,
|
| 274391 |
+
"step": 137565
|
| 274392 |
+
},
|
| 274393 |
+
{
|
| 274394 |
+
"epoch": 1100.46,
|
| 274395 |
+
"learning_rate": 7.812227564102565e-06,
|
| 274396 |
+
"loss": 0.2346,
|
| 274397 |
+
"step": 137570
|
| 274398 |
+
},
|
| 274399 |
+
{
|
| 274400 |
+
"epoch": 1100.5,
|
| 274401 |
+
"learning_rate": 7.812147435897437e-06,
|
| 274402 |
+
"loss": 0.3372,
|
| 274403 |
+
"step": 137575
|
| 274404 |
+
},
|
| 274405 |
+
{
|
| 274406 |
+
"epoch": 1100.54,
|
| 274407 |
+
"learning_rate": 7.812067307692308e-06,
|
| 274408 |
+
"loss": 0.3696,
|
| 274409 |
+
"step": 137580
|
| 274410 |
+
},
|
| 274411 |
+
{
|
| 274412 |
+
"epoch": 1100.58,
|
| 274413 |
+
"learning_rate": 7.811987179487181e-06,
|
| 274414 |
+
"loss": 0.5708,
|
| 274415 |
+
"step": 137585
|
| 274416 |
+
},
|
| 274417 |
+
{
|
| 274418 |
+
"epoch": 1100.62,
|
| 274419 |
+
"learning_rate": 7.811907051282051e-06,
|
| 274420 |
+
"loss": 1.0029,
|
| 274421 |
+
"step": 137590
|
| 274422 |
+
},
|
| 274423 |
+
{
|
| 274424 |
+
"epoch": 1100.66,
|
| 274425 |
+
"learning_rate": 7.811826923076924e-06,
|
| 274426 |
+
"loss": 0.2572,
|
| 274427 |
+
"step": 137595
|
| 274428 |
+
},
|
| 274429 |
+
{
|
| 274430 |
+
"epoch": 1100.7,
|
| 274431 |
+
"learning_rate": 7.811746794871795e-06,
|
| 274432 |
+
"loss": 0.2712,
|
| 274433 |
+
"step": 137600
|
| 274434 |
+
},
|
| 274435 |
+
{
|
| 274436 |
+
"epoch": 1100.74,
|
| 274437 |
+
"learning_rate": 7.811666666666667e-06,
|
| 274438 |
+
"loss": 0.3289,
|
| 274439 |
+
"step": 137605
|
| 274440 |
+
},
|
| 274441 |
+
{
|
| 274442 |
+
"epoch": 1100.78,
|
| 274443 |
+
"learning_rate": 7.811586538461538e-06,
|
| 274444 |
+
"loss": 0.636,
|
| 274445 |
+
"step": 137610
|
| 274446 |
+
},
|
| 274447 |
+
{
|
| 274448 |
+
"epoch": 1100.82,
|
| 274449 |
+
"learning_rate": 7.811506410256411e-06,
|
| 274450 |
+
"loss": 0.9177,
|
| 274451 |
+
"step": 137615
|
| 274452 |
+
},
|
| 274453 |
+
{
|
| 274454 |
+
"epoch": 1100.86,
|
| 274455 |
+
"learning_rate": 7.811426282051282e-06,
|
| 274456 |
+
"loss": 0.2958,
|
| 274457 |
+
"step": 137620
|
| 274458 |
+
},
|
| 274459 |
+
{
|
| 274460 |
+
"epoch": 1100.9,
|
| 274461 |
+
"learning_rate": 7.811346153846154e-06,
|
| 274462 |
+
"loss": 0.2554,
|
| 274463 |
+
"step": 137625
|
| 274464 |
+
},
|
| 274465 |
+
{
|
| 274466 |
+
"epoch": 1100.94,
|
| 274467 |
+
"learning_rate": 7.811266025641027e-06,
|
| 274468 |
+
"loss": 0.3038,
|
| 274469 |
+
"step": 137630
|
| 274470 |
+
},
|
| 274471 |
+
{
|
| 274472 |
+
"epoch": 1100.98,
|
| 274473 |
+
"learning_rate": 7.811185897435898e-06,
|
| 274474 |
+
"loss": 0.6366,
|
| 274475 |
+
"step": 137635
|
| 274476 |
+
},
|
| 274477 |
+
{
|
| 274478 |
+
"epoch": 1101.0,
|
| 274479 |
+
"eval_loss": 0.3657839000225067,
|
| 274480 |
+
"eval_runtime": 35.612,
|
| 274481 |
+
"eval_samples_per_second": 23.672,
|
| 274482 |
+
"eval_steps_per_second": 0.758,
|
| 274483 |
+
"eval_wer": 0.176545332237088,
|
| 274484 |
+
"step": 137638
|
| 274485 |
+
},
|
| 274486 |
+
{
|
| 274487 |
+
"epoch": 1101.02,
|
| 274488 |
+
"learning_rate": 7.81110576923077e-06,
|
| 274489 |
+
"loss": 0.3074,
|
| 274490 |
+
"step": 137640
|
| 274491 |
+
},
|
| 274492 |
+
{
|
| 274493 |
+
"epoch": 1101.06,
|
| 274494 |
+
"learning_rate": 7.811025641025641e-06,
|
| 274495 |
+
"loss": 0.2618,
|
| 274496 |
+
"step": 137645
|
| 274497 |
+
},
|
| 274498 |
+
{
|
| 274499 |
+
"epoch": 1101.1,
|
| 274500 |
+
"learning_rate": 7.810945512820514e-06,
|
| 274501 |
+
"loss": 0.3976,
|
| 274502 |
+
"step": 137650
|
| 274503 |
+
},
|
| 274504 |
+
{
|
| 274505 |
+
"epoch": 1101.14,
|
| 274506 |
+
"learning_rate": 7.810865384615385e-06,
|
| 274507 |
+
"loss": 0.3164,
|
| 274508 |
+
"step": 137655
|
| 274509 |
+
},
|
| 274510 |
+
{
|
| 274511 |
+
"epoch": 1101.18,
|
| 274512 |
+
"learning_rate": 7.810785256410257e-06,
|
| 274513 |
+
"loss": 0.6068,
|
| 274514 |
+
"step": 137660
|
| 274515 |
+
},
|
| 274516 |
+
{
|
| 274517 |
+
"epoch": 1101.22,
|
| 274518 |
+
"learning_rate": 7.81070512820513e-06,
|
| 274519 |
+
"loss": 0.855,
|
| 274520 |
+
"step": 137665
|
| 274521 |
+
},
|
| 274522 |
+
{
|
| 274523 |
+
"epoch": 1101.26,
|
| 274524 |
+
"learning_rate": 7.810625000000001e-06,
|
| 274525 |
+
"loss": 0.2925,
|
| 274526 |
+
"step": 137670
|
| 274527 |
+
},
|
| 274528 |
+
{
|
| 274529 |
+
"epoch": 1101.3,
|
| 274530 |
+
"learning_rate": 7.810544871794872e-06,
|
| 274531 |
+
"loss": 0.2918,
|
| 274532 |
+
"step": 137675
|
| 274533 |
+
},
|
| 274534 |
+
{
|
| 274535 |
+
"epoch": 1101.34,
|
| 274536 |
+
"learning_rate": 7.810464743589744e-06,
|
| 274537 |
+
"loss": 0.3933,
|
| 274538 |
+
"step": 137680
|
| 274539 |
+
},
|
| 274540 |
+
{
|
| 274541 |
+
"epoch": 1101.38,
|
| 274542 |
+
"learning_rate": 7.810384615384617e-06,
|
| 274543 |
+
"loss": 0.6153,
|
| 274544 |
+
"step": 137685
|
| 274545 |
+
},
|
| 274546 |
+
{
|
| 274547 |
+
"epoch": 1101.42,
|
| 274548 |
+
"learning_rate": 7.810304487179488e-06,
|
| 274549 |
+
"loss": 1.062,
|
| 274550 |
+
"step": 137690
|
| 274551 |
+
},
|
| 274552 |
+
{
|
| 274553 |
+
"epoch": 1101.46,
|
| 274554 |
+
"learning_rate": 7.81022435897436e-06,
|
| 274555 |
+
"loss": 0.2541,
|
| 274556 |
+
"step": 137695
|
| 274557 |
+
},
|
| 274558 |
+
{
|
| 274559 |
+
"epoch": 1101.5,
|
| 274560 |
+
"learning_rate": 7.810144230769231e-06,
|
| 274561 |
+
"loss": 0.3637,
|
| 274562 |
+
"step": 137700
|
| 274563 |
+
},
|
| 274564 |
+
{
|
| 274565 |
+
"epoch": 1101.54,
|
| 274566 |
+
"learning_rate": 7.810064102564104e-06,
|
| 274567 |
+
"loss": 0.3795,
|
| 274568 |
+
"step": 137705
|
| 274569 |
+
},
|
| 274570 |
+
{
|
| 274571 |
+
"epoch": 1101.58,
|
| 274572 |
+
"learning_rate": 7.809983974358974e-06,
|
| 274573 |
+
"loss": 0.5855,
|
| 274574 |
+
"step": 137710
|
| 274575 |
+
},
|
| 274576 |
+
{
|
| 274577 |
+
"epoch": 1101.62,
|
| 274578 |
+
"learning_rate": 7.809903846153847e-06,
|
| 274579 |
+
"loss": 0.9056,
|
| 274580 |
+
"step": 137715
|
| 274581 |
+
},
|
| 274582 |
+
{
|
| 274583 |
+
"epoch": 1101.66,
|
| 274584 |
+
"learning_rate": 7.80982371794872e-06,
|
| 274585 |
+
"loss": 0.264,
|
| 274586 |
+
"step": 137720
|
| 274587 |
+
},
|
| 274588 |
+
{
|
| 274589 |
+
"epoch": 1101.7,
|
| 274590 |
+
"learning_rate": 7.80974358974359e-06,
|
| 274591 |
+
"loss": 0.2906,
|
| 274592 |
+
"step": 137725
|
| 274593 |
+
},
|
| 274594 |
+
{
|
| 274595 |
+
"epoch": 1101.74,
|
| 274596 |
+
"learning_rate": 7.809663461538463e-06,
|
| 274597 |
+
"loss": 0.3041,
|
| 274598 |
+
"step": 137730
|
| 274599 |
+
},
|
| 274600 |
+
{
|
| 274601 |
+
"epoch": 1101.78,
|
| 274602 |
+
"learning_rate": 7.809583333333334e-06,
|
| 274603 |
+
"loss": 0.5994,
|
| 274604 |
+
"step": 137735
|
| 274605 |
+
},
|
| 274606 |
+
{
|
| 274607 |
+
"epoch": 1101.82,
|
| 274608 |
+
"learning_rate": 7.809503205128205e-06,
|
| 274609 |
+
"loss": 0.9407,
|
| 274610 |
+
"step": 137740
|
| 274611 |
+
},
|
| 274612 |
+
{
|
| 274613 |
+
"epoch": 1101.86,
|
| 274614 |
+
"learning_rate": 7.809423076923077e-06,
|
| 274615 |
+
"loss": 0.3137,
|
| 274616 |
+
"step": 137745
|
| 274617 |
+
},
|
| 274618 |
+
{
|
| 274619 |
+
"epoch": 1101.9,
|
| 274620 |
+
"learning_rate": 7.80934294871795e-06,
|
| 274621 |
+
"loss": 0.2753,
|
| 274622 |
+
"step": 137750
|
| 274623 |
+
},
|
| 274624 |
+
{
|
| 274625 |
+
"epoch": 1101.94,
|
| 274626 |
+
"learning_rate": 7.809262820512821e-06,
|
| 274627 |
+
"loss": 0.3856,
|
| 274628 |
+
"step": 137755
|
| 274629 |
+
},
|
| 274630 |
+
{
|
| 274631 |
+
"epoch": 1101.98,
|
| 274632 |
+
"learning_rate": 7.809182692307692e-06,
|
| 274633 |
+
"loss": 0.5517,
|
| 274634 |
+
"step": 137760
|
| 274635 |
+
},
|
| 274636 |
+
{
|
| 274637 |
+
"epoch": 1102.0,
|
| 274638 |
+
"eval_loss": 0.5093977451324463,
|
| 274639 |
+
"eval_runtime": 37.1907,
|
| 274640 |
+
"eval_samples_per_second": 22.667,
|
| 274641 |
+
"eval_steps_per_second": 0.726,
|
| 274642 |
+
"eval_wer": 0.1790757381258023,
|
| 274643 |
+
"step": 137763
|
| 274644 |
+
},
|
| 274645 |
+
{
|
| 274646 |
+
"epoch": 1102.02,
|
| 274647 |
+
"learning_rate": 7.809102564102565e-06,
|
| 274648 |
+
"loss": 0.3526,
|
| 274649 |
+
"step": 137765
|
| 274650 |
+
},
|
| 274651 |
+
{
|
| 274652 |
+
"epoch": 1102.06,
|
| 274653 |
+
"learning_rate": 7.809022435897437e-06,
|
| 274654 |
+
"loss": 0.2889,
|
| 274655 |
+
"step": 137770
|
| 274656 |
+
},
|
| 274657 |
+
{
|
| 274658 |
+
"epoch": 1102.1,
|
| 274659 |
+
"learning_rate": 7.808942307692308e-06,
|
| 274660 |
+
"loss": 0.2988,
|
| 274661 |
+
"step": 137775
|
| 274662 |
+
},
|
| 274663 |
+
{
|
| 274664 |
+
"epoch": 1102.14,
|
| 274665 |
+
"learning_rate": 7.80886217948718e-06,
|
| 274666 |
+
"loss": 0.2799,
|
| 274667 |
+
"step": 137780
|
| 274668 |
+
},
|
| 274669 |
+
{
|
| 274670 |
+
"epoch": 1102.18,
|
| 274671 |
+
"learning_rate": 7.808782051282053e-06,
|
| 274672 |
+
"loss": 0.5894,
|
| 274673 |
+
"step": 137785
|
| 274674 |
+
},
|
| 274675 |
+
{
|
| 274676 |
+
"epoch": 1102.22,
|
| 274677 |
+
"learning_rate": 7.808701923076924e-06,
|
| 274678 |
+
"loss": 0.8977,
|
| 274679 |
+
"step": 137790
|
| 274680 |
+
},
|
| 274681 |
+
{
|
| 274682 |
+
"epoch": 1102.26,
|
| 274683 |
+
"learning_rate": 7.808621794871795e-06,
|
| 274684 |
+
"loss": 0.2808,
|
| 274685 |
+
"step": 137795
|
| 274686 |
+
},
|
| 274687 |
+
{
|
| 274688 |
+
"epoch": 1102.3,
|
| 274689 |
+
"learning_rate": 7.808541666666667e-06,
|
| 274690 |
+
"loss": 0.2868,
|
| 274691 |
+
"step": 137800
|
| 274692 |
+
},
|
| 274693 |
+
{
|
| 274694 |
+
"epoch": 1102.34,
|
| 274695 |
+
"learning_rate": 7.80846153846154e-06,
|
| 274696 |
+
"loss": 0.3436,
|
| 274697 |
+
"step": 137805
|
| 274698 |
+
},
|
| 274699 |
+
{
|
| 274700 |
+
"epoch": 1102.38,
|
| 274701 |
+
"learning_rate": 7.808381410256411e-06,
|
| 274702 |
+
"loss": 0.4889,
|
| 274703 |
+
"step": 137810
|
| 274704 |
+
},
|
| 274705 |
+
{
|
| 274706 |
+
"epoch": 1102.42,
|
| 274707 |
+
"learning_rate": 7.808301282051282e-06,
|
| 274708 |
+
"loss": 0.9196,
|
| 274709 |
+
"step": 137815
|
| 274710 |
+
},
|
| 274711 |
+
{
|
| 274712 |
+
"epoch": 1102.46,
|
| 274713 |
+
"learning_rate": 7.808221153846155e-06,
|
| 274714 |
+
"loss": 0.2466,
|
| 274715 |
+
"step": 137820
|
| 274716 |
+
},
|
| 274717 |
+
{
|
| 274718 |
+
"epoch": 1102.5,
|
| 274719 |
+
"learning_rate": 7.808141025641027e-06,
|
| 274720 |
+
"loss": 0.3175,
|
| 274721 |
+
"step": 137825
|
| 274722 |
+
},
|
| 274723 |
+
{
|
| 274724 |
+
"epoch": 1102.54,
|
| 274725 |
+
"learning_rate": 7.808060897435898e-06,
|
| 274726 |
+
"loss": 0.3459,
|
| 274727 |
+
"step": 137830
|
| 274728 |
+
},
|
| 274729 |
+
{
|
| 274730 |
+
"epoch": 1102.58,
|
| 274731 |
+
"learning_rate": 7.80798076923077e-06,
|
| 274732 |
+
"loss": 0.6213,
|
| 274733 |
+
"step": 137835
|
| 274734 |
+
},
|
| 274735 |
+
{
|
| 274736 |
+
"epoch": 1102.62,
|
| 274737 |
+
"learning_rate": 7.807900641025643e-06,
|
| 274738 |
+
"loss": 1.0231,
|
| 274739 |
+
"step": 137840
|
| 274740 |
+
},
|
| 274741 |
+
{
|
| 274742 |
+
"epoch": 1102.66,
|
| 274743 |
+
"learning_rate": 7.807820512820514e-06,
|
| 274744 |
+
"loss": 0.2693,
|
| 274745 |
+
"step": 137845
|
| 274746 |
+
},
|
| 274747 |
+
{
|
| 274748 |
+
"epoch": 1102.7,
|
| 274749 |
+
"learning_rate": 7.807740384615385e-06,
|
| 274750 |
+
"loss": 0.3046,
|
| 274751 |
+
"step": 137850
|
| 274752 |
+
},
|
| 274753 |
+
{
|
| 274754 |
+
"epoch": 1102.74,
|
| 274755 |
+
"learning_rate": 7.807660256410257e-06,
|
| 274756 |
+
"loss": 0.349,
|
| 274757 |
+
"step": 137855
|
| 274758 |
+
},
|
| 274759 |
+
{
|
| 274760 |
+
"epoch": 1102.78,
|
| 274761 |
+
"learning_rate": 7.80758012820513e-06,
|
| 274762 |
+
"loss": 0.6813,
|
| 274763 |
+
"step": 137860
|
| 274764 |
+
},
|
| 274765 |
+
{
|
| 274766 |
+
"epoch": 1102.82,
|
| 274767 |
+
"learning_rate": 7.807500000000001e-06,
|
| 274768 |
+
"loss": 0.9494,
|
| 274769 |
+
"step": 137865
|
| 274770 |
+
},
|
| 274771 |
+
{
|
| 274772 |
+
"epoch": 1102.86,
|
| 274773 |
+
"learning_rate": 7.807419871794872e-06,
|
| 274774 |
+
"loss": 0.2793,
|
| 274775 |
+
"step": 137870
|
| 274776 |
+
},
|
| 274777 |
+
{
|
| 274778 |
+
"epoch": 1102.9,
|
| 274779 |
+
"learning_rate": 7.807339743589745e-06,
|
| 274780 |
+
"loss": 0.278,
|
| 274781 |
+
"step": 137875
|
| 274782 |
+
},
|
| 274783 |
+
{
|
| 274784 |
+
"epoch": 1102.94,
|
| 274785 |
+
"learning_rate": 7.807259615384615e-06,
|
| 274786 |
+
"loss": 0.3496,
|
| 274787 |
+
"step": 137880
|
| 274788 |
+
},
|
| 274789 |
+
{
|
| 274790 |
+
"epoch": 1102.98,
|
| 274791 |
+
"learning_rate": 7.807179487179488e-06,
|
| 274792 |
+
"loss": 0.7243,
|
| 274793 |
+
"step": 137885
|
| 274794 |
+
},
|
| 274795 |
+
{
|
| 274796 |
+
"epoch": 1103.0,
|
| 274797 |
+
"eval_loss": 0.35368233919143677,
|
| 274798 |
+
"eval_runtime": 36.042,
|
| 274799 |
+
"eval_samples_per_second": 23.389,
|
| 274800 |
+
"eval_steps_per_second": 0.749,
|
| 274801 |
+
"eval_wer": 0.19015407741450582,
|
| 274802 |
+
"step": 137888
|
| 274803 |
}
|
| 274804 |
],
|
| 274805 |
"max_steps": 625000,
|
| 274806 |
"num_train_epochs": 5000,
|
| 274807 |
+
"total_flos": 3.880111263484614e+20,
|
| 274808 |
"trial_name": null,
|
| 274809 |
"trial_params": null
|
| 274810 |
}
|
model-bin/finetune/base/{checkpoint-137263 β checkpoint-137888}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1630185503.3679285/events.out.tfevents.1630185503.86bb0ddabf9b.4092.311
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f3bf4bd2af253b5715776dbd76716207a7d200a52737f11559fab6c6257bd848
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630185910.66528/events.out.tfevents.1630185910.86bb0ddabf9b.4092.313
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7290fa65d2aee61bae2e383e1afdbb1499dd951d799d6de689baa695deee8e16
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630186302.3971124/events.out.tfevents.1630186302.86bb0ddabf9b.4092.315
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:583752ddf4945ee796cf9e0eeb1e5f63e910864da21ca2064ff5e1e81f33ec6f
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630186692.7030358/events.out.tfevents.1630186692.86bb0ddabf9b.4092.317
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c19a78afc8b2dc7fa3ff6f7fc88c144de620f48e373fbfd68c49a8062d946148
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630187079.2727177/events.out.tfevents.1630187079.86bb0ddabf9b.4092.319
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ede442494fd299a3cb0c0e42ae997d7d447a13a439c3fb5eed84f9fc4aeecdc0
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630185503.86bb0ddabf9b.4092.310
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:db4100b26eaf992aa8397e74d2165a0662371941137d0991f692f51773d58be1
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630185910.86bb0ddabf9b.4092.312
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:23ac6cdc07e294b6e3247e161f8f9323492fd3a763f4929182f48e3ef605dba3
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630186302.86bb0ddabf9b.4092.314
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:46b530fdd5a28912c3c5654e59e6bb6158fb54694e266c868341d196408bfdf2
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630186692.86bb0ddabf9b.4092.316
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:393d3439e3a024153c73bbfbb3dd63648a3fe9253236b4e4b0afc279d48fbf71
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630187079.86bb0ddabf9b.4092.318
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:10d8c5d88d5c9e3d656a7cca9fd2b9ac80e31a24b0240d364133350212c3d480
|
| 3 |
+
size 8622
|