"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/trainer_state.json +798 -3
- model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630259719.897293/events.out.tfevents.1630259719.cc93b136ebf5.1086.249 +3 -0
- model-bin/finetune/base/log/1630260157.7758684/events.out.tfevents.1630260157.cc93b136ebf5.1086.251 +3 -0
- model-bin/finetune/base/log/1630260591.4709191/events.out.tfevents.1630260591.cc93b136ebf5.1086.253 +3 -0
- model-bin/finetune/base/log/1630261026.7725165/events.out.tfevents.1630261026.cc93b136ebf5.1086.255 +3 -0
- model-bin/finetune/base/log/1630261460.190783/events.out.tfevents.1630261460.cc93b136ebf5.1086.257 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630259719.cc93b136ebf5.1086.248 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630260157.cc93b136ebf5.1086.250 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630260591.cc93b136ebf5.1086.252 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630261026.cc93b136ebf5.1086.254 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630261460.cc93b136ebf5.1086.256 +3 -0
model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:470c3ea26795136bd282cb4eb7e01b82e7f74de658b159421d45cd8357c5df2e
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b99febcccd6a1f50c15709024b6aadd57bf558412f09f39446be17d92b116544
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:847621cdc9516237f2d0544ce00f3369310442cbe0ff46e33f4438f8246622c7
|
| 3 |
size 14503
|
model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1adce13f705e5edbbf832291b3d4b862eb38f0d353dd7a919d7e7015308b5f53
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ccc55077eebec856cae793ae16f745e98931efbbdb640f5646e5c41e6164ecd5
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1668011442822563,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -296010,11 +296010,806 @@
|
|
| 296010 |
"eval_steps_per_second": 0.654,
|
| 296011 |
"eval_wer": 0.17398648648648649,
|
| 296012 |
"step": 154557
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 296013 |
}
|
| 296014 |
],
|
| 296015 |
"max_steps": 620000,
|
| 296016 |
"num_train_epochs": 5000,
|
| 296017 |
-
"total_flos": 4.
|
| 296018 |
"trial_name": null,
|
| 296019 |
"trial_params": null
|
| 296020 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1668011442822563,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
| 4 |
+
"epoch": 1250.995983935743,
|
| 5 |
+
"global_step": 155180,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 296010 |
"eval_steps_per_second": 0.654,
|
| 296011 |
"eval_wer": 0.17398648648648649,
|
| 296012 |
"step": 154557
|
| 296013 |
+
},
|
| 296014 |
+
{
|
| 296015 |
+
"epoch": 1246.02,
|
| 296016 |
+
"learning_rate": 7.5202100161550895e-06,
|
| 296017 |
+
"loss": 0.3359,
|
| 296018 |
+
"step": 154560
|
| 296019 |
+
},
|
| 296020 |
+
{
|
| 296021 |
+
"epoch": 1246.06,
|
| 296022 |
+
"learning_rate": 7.520129240710825e-06,
|
| 296023 |
+
"loss": 0.2438,
|
| 296024 |
+
"step": 154565
|
| 296025 |
+
},
|
| 296026 |
+
{
|
| 296027 |
+
"epoch": 1246.1,
|
| 296028 |
+
"learning_rate": 7.5200484652665595e-06,
|
| 296029 |
+
"loss": 0.2681,
|
| 296030 |
+
"step": 154570
|
| 296031 |
+
},
|
| 296032 |
+
{
|
| 296033 |
+
"epoch": 1246.14,
|
| 296034 |
+
"learning_rate": 7.519967689822295e-06,
|
| 296035 |
+
"loss": 0.3789,
|
| 296036 |
+
"step": 154575
|
| 296037 |
+
},
|
| 296038 |
+
{
|
| 296039 |
+
"epoch": 1246.18,
|
| 296040 |
+
"learning_rate": 7.5198869143780295e-06,
|
| 296041 |
+
"loss": 0.6661,
|
| 296042 |
+
"step": 154580
|
| 296043 |
+
},
|
| 296044 |
+
{
|
| 296045 |
+
"epoch": 1246.22,
|
| 296046 |
+
"learning_rate": 7.519806138933765e-06,
|
| 296047 |
+
"loss": 0.8786,
|
| 296048 |
+
"step": 154585
|
| 296049 |
+
},
|
| 296050 |
+
{
|
| 296051 |
+
"epoch": 1246.27,
|
| 296052 |
+
"learning_rate": 7.5197253634894995e-06,
|
| 296053 |
+
"loss": 0.2525,
|
| 296054 |
+
"step": 154590
|
| 296055 |
+
},
|
| 296056 |
+
{
|
| 296057 |
+
"epoch": 1246.31,
|
| 296058 |
+
"learning_rate": 7.519644588045235e-06,
|
| 296059 |
+
"loss": 0.2866,
|
| 296060 |
+
"step": 154595
|
| 296061 |
+
},
|
| 296062 |
+
{
|
| 296063 |
+
"epoch": 1246.35,
|
| 296064 |
+
"learning_rate": 7.5195638126009694e-06,
|
| 296065 |
+
"loss": 0.3159,
|
| 296066 |
+
"step": 154600
|
| 296067 |
+
},
|
| 296068 |
+
{
|
| 296069 |
+
"epoch": 1246.39,
|
| 296070 |
+
"learning_rate": 7.519483037156705e-06,
|
| 296071 |
+
"loss": 0.6285,
|
| 296072 |
+
"step": 154605
|
| 296073 |
+
},
|
| 296074 |
+
{
|
| 296075 |
+
"epoch": 1246.43,
|
| 296076 |
+
"learning_rate": 7.519402261712439e-06,
|
| 296077 |
+
"loss": 0.8862,
|
| 296078 |
+
"step": 154610
|
| 296079 |
+
},
|
| 296080 |
+
{
|
| 296081 |
+
"epoch": 1246.47,
|
| 296082 |
+
"learning_rate": 7.519321486268175e-06,
|
| 296083 |
+
"loss": 0.2578,
|
| 296084 |
+
"step": 154615
|
| 296085 |
+
},
|
| 296086 |
+
{
|
| 296087 |
+
"epoch": 1246.51,
|
| 296088 |
+
"learning_rate": 7.519240710823909e-06,
|
| 296089 |
+
"loss": 0.2415,
|
| 296090 |
+
"step": 154620
|
| 296091 |
+
},
|
| 296092 |
+
{
|
| 296093 |
+
"epoch": 1246.55,
|
| 296094 |
+
"learning_rate": 7.519159935379645e-06,
|
| 296095 |
+
"loss": 0.3066,
|
| 296096 |
+
"step": 154625
|
| 296097 |
+
},
|
| 296098 |
+
{
|
| 296099 |
+
"epoch": 1246.59,
|
| 296100 |
+
"learning_rate": 7.519079159935381e-06,
|
| 296101 |
+
"loss": 0.5628,
|
| 296102 |
+
"step": 154630
|
| 296103 |
+
},
|
| 296104 |
+
{
|
| 296105 |
+
"epoch": 1246.63,
|
| 296106 |
+
"learning_rate": 7.518998384491115e-06,
|
| 296107 |
+
"loss": 0.865,
|
| 296108 |
+
"step": 154635
|
| 296109 |
+
},
|
| 296110 |
+
{
|
| 296111 |
+
"epoch": 1246.67,
|
| 296112 |
+
"learning_rate": 7.518917609046851e-06,
|
| 296113 |
+
"loss": 0.2865,
|
| 296114 |
+
"step": 154640
|
| 296115 |
+
},
|
| 296116 |
+
{
|
| 296117 |
+
"epoch": 1246.71,
|
| 296118 |
+
"learning_rate": 7.518836833602585e-06,
|
| 296119 |
+
"loss": 0.2618,
|
| 296120 |
+
"step": 154645
|
| 296121 |
+
},
|
| 296122 |
+
{
|
| 296123 |
+
"epoch": 1246.75,
|
| 296124 |
+
"learning_rate": 7.518756058158321e-06,
|
| 296125 |
+
"loss": 0.3727,
|
| 296126 |
+
"step": 154650
|
| 296127 |
+
},
|
| 296128 |
+
{
|
| 296129 |
+
"epoch": 1246.79,
|
| 296130 |
+
"learning_rate": 7.518675282714055e-06,
|
| 296131 |
+
"loss": 0.8115,
|
| 296132 |
+
"step": 154655
|
| 296133 |
+
},
|
| 296134 |
+
{
|
| 296135 |
+
"epoch": 1246.83,
|
| 296136 |
+
"learning_rate": 7.518594507269791e-06,
|
| 296137 |
+
"loss": 0.7482,
|
| 296138 |
+
"step": 154660
|
| 296139 |
+
},
|
| 296140 |
+
{
|
| 296141 |
+
"epoch": 1246.87,
|
| 296142 |
+
"learning_rate": 7.518513731825525e-06,
|
| 296143 |
+
"loss": 0.3001,
|
| 296144 |
+
"step": 154665
|
| 296145 |
+
},
|
| 296146 |
+
{
|
| 296147 |
+
"epoch": 1246.91,
|
| 296148 |
+
"learning_rate": 7.518432956381261e-06,
|
| 296149 |
+
"loss": 0.3185,
|
| 296150 |
+
"step": 154670
|
| 296151 |
+
},
|
| 296152 |
+
{
|
| 296153 |
+
"epoch": 1246.95,
|
| 296154 |
+
"learning_rate": 7.518352180936995e-06,
|
| 296155 |
+
"loss": 0.3938,
|
| 296156 |
+
"step": 154675
|
| 296157 |
+
},
|
| 296158 |
+
{
|
| 296159 |
+
"epoch": 1246.99,
|
| 296160 |
+
"learning_rate": 7.518271405492731e-06,
|
| 296161 |
+
"loss": 0.8753,
|
| 296162 |
+
"step": 154680
|
| 296163 |
+
},
|
| 296164 |
+
{
|
| 296165 |
+
"epoch": 1247.0,
|
| 296166 |
+
"eval_loss": 0.3975435793399811,
|
| 296167 |
+
"eval_runtime": 41.5728,
|
| 296168 |
+
"eval_samples_per_second": 20.326,
|
| 296169 |
+
"eval_steps_per_second": 0.649,
|
| 296170 |
+
"eval_wer": 0.1845796090400407,
|
| 296171 |
+
"step": 154681
|
| 296172 |
+
},
|
| 296173 |
+
{
|
| 296174 |
+
"epoch": 1237.03,
|
| 296175 |
+
"learning_rate": 7.518190630048467e-06,
|
| 296176 |
+
"loss": 0.3444,
|
| 296177 |
+
"step": 154685
|
| 296178 |
+
},
|
| 296179 |
+
{
|
| 296180 |
+
"epoch": 1237.07,
|
| 296181 |
+
"learning_rate": 7.518109854604201e-06,
|
| 296182 |
+
"loss": 0.2737,
|
| 296183 |
+
"step": 154690
|
| 296184 |
+
},
|
| 296185 |
+
{
|
| 296186 |
+
"epoch": 1237.11,
|
| 296187 |
+
"learning_rate": 7.518029079159937e-06,
|
| 296188 |
+
"loss": 0.2676,
|
| 296189 |
+
"step": 154695
|
| 296190 |
+
},
|
| 296191 |
+
{
|
| 296192 |
+
"epoch": 1237.15,
|
| 296193 |
+
"learning_rate": 7.517948303715671e-06,
|
| 296194 |
+
"loss": 0.3919,
|
| 296195 |
+
"step": 154700
|
| 296196 |
+
},
|
| 296197 |
+
{
|
| 296198 |
+
"epoch": 1237.19,
|
| 296199 |
+
"learning_rate": 7.517867528271407e-06,
|
| 296200 |
+
"loss": 1.0205,
|
| 296201 |
+
"step": 154705
|
| 296202 |
+
},
|
| 296203 |
+
{
|
| 296204 |
+
"epoch": 1237.23,
|
| 296205 |
+
"learning_rate": 7.517786752827141e-06,
|
| 296206 |
+
"loss": 0.6594,
|
| 296207 |
+
"step": 154710
|
| 296208 |
+
},
|
| 296209 |
+
{
|
| 296210 |
+
"epoch": 1237.27,
|
| 296211 |
+
"learning_rate": 7.517705977382877e-06,
|
| 296212 |
+
"loss": 0.2955,
|
| 296213 |
+
"step": 154715
|
| 296214 |
+
},
|
| 296215 |
+
{
|
| 296216 |
+
"epoch": 1237.31,
|
| 296217 |
+
"learning_rate": 7.517625201938611e-06,
|
| 296218 |
+
"loss": 0.2682,
|
| 296219 |
+
"step": 154720
|
| 296220 |
+
},
|
| 296221 |
+
{
|
| 296222 |
+
"epoch": 1237.35,
|
| 296223 |
+
"learning_rate": 7.517544426494347e-06,
|
| 296224 |
+
"loss": 0.3546,
|
| 296225 |
+
"step": 154725
|
| 296226 |
+
},
|
| 296227 |
+
{
|
| 296228 |
+
"epoch": 1237.39,
|
| 296229 |
+
"learning_rate": 7.517463651050081e-06,
|
| 296230 |
+
"loss": 0.8458,
|
| 296231 |
+
"step": 154730
|
| 296232 |
+
},
|
| 296233 |
+
{
|
| 296234 |
+
"epoch": 1237.43,
|
| 296235 |
+
"learning_rate": 7.517382875605817e-06,
|
| 296236 |
+
"loss": 0.6329,
|
| 296237 |
+
"step": 154735
|
| 296238 |
+
},
|
| 296239 |
+
{
|
| 296240 |
+
"epoch": 1237.47,
|
| 296241 |
+
"learning_rate": 7.517302100161552e-06,
|
| 296242 |
+
"loss": 0.2381,
|
| 296243 |
+
"step": 154740
|
| 296244 |
+
},
|
| 296245 |
+
{
|
| 296246 |
+
"epoch": 1237.51,
|
| 296247 |
+
"learning_rate": 7.517221324717287e-06,
|
| 296248 |
+
"loss": 0.3116,
|
| 296249 |
+
"step": 154745
|
| 296250 |
+
},
|
| 296251 |
+
{
|
| 296252 |
+
"epoch": 1237.55,
|
| 296253 |
+
"learning_rate": 7.5171405492730225e-06,
|
| 296254 |
+
"loss": 0.3383,
|
| 296255 |
+
"step": 154750
|
| 296256 |
+
},
|
| 296257 |
+
{
|
| 296258 |
+
"epoch": 1237.59,
|
| 296259 |
+
"learning_rate": 7.517059773828757e-06,
|
| 296260 |
+
"loss": 0.8708,
|
| 296261 |
+
"step": 154755
|
| 296262 |
+
},
|
| 296263 |
+
{
|
| 296264 |
+
"epoch": 1237.63,
|
| 296265 |
+
"learning_rate": 7.5169789983844924e-06,
|
| 296266 |
+
"loss": 0.5733,
|
| 296267 |
+
"step": 154760
|
| 296268 |
+
},
|
| 296269 |
+
{
|
| 296270 |
+
"epoch": 1237.67,
|
| 296271 |
+
"learning_rate": 7.5168982229402266e-06,
|
| 296272 |
+
"loss": 0.247,
|
| 296273 |
+
"step": 154765
|
| 296274 |
+
},
|
| 296275 |
+
{
|
| 296276 |
+
"epoch": 1237.71,
|
| 296277 |
+
"learning_rate": 7.516817447495962e-06,
|
| 296278 |
+
"loss": 0.3016,
|
| 296279 |
+
"step": 154770
|
| 296280 |
+
},
|
| 296281 |
+
{
|
| 296282 |
+
"epoch": 1237.75,
|
| 296283 |
+
"learning_rate": 7.5167366720516965e-06,
|
| 296284 |
+
"loss": 0.4686,
|
| 296285 |
+
"step": 154775
|
| 296286 |
+
},
|
| 296287 |
+
{
|
| 296288 |
+
"epoch": 1237.79,
|
| 296289 |
+
"learning_rate": 7.516655896607432e-06,
|
| 296290 |
+
"loss": 0.8122,
|
| 296291 |
+
"step": 154780
|
| 296292 |
+
},
|
| 296293 |
+
{
|
| 296294 |
+
"epoch": 1237.83,
|
| 296295 |
+
"learning_rate": 7.5165751211631665e-06,
|
| 296296 |
+
"loss": 0.5604,
|
| 296297 |
+
"step": 154785
|
| 296298 |
+
},
|
| 296299 |
+
{
|
| 296300 |
+
"epoch": 1237.87,
|
| 296301 |
+
"learning_rate": 7.516494345718902e-06,
|
| 296302 |
+
"loss": 0.3368,
|
| 296303 |
+
"step": 154790
|
| 296304 |
+
},
|
| 296305 |
+
{
|
| 296306 |
+
"epoch": 1237.91,
|
| 296307 |
+
"learning_rate": 7.5164135702746365e-06,
|
| 296308 |
+
"loss": 0.2458,
|
| 296309 |
+
"step": 154795
|
| 296310 |
+
},
|
| 296311 |
+
{
|
| 296312 |
+
"epoch": 1237.95,
|
| 296313 |
+
"learning_rate": 7.516332794830372e-06,
|
| 296314 |
+
"loss": 0.3618,
|
| 296315 |
+
"step": 154800
|
| 296316 |
+
},
|
| 296317 |
+
{
|
| 296318 |
+
"epoch": 1237.99,
|
| 296319 |
+
"learning_rate": 7.516252019386107e-06,
|
| 296320 |
+
"loss": 0.931,
|
| 296321 |
+
"step": 154805
|
| 296322 |
+
},
|
| 296323 |
+
{
|
| 296324 |
+
"epoch": 1238.0,
|
| 296325 |
+
"eval_loss": 0.42138466238975525,
|
| 296326 |
+
"eval_runtime": 40.226,
|
| 296327 |
+
"eval_samples_per_second": 20.981,
|
| 296328 |
+
"eval_steps_per_second": 0.671,
|
| 296329 |
+
"eval_wer": 0.17503371796793046,
|
| 296330 |
+
"step": 154806
|
| 296331 |
+
},
|
| 296332 |
+
{
|
| 296333 |
+
"epoch": 1238.03,
|
| 296334 |
+
"learning_rate": 7.516171243941842e-06,
|
| 296335 |
+
"loss": 0.2782,
|
| 296336 |
+
"step": 154810
|
| 296337 |
+
},
|
| 296338 |
+
{
|
| 296339 |
+
"epoch": 1238.07,
|
| 296340 |
+
"learning_rate": 7.516090468497578e-06,
|
| 296341 |
+
"loss": 0.2432,
|
| 296342 |
+
"step": 154815
|
| 296343 |
+
},
|
| 296344 |
+
{
|
| 296345 |
+
"epoch": 1238.11,
|
| 296346 |
+
"learning_rate": 7.516009693053312e-06,
|
| 296347 |
+
"loss": 0.2524,
|
| 296348 |
+
"step": 154820
|
| 296349 |
+
},
|
| 296350 |
+
{
|
| 296351 |
+
"epoch": 1238.15,
|
| 296352 |
+
"learning_rate": 7.515928917609048e-06,
|
| 296353 |
+
"loss": 0.3604,
|
| 296354 |
+
"step": 154825
|
| 296355 |
+
},
|
| 296356 |
+
{
|
| 296357 |
+
"epoch": 1238.19,
|
| 296358 |
+
"learning_rate": 7.515848142164782e-06,
|
| 296359 |
+
"loss": 0.934,
|
| 296360 |
+
"step": 154830
|
| 296361 |
+
},
|
| 296362 |
+
{
|
| 296363 |
+
"epoch": 1238.23,
|
| 296364 |
+
"learning_rate": 7.515767366720518e-06,
|
| 296365 |
+
"loss": 0.5878,
|
| 296366 |
+
"step": 154835
|
| 296367 |
+
},
|
| 296368 |
+
{
|
| 296369 |
+
"epoch": 1238.27,
|
| 296370 |
+
"learning_rate": 7.515686591276252e-06,
|
| 296371 |
+
"loss": 0.2449,
|
| 296372 |
+
"step": 154840
|
| 296373 |
+
},
|
| 296374 |
+
{
|
| 296375 |
+
"epoch": 1238.31,
|
| 296376 |
+
"learning_rate": 7.515605815831988e-06,
|
| 296377 |
+
"loss": 0.2886,
|
| 296378 |
+
"step": 154845
|
| 296379 |
+
},
|
| 296380 |
+
{
|
| 296381 |
+
"epoch": 1238.35,
|
| 296382 |
+
"learning_rate": 7.515525040387722e-06,
|
| 296383 |
+
"loss": 0.3381,
|
| 296384 |
+
"step": 154850
|
| 296385 |
+
},
|
| 296386 |
+
{
|
| 296387 |
+
"epoch": 1238.39,
|
| 296388 |
+
"learning_rate": 7.515444264943458e-06,
|
| 296389 |
+
"loss": 0.7834,
|
| 296390 |
+
"step": 154855
|
| 296391 |
+
},
|
| 296392 |
+
{
|
| 296393 |
+
"epoch": 1238.43,
|
| 296394 |
+
"learning_rate": 7.515363489499193e-06,
|
| 296395 |
+
"loss": 0.7959,
|
| 296396 |
+
"step": 154860
|
| 296397 |
+
},
|
| 296398 |
+
{
|
| 296399 |
+
"epoch": 1238.47,
|
| 296400 |
+
"learning_rate": 7.515282714054928e-06,
|
| 296401 |
+
"loss": 0.2812,
|
| 296402 |
+
"step": 154865
|
| 296403 |
+
},
|
| 296404 |
+
{
|
| 296405 |
+
"epoch": 1238.51,
|
| 296406 |
+
"learning_rate": 7.515201938610663e-06,
|
| 296407 |
+
"loss": 0.2775,
|
| 296408 |
+
"step": 154870
|
| 296409 |
+
},
|
| 296410 |
+
{
|
| 296411 |
+
"epoch": 1238.55,
|
| 296412 |
+
"learning_rate": 7.515121163166398e-06,
|
| 296413 |
+
"loss": 0.4103,
|
| 296414 |
+
"step": 154875
|
| 296415 |
+
},
|
| 296416 |
+
{
|
| 296417 |
+
"epoch": 1238.59,
|
| 296418 |
+
"learning_rate": 7.515040387722134e-06,
|
| 296419 |
+
"loss": 0.8922,
|
| 296420 |
+
"step": 154880
|
| 296421 |
+
},
|
| 296422 |
+
{
|
| 296423 |
+
"epoch": 1238.63,
|
| 296424 |
+
"learning_rate": 7.514959612277868e-06,
|
| 296425 |
+
"loss": 0.696,
|
| 296426 |
+
"step": 154885
|
| 296427 |
+
},
|
| 296428 |
+
{
|
| 296429 |
+
"epoch": 1238.67,
|
| 296430 |
+
"learning_rate": 7.514878836833604e-06,
|
| 296431 |
+
"loss": 0.2623,
|
| 296432 |
+
"step": 154890
|
| 296433 |
+
},
|
| 296434 |
+
{
|
| 296435 |
+
"epoch": 1238.71,
|
| 296436 |
+
"learning_rate": 7.514798061389338e-06,
|
| 296437 |
+
"loss": 0.2932,
|
| 296438 |
+
"step": 154895
|
| 296439 |
+
},
|
| 296440 |
+
{
|
| 296441 |
+
"epoch": 1238.75,
|
| 296442 |
+
"learning_rate": 7.514717285945074e-06,
|
| 296443 |
+
"loss": 0.3314,
|
| 296444 |
+
"step": 154900
|
| 296445 |
+
},
|
| 296446 |
+
{
|
| 296447 |
+
"epoch": 1238.79,
|
| 296448 |
+
"learning_rate": 7.514636510500808e-06,
|
| 296449 |
+
"loss": 0.8592,
|
| 296450 |
+
"step": 154905
|
| 296451 |
+
},
|
| 296452 |
+
{
|
| 296453 |
+
"epoch": 1238.83,
|
| 296454 |
+
"learning_rate": 7.514555735056544e-06,
|
| 296455 |
+
"loss": 0.5791,
|
| 296456 |
+
"step": 154910
|
| 296457 |
+
},
|
| 296458 |
+
{
|
| 296459 |
+
"epoch": 1238.87,
|
| 296460 |
+
"learning_rate": 7.514474959612279e-06,
|
| 296461 |
+
"loss": 0.2748,
|
| 296462 |
+
"step": 154915
|
| 296463 |
+
},
|
| 296464 |
+
{
|
| 296465 |
+
"epoch": 1238.91,
|
| 296466 |
+
"learning_rate": 7.514394184168014e-06,
|
| 296467 |
+
"loss": 0.3728,
|
| 296468 |
+
"step": 154920
|
| 296469 |
+
},
|
| 296470 |
+
{
|
| 296471 |
+
"epoch": 1238.95,
|
| 296472 |
+
"learning_rate": 7.514313408723749e-06,
|
| 296473 |
+
"loss": 0.4352,
|
| 296474 |
+
"step": 154925
|
| 296475 |
+
},
|
| 296476 |
+
{
|
| 296477 |
+
"epoch": 1238.99,
|
| 296478 |
+
"learning_rate": 7.514232633279484e-06,
|
| 296479 |
+
"loss": 0.8688,
|
| 296480 |
+
"step": 154930
|
| 296481 |
+
},
|
| 296482 |
+
{
|
| 296483 |
+
"epoch": 1239.0,
|
| 296484 |
+
"eval_loss": 0.36569637060165405,
|
| 296485 |
+
"eval_runtime": 40.9116,
|
| 296486 |
+
"eval_samples_per_second": 20.654,
|
| 296487 |
+
"eval_steps_per_second": 0.66,
|
| 296488 |
+
"eval_wer": 0.18007352941176472,
|
| 296489 |
+
"step": 154931
|
| 296490 |
+
},
|
| 296491 |
+
{
|
| 296492 |
+
"epoch": 1239.03,
|
| 296493 |
+
"learning_rate": 7.514151857835219e-06,
|
| 296494 |
+
"loss": 0.3529,
|
| 296495 |
+
"step": 154935
|
| 296496 |
+
},
|
| 296497 |
+
{
|
| 296498 |
+
"epoch": 1239.07,
|
| 296499 |
+
"learning_rate": 7.514071082390954e-06,
|
| 296500 |
+
"loss": 0.2858,
|
| 296501 |
+
"step": 154940
|
| 296502 |
+
},
|
| 296503 |
+
{
|
| 296504 |
+
"epoch": 1239.11,
|
| 296505 |
+
"learning_rate": 7.513990306946689e-06,
|
| 296506 |
+
"loss": 0.3375,
|
| 296507 |
+
"step": 154945
|
| 296508 |
+
},
|
| 296509 |
+
{
|
| 296510 |
+
"epoch": 1239.15,
|
| 296511 |
+
"learning_rate": 7.513909531502424e-06,
|
| 296512 |
+
"loss": 0.3088,
|
| 296513 |
+
"step": 154950
|
| 296514 |
+
},
|
| 296515 |
+
{
|
| 296516 |
+
"epoch": 1239.19,
|
| 296517 |
+
"learning_rate": 7.5138287560581595e-06,
|
| 296518 |
+
"loss": 0.7962,
|
| 296519 |
+
"step": 154955
|
| 296520 |
+
},
|
| 296521 |
+
{
|
| 296522 |
+
"epoch": 1239.23,
|
| 296523 |
+
"learning_rate": 7.513747980613894e-06,
|
| 296524 |
+
"loss": 0.6496,
|
| 296525 |
+
"step": 154960
|
| 296526 |
+
},
|
| 296527 |
+
{
|
| 296528 |
+
"epoch": 1239.27,
|
| 296529 |
+
"learning_rate": 7.5136672051696295e-06,
|
| 296530 |
+
"loss": 0.3379,
|
| 296531 |
+
"step": 154965
|
| 296532 |
+
},
|
| 296533 |
+
{
|
| 296534 |
+
"epoch": 1239.31,
|
| 296535 |
+
"learning_rate": 7.513586429725364e-06,
|
| 296536 |
+
"loss": 0.2975,
|
| 296537 |
+
"step": 154970
|
| 296538 |
+
},
|
| 296539 |
+
{
|
| 296540 |
+
"epoch": 1239.35,
|
| 296541 |
+
"learning_rate": 7.5135056542810994e-06,
|
| 296542 |
+
"loss": 0.3822,
|
| 296543 |
+
"step": 154975
|
| 296544 |
+
},
|
| 296545 |
+
{
|
| 296546 |
+
"epoch": 1239.39,
|
| 296547 |
+
"learning_rate": 7.513424878836834e-06,
|
| 296548 |
+
"loss": 0.9315,
|
| 296549 |
+
"step": 154980
|
| 296550 |
+
},
|
| 296551 |
+
{
|
| 296552 |
+
"epoch": 1239.43,
|
| 296553 |
+
"learning_rate": 7.513344103392569e-06,
|
| 296554 |
+
"loss": 0.6303,
|
| 296555 |
+
"step": 154985
|
| 296556 |
+
},
|
| 296557 |
+
{
|
| 296558 |
+
"epoch": 1239.47,
|
| 296559 |
+
"learning_rate": 7.513263327948304e-06,
|
| 296560 |
+
"loss": 0.3489,
|
| 296561 |
+
"step": 154990
|
| 296562 |
+
},
|
| 296563 |
+
{
|
| 296564 |
+
"epoch": 1239.51,
|
| 296565 |
+
"learning_rate": 7.513182552504039e-06,
|
| 296566 |
+
"loss": 0.3152,
|
| 296567 |
+
"step": 154995
|
| 296568 |
+
},
|
| 296569 |
+
{
|
| 296570 |
+
"epoch": 1239.55,
|
| 296571 |
+
"learning_rate": 7.513101777059774e-06,
|
| 296572 |
+
"loss": 0.3954,
|
| 296573 |
+
"step": 155000
|
| 296574 |
+
},
|
| 296575 |
+
{
|
| 296576 |
+
"epoch": 1239.59,
|
| 296577 |
+
"learning_rate": 7.513021001615509e-06,
|
| 296578 |
+
"loss": 0.878,
|
| 296579 |
+
"step": 155005
|
| 296580 |
+
},
|
| 296581 |
+
{
|
| 296582 |
+
"epoch": 1239.63,
|
| 296583 |
+
"learning_rate": 7.512940226171244e-06,
|
| 296584 |
+
"loss": 0.627,
|
| 296585 |
+
"step": 155010
|
| 296586 |
+
},
|
| 296587 |
+
{
|
| 296588 |
+
"epoch": 1239.67,
|
| 296589 |
+
"learning_rate": 7.512859450726979e-06,
|
| 296590 |
+
"loss": 0.2331,
|
| 296591 |
+
"step": 155015
|
| 296592 |
+
},
|
| 296593 |
+
{
|
| 296594 |
+
"epoch": 1239.71,
|
| 296595 |
+
"learning_rate": 7.512778675282715e-06,
|
| 296596 |
+
"loss": 0.2644,
|
| 296597 |
+
"step": 155020
|
| 296598 |
+
},
|
| 296599 |
+
{
|
| 296600 |
+
"epoch": 1239.75,
|
| 296601 |
+
"learning_rate": 7.512697899838449e-06,
|
| 296602 |
+
"loss": 0.3715,
|
| 296603 |
+
"step": 155025
|
| 296604 |
+
},
|
| 296605 |
+
{
|
| 296606 |
+
"epoch": 1239.79,
|
| 296607 |
+
"learning_rate": 7.512617124394185e-06,
|
| 296608 |
+
"loss": 0.9216,
|
| 296609 |
+
"step": 155030
|
| 296610 |
+
},
|
| 296611 |
+
{
|
| 296612 |
+
"epoch": 1239.83,
|
| 296613 |
+
"learning_rate": 7.51253634894992e-06,
|
| 296614 |
+
"loss": 0.6866,
|
| 296615 |
+
"step": 155035
|
| 296616 |
+
},
|
| 296617 |
+
{
|
| 296618 |
+
"epoch": 1239.87,
|
| 296619 |
+
"learning_rate": 7.512455573505655e-06,
|
| 296620 |
+
"loss": 0.2765,
|
| 296621 |
+
"step": 155040
|
| 296622 |
+
},
|
| 296623 |
+
{
|
| 296624 |
+
"epoch": 1239.91,
|
| 296625 |
+
"learning_rate": 7.51237479806139e-06,
|
| 296626 |
+
"loss": 0.2844,
|
| 296627 |
+
"step": 155045
|
| 296628 |
+
},
|
| 296629 |
+
{
|
| 296630 |
+
"epoch": 1239.95,
|
| 296631 |
+
"learning_rate": 7.512294022617125e-06,
|
| 296632 |
+
"loss": 0.3856,
|
| 296633 |
+
"step": 155050
|
| 296634 |
+
},
|
| 296635 |
+
{
|
| 296636 |
+
"epoch": 1239.99,
|
| 296637 |
+
"learning_rate": 7.51221324717286e-06,
|
| 296638 |
+
"loss": 1.0202,
|
| 296639 |
+
"step": 155055
|
| 296640 |
+
},
|
| 296641 |
+
{
|
| 296642 |
+
"epoch": 1240.0,
|
| 296643 |
+
"eval_loss": 0.30646491050720215,
|
| 296644 |
+
"eval_runtime": 40.3619,
|
| 296645 |
+
"eval_samples_per_second": 20.936,
|
| 296646 |
+
"eval_steps_per_second": 0.669,
|
| 296647 |
+
"eval_wer": 0.17656983568075119,
|
| 296648 |
+
"step": 155056
|
| 296649 |
+
},
|
| 296650 |
+
{
|
| 296651 |
+
"epoch": 1250.03,
|
| 296652 |
+
"learning_rate": 7.512132471728595e-06,
|
| 296653 |
+
"loss": 0.2716,
|
| 296654 |
+
"step": 155060
|
| 296655 |
+
},
|
| 296656 |
+
{
|
| 296657 |
+
"epoch": 1250.07,
|
| 296658 |
+
"learning_rate": 7.51205169628433e-06,
|
| 296659 |
+
"loss": 0.2815,
|
| 296660 |
+
"step": 155065
|
| 296661 |
+
},
|
| 296662 |
+
{
|
| 296663 |
+
"epoch": 1250.11,
|
| 296664 |
+
"learning_rate": 7.511970920840065e-06,
|
| 296665 |
+
"loss": 0.3168,
|
| 296666 |
+
"step": 155070
|
| 296667 |
+
},
|
| 296668 |
+
{
|
| 296669 |
+
"epoch": 1250.15,
|
| 296670 |
+
"learning_rate": 7.5118901453958e-06,
|
| 296671 |
+
"loss": 0.3727,
|
| 296672 |
+
"step": 155075
|
| 296673 |
+
},
|
| 296674 |
+
{
|
| 296675 |
+
"epoch": 1250.19,
|
| 296676 |
+
"learning_rate": 7.511809369951535e-06,
|
| 296677 |
+
"loss": 0.9729,
|
| 296678 |
+
"step": 155080
|
| 296679 |
+
},
|
| 296680 |
+
{
|
| 296681 |
+
"epoch": 1250.23,
|
| 296682 |
+
"learning_rate": 7.511728594507271e-06,
|
| 296683 |
+
"loss": 0.7132,
|
| 296684 |
+
"step": 155085
|
| 296685 |
+
},
|
| 296686 |
+
{
|
| 296687 |
+
"epoch": 1250.27,
|
| 296688 |
+
"learning_rate": 7.511647819063006e-06,
|
| 296689 |
+
"loss": 0.3155,
|
| 296690 |
+
"step": 155090
|
| 296691 |
+
},
|
| 296692 |
+
{
|
| 296693 |
+
"epoch": 1250.31,
|
| 296694 |
+
"learning_rate": 7.511567043618741e-06,
|
| 296695 |
+
"loss": 0.3005,
|
| 296696 |
+
"step": 155095
|
| 296697 |
+
},
|
| 296698 |
+
{
|
| 296699 |
+
"epoch": 1250.35,
|
| 296700 |
+
"learning_rate": 7.511486268174476e-06,
|
| 296701 |
+
"loss": 0.3456,
|
| 296702 |
+
"step": 155100
|
| 296703 |
+
},
|
| 296704 |
+
{
|
| 296705 |
+
"epoch": 1250.39,
|
| 296706 |
+
"learning_rate": 7.511405492730211e-06,
|
| 296707 |
+
"loss": 0.7762,
|
| 296708 |
+
"step": 155105
|
| 296709 |
+
},
|
| 296710 |
+
{
|
| 296711 |
+
"epoch": 1250.43,
|
| 296712 |
+
"learning_rate": 7.511324717285946e-06,
|
| 296713 |
+
"loss": 0.6201,
|
| 296714 |
+
"step": 155110
|
| 296715 |
+
},
|
| 296716 |
+
{
|
| 296717 |
+
"epoch": 1250.47,
|
| 296718 |
+
"learning_rate": 7.511243941841681e-06,
|
| 296719 |
+
"loss": 0.2818,
|
| 296720 |
+
"step": 155115
|
| 296721 |
+
},
|
| 296722 |
+
{
|
| 296723 |
+
"epoch": 1250.51,
|
| 296724 |
+
"learning_rate": 7.511163166397416e-06,
|
| 296725 |
+
"loss": 0.2924,
|
| 296726 |
+
"step": 155120
|
| 296727 |
+
},
|
| 296728 |
+
{
|
| 296729 |
+
"epoch": 1250.55,
|
| 296730 |
+
"learning_rate": 7.511082390953151e-06,
|
| 296731 |
+
"loss": 0.3261,
|
| 296732 |
+
"step": 155125
|
| 296733 |
+
},
|
| 296734 |
+
{
|
| 296735 |
+
"epoch": 1250.59,
|
| 296736 |
+
"learning_rate": 7.511001615508886e-06,
|
| 296737 |
+
"loss": 0.8687,
|
| 296738 |
+
"step": 155130
|
| 296739 |
+
},
|
| 296740 |
+
{
|
| 296741 |
+
"epoch": 1250.63,
|
| 296742 |
+
"learning_rate": 7.510920840064621e-06,
|
| 296743 |
+
"loss": 0.5005,
|
| 296744 |
+
"step": 155135
|
| 296745 |
+
},
|
| 296746 |
+
{
|
| 296747 |
+
"epoch": 1250.67,
|
| 296748 |
+
"learning_rate": 7.510840064620356e-06,
|
| 296749 |
+
"loss": 0.2698,
|
| 296750 |
+
"step": 155140
|
| 296751 |
+
},
|
| 296752 |
+
{
|
| 296753 |
+
"epoch": 1250.71,
|
| 296754 |
+
"learning_rate": 7.510759289176091e-06,
|
| 296755 |
+
"loss": 0.2844,
|
| 296756 |
+
"step": 155145
|
| 296757 |
+
},
|
| 296758 |
+
{
|
| 296759 |
+
"epoch": 1250.76,
|
| 296760 |
+
"learning_rate": 7.510678513731826e-06,
|
| 296761 |
+
"loss": 0.3468,
|
| 296762 |
+
"step": 155150
|
| 296763 |
+
},
|
| 296764 |
+
{
|
| 296765 |
+
"epoch": 1250.8,
|
| 296766 |
+
"learning_rate": 7.5105977382875615e-06,
|
| 296767 |
+
"loss": 0.9883,
|
| 296768 |
+
"step": 155155
|
| 296769 |
+
},
|
| 296770 |
+
{
|
| 296771 |
+
"epoch": 1250.84,
|
| 296772 |
+
"learning_rate": 7.5105169628432965e-06,
|
| 296773 |
+
"loss": 0.6705,
|
| 296774 |
+
"step": 155160
|
| 296775 |
+
},
|
| 296776 |
+
{
|
| 296777 |
+
"epoch": 1250.88,
|
| 296778 |
+
"learning_rate": 7.5104361873990315e-06,
|
| 296779 |
+
"loss": 0.2834,
|
| 296780 |
+
"step": 155165
|
| 296781 |
+
},
|
| 296782 |
+
{
|
| 296783 |
+
"epoch": 1250.92,
|
| 296784 |
+
"learning_rate": 7.5103554119547665e-06,
|
| 296785 |
+
"loss": 0.3121,
|
| 296786 |
+
"step": 155170
|
| 296787 |
+
},
|
| 296788 |
+
{
|
| 296789 |
+
"epoch": 1250.96,
|
| 296790 |
+
"learning_rate": 7.5102746365105015e-06,
|
| 296791 |
+
"loss": 0.4157,
|
| 296792 |
+
"step": 155175
|
| 296793 |
+
},
|
| 296794 |
+
{
|
| 296795 |
+
"epoch": 1251.0,
|
| 296796 |
+
"learning_rate": 7.5101938610662365e-06,
|
| 296797 |
+
"loss": 0.9646,
|
| 296798 |
+
"step": 155180
|
| 296799 |
+
},
|
| 296800 |
+
{
|
| 296801 |
+
"epoch": 1251.0,
|
| 296802 |
+
"eval_loss": 0.518424391746521,
|
| 296803 |
+
"eval_runtime": 41.1845,
|
| 296804 |
+
"eval_samples_per_second": 20.517,
|
| 296805 |
+
"eval_steps_per_second": 0.656,
|
| 296806 |
+
"eval_wer": 0.18026912388474478,
|
| 296807 |
+
"step": 155180
|
| 296808 |
}
|
| 296809 |
],
|
| 296810 |
"max_steps": 620000,
|
| 296811 |
"num_train_epochs": 5000,
|
| 296812 |
+
"total_flos": 4.366899653515725e+20,
|
| 296813 |
"trial_name": null,
|
| 296814 |
"trial_params": null
|
| 296815 |
}
|
model-bin/finetune/base/{checkpoint-154557 β checkpoint-155180}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1630259719.897293/events.out.tfevents.1630259719.cc93b136ebf5.1086.249
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b026cb9d789aece99a3e446b51eef6d6eafe7463ff3affb3acdff973c9cecafb
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630260157.7758684/events.out.tfevents.1630260157.cc93b136ebf5.1086.251
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:eb971f3865430bebfa1d58eda62f42d75ec70030bc30ab5b02fcf85757e2658f
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630260591.4709191/events.out.tfevents.1630260591.cc93b136ebf5.1086.253
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6a1abfdef8f5e60f550201561885525bc0f08b046f43b14dd16bd319ab43f679
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630261026.7725165/events.out.tfevents.1630261026.cc93b136ebf5.1086.255
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1e4d315fde2bca0802d4b45f0f764ad61cbd51cadef039b9a244dc44c1e01520
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630261460.190783/events.out.tfevents.1630261460.cc93b136ebf5.1086.257
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:620ac2c75253b61fa00bf008d66e61a212b0eeb51edec85242ba1b84ca784dd7
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630259719.cc93b136ebf5.1086.248
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f8e93d8cc698e64dca2ac9e57763f5c78feb18455eabfac9c1e2580edc192eb2
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630260157.cc93b136ebf5.1086.250
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fb210509e463781ed894247a0beaf3c1bbd3f02168ab0b5313909da570cfed24
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630260591.cc93b136ebf5.1086.252
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1cac3e5992df5b67b83acf2fc649f479413bf5f72958cbdefca5ba2ed0761bc1
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630261026.cc93b136ebf5.1086.254
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:42bd2ba8b85647d66baebed9e3e1cff725f407db4785030a2cf15ecbc6bdb6ca
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630261460.cc93b136ebf5.1086.256
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:541f0e49276ec7aa96edb6c944e0410f0b0b1cd347b64a03ba4fdbd5c0b0d642
|
| 3 |
+
size 8622
|