"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-133530 β checkpoint-134152}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-133530 β checkpoint-134152}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-133530 β checkpoint-134152}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-133530 β checkpoint-134152}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-133530 β checkpoint-134152}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-133530 β checkpoint-134152}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-133530 β checkpoint-134152}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-133530 β checkpoint-134152}/trainer_state.json +792 -3
- model-bin/finetune/base/{checkpoint-133530 β checkpoint-134152}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630173460.284948/events.out.tfevents.1630173460.86bb0ddabf9b.4092.251 +3 -0
- model-bin/finetune/base/log/1630173847.4052565/events.out.tfevents.1630173847.86bb0ddabf9b.4092.253 +3 -0
- model-bin/finetune/base/log/1630174238.155075/events.out.tfevents.1630174238.86bb0ddabf9b.4092.255 +3 -0
- model-bin/finetune/base/log/1630174625.3833299/events.out.tfevents.1630174625.86bb0ddabf9b.4092.257 +3 -0
- model-bin/finetune/base/log/1630175012.849805/events.out.tfevents.1630175012.86bb0ddabf9b.4092.259 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630173460.86bb0ddabf9b.4092.250 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630173847.86bb0ddabf9b.4092.252 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630174238.86bb0ddabf9b.4092.254 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630174625.86bb0ddabf9b.4092.256 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630175012.86bb0ddabf9b.4092.258 +3 -0
model-bin/finetune/base/{checkpoint-133530 β checkpoint-134152}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-133530 β checkpoint-134152}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3d508d8a35a8f9a643497f1992285b1c2f7877eb4215a0840cbe0a3f99384b88
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-133530 β checkpoint-134152}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-133530 β checkpoint-134152}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0e2bd640a75fb59f050e69639bbc44f6c152b69f570c1567b4a338b6476efa97
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-133530 β checkpoint-134152}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f6d3d8ba5c02cd61971126a99f8567c0c592c0bc2b36c2483b358d10fd14ecb4
|
| 3 |
size 14503
|
model-bin/finetune/base/{checkpoint-133530 β checkpoint-134152}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bc49d6fa99b57306d86d74ceb133d6a9314c50a72c07e34fb7acc7abe68fa707
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-133530 β checkpoint-134152}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8927c9b9c4b0eca7da703c920593d2940ba00dd84b20f83a2fca6543dd58d6cc
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-133530 β checkpoint-134152}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1689111747851003,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-132910",
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -269259,11 +269259,800 @@
|
|
| 269259 |
"eval_steps_per_second": 0.757,
|
| 269260 |
"eval_wer": 0.17710974284679465,
|
| 269261 |
"step": 133530
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 269262 |
}
|
| 269263 |
],
|
| 269264 |
"max_steps": 620000,
|
| 269265 |
"num_train_epochs": 5000,
|
| 269266 |
-
"total_flos": 3.
|
| 269267 |
"trial_name": null,
|
| 269268 |
"trial_params": null
|
| 269269 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1689111747851003,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-132910",
|
| 4 |
+
"epoch": 1080.995983935743,
|
| 5 |
+
"global_step": 134152,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 269259 |
"eval_steps_per_second": 0.757,
|
| 269260 |
"eval_wer": 0.17710974284679465,
|
| 269261 |
"step": 133530
|
| 269262 |
+
},
|
| 269263 |
+
{
|
| 269264 |
+
"epoch": 1068.04,
|
| 269265 |
+
"learning_rate": 7.876891025641025e-06,
|
| 269266 |
+
"loss": 0.283,
|
| 269267 |
+
"step": 133535
|
| 269268 |
+
},
|
| 269269 |
+
{
|
| 269270 |
+
"epoch": 1068.08,
|
| 269271 |
+
"learning_rate": 7.876810897435898e-06,
|
| 269272 |
+
"loss": 0.6511,
|
| 269273 |
+
"step": 133540
|
| 269274 |
+
},
|
| 269275 |
+
{
|
| 269276 |
+
"epoch": 1068.12,
|
| 269277 |
+
"learning_rate": 7.876730769230771e-06,
|
| 269278 |
+
"loss": 0.2785,
|
| 269279 |
+
"step": 133545
|
| 269280 |
+
},
|
| 269281 |
+
{
|
| 269282 |
+
"epoch": 1068.16,
|
| 269283 |
+
"learning_rate": 7.876650641025641e-06,
|
| 269284 |
+
"loss": 0.438,
|
| 269285 |
+
"step": 133550
|
| 269286 |
+
},
|
| 269287 |
+
{
|
| 269288 |
+
"epoch": 1068.2,
|
| 269289 |
+
"learning_rate": 7.876570512820514e-06,
|
| 269290 |
+
"loss": 1.1175,
|
| 269291 |
+
"step": 133555
|
| 269292 |
+
},
|
| 269293 |
+
{
|
| 269294 |
+
"epoch": 1068.24,
|
| 269295 |
+
"learning_rate": 7.876490384615385e-06,
|
| 269296 |
+
"loss": 0.2989,
|
| 269297 |
+
"step": 133560
|
| 269298 |
+
},
|
| 269299 |
+
{
|
| 269300 |
+
"epoch": 1068.28,
|
| 269301 |
+
"learning_rate": 7.876410256410257e-06,
|
| 269302 |
+
"loss": 0.2486,
|
| 269303 |
+
"step": 133565
|
| 269304 |
+
},
|
| 269305 |
+
{
|
| 269306 |
+
"epoch": 1068.32,
|
| 269307 |
+
"learning_rate": 7.876330128205128e-06,
|
| 269308 |
+
"loss": 0.31,
|
| 269309 |
+
"step": 133570
|
| 269310 |
+
},
|
| 269311 |
+
{
|
| 269312 |
+
"epoch": 1068.36,
|
| 269313 |
+
"learning_rate": 7.876250000000001e-06,
|
| 269314 |
+
"loss": 0.3822,
|
| 269315 |
+
"step": 133575
|
| 269316 |
+
},
|
| 269317 |
+
{
|
| 269318 |
+
"epoch": 1068.4,
|
| 269319 |
+
"learning_rate": 7.876169871794872e-06,
|
| 269320 |
+
"loss": 1.2008,
|
| 269321 |
+
"step": 133580
|
| 269322 |
+
},
|
| 269323 |
+
{
|
| 269324 |
+
"epoch": 1068.44,
|
| 269325 |
+
"learning_rate": 7.876089743589744e-06,
|
| 269326 |
+
"loss": 0.3558,
|
| 269327 |
+
"step": 133585
|
| 269328 |
+
},
|
| 269329 |
+
{
|
| 269330 |
+
"epoch": 1068.48,
|
| 269331 |
+
"learning_rate": 7.876009615384615e-06,
|
| 269332 |
+
"loss": 0.2611,
|
| 269333 |
+
"step": 133590
|
| 269334 |
+
},
|
| 269335 |
+
{
|
| 269336 |
+
"epoch": 1068.52,
|
| 269337 |
+
"learning_rate": 7.875929487179488e-06,
|
| 269338 |
+
"loss": 0.282,
|
| 269339 |
+
"step": 133595
|
| 269340 |
+
},
|
| 269341 |
+
{
|
| 269342 |
+
"epoch": 1068.56,
|
| 269343 |
+
"learning_rate": 7.87584935897436e-06,
|
| 269344 |
+
"loss": 0.4253,
|
| 269345 |
+
"step": 133600
|
| 269346 |
+
},
|
| 269347 |
+
{
|
| 269348 |
+
"epoch": 1068.6,
|
| 269349 |
+
"learning_rate": 7.875769230769231e-06,
|
| 269350 |
+
"loss": 1.191,
|
| 269351 |
+
"step": 133605
|
| 269352 |
+
},
|
| 269353 |
+
{
|
| 269354 |
+
"epoch": 1068.64,
|
| 269355 |
+
"learning_rate": 7.875689102564104e-06,
|
| 269356 |
+
"loss": 0.3124,
|
| 269357 |
+
"step": 133610
|
| 269358 |
+
},
|
| 269359 |
+
{
|
| 269360 |
+
"epoch": 1068.68,
|
| 269361 |
+
"learning_rate": 7.875608974358975e-06,
|
| 269362 |
+
"loss": 0.3442,
|
| 269363 |
+
"step": 133615
|
| 269364 |
+
},
|
| 269365 |
+
{
|
| 269366 |
+
"epoch": 1068.72,
|
| 269367 |
+
"learning_rate": 7.875528846153847e-06,
|
| 269368 |
+
"loss": 0.2791,
|
| 269369 |
+
"step": 133620
|
| 269370 |
+
},
|
| 269371 |
+
{
|
| 269372 |
+
"epoch": 1068.76,
|
| 269373 |
+
"learning_rate": 7.875448717948718e-06,
|
| 269374 |
+
"loss": 0.4265,
|
| 269375 |
+
"step": 133625
|
| 269376 |
+
},
|
| 269377 |
+
{
|
| 269378 |
+
"epoch": 1068.8,
|
| 269379 |
+
"learning_rate": 7.875368589743591e-06,
|
| 269380 |
+
"loss": 1.2426,
|
| 269381 |
+
"step": 133630
|
| 269382 |
+
},
|
| 269383 |
+
{
|
| 269384 |
+
"epoch": 1068.84,
|
| 269385 |
+
"learning_rate": 7.875288461538462e-06,
|
| 269386 |
+
"loss": 0.2532,
|
| 269387 |
+
"step": 133635
|
| 269388 |
+
},
|
| 269389 |
+
{
|
| 269390 |
+
"epoch": 1068.88,
|
| 269391 |
+
"learning_rate": 7.875208333333334e-06,
|
| 269392 |
+
"loss": 0.2682,
|
| 269393 |
+
"step": 133640
|
| 269394 |
+
},
|
| 269395 |
+
{
|
| 269396 |
+
"epoch": 1068.92,
|
| 269397 |
+
"learning_rate": 7.875128205128207e-06,
|
| 269398 |
+
"loss": 0.3925,
|
| 269399 |
+
"step": 133645
|
| 269400 |
+
},
|
| 269401 |
+
{
|
| 269402 |
+
"epoch": 1068.96,
|
| 269403 |
+
"learning_rate": 7.875048076923078e-06,
|
| 269404 |
+
"loss": 0.4542,
|
| 269405 |
+
"step": 133650
|
| 269406 |
+
},
|
| 269407 |
+
{
|
| 269408 |
+
"epoch": 1069.0,
|
| 269409 |
+
"learning_rate": 7.87496794871795e-06,
|
| 269410 |
+
"loss": 1.3908,
|
| 269411 |
+
"step": 133655
|
| 269412 |
+
},
|
| 269413 |
+
{
|
| 269414 |
+
"epoch": 1069.0,
|
| 269415 |
+
"eval_loss": 0.3515404164791107,
|
| 269416 |
+
"eval_runtime": 35.3248,
|
| 269417 |
+
"eval_samples_per_second": 23.638,
|
| 269418 |
+
"eval_steps_per_second": 0.764,
|
| 269419 |
+
"eval_wer": 0.1784813881061842,
|
| 269420 |
+
"step": 133655
|
| 269421 |
+
},
|
| 269422 |
+
{
|
| 269423 |
+
"epoch": 1077.04,
|
| 269424 |
+
"learning_rate": 7.874887820512821e-06,
|
| 269425 |
+
"loss": 0.2722,
|
| 269426 |
+
"step": 133660
|
| 269427 |
+
},
|
| 269428 |
+
{
|
| 269429 |
+
"epoch": 1077.08,
|
| 269430 |
+
"learning_rate": 7.874807692307694e-06,
|
| 269431 |
+
"loss": 0.2621,
|
| 269432 |
+
"step": 133665
|
| 269433 |
+
},
|
| 269434 |
+
{
|
| 269435 |
+
"epoch": 1077.12,
|
| 269436 |
+
"learning_rate": 7.874727564102564e-06,
|
| 269437 |
+
"loss": 0.285,
|
| 269438 |
+
"step": 133670
|
| 269439 |
+
},
|
| 269440 |
+
{
|
| 269441 |
+
"epoch": 1077.16,
|
| 269442 |
+
"learning_rate": 7.874647435897437e-06,
|
| 269443 |
+
"loss": 0.4751,
|
| 269444 |
+
"step": 133675
|
| 269445 |
+
},
|
| 269446 |
+
{
|
| 269447 |
+
"epoch": 1077.2,
|
| 269448 |
+
"learning_rate": 7.874567307692308e-06,
|
| 269449 |
+
"loss": 1.0416,
|
| 269450 |
+
"step": 133680
|
| 269451 |
+
},
|
| 269452 |
+
{
|
| 269453 |
+
"epoch": 1077.24,
|
| 269454 |
+
"learning_rate": 7.87448717948718e-06,
|
| 269455 |
+
"loss": 0.32,
|
| 269456 |
+
"step": 133685
|
| 269457 |
+
},
|
| 269458 |
+
{
|
| 269459 |
+
"epoch": 1077.28,
|
| 269460 |
+
"learning_rate": 7.87440705128205e-06,
|
| 269461 |
+
"loss": 0.2579,
|
| 269462 |
+
"step": 133690
|
| 269463 |
+
},
|
| 269464 |
+
{
|
| 269465 |
+
"epoch": 1077.32,
|
| 269466 |
+
"learning_rate": 7.874326923076924e-06,
|
| 269467 |
+
"loss": 0.2939,
|
| 269468 |
+
"step": 133695
|
| 269469 |
+
},
|
| 269470 |
+
{
|
| 269471 |
+
"epoch": 1077.36,
|
| 269472 |
+
"learning_rate": 7.874246794871795e-06,
|
| 269473 |
+
"loss": 0.4655,
|
| 269474 |
+
"step": 133700
|
| 269475 |
+
},
|
| 269476 |
+
{
|
| 269477 |
+
"epoch": 1077.4,
|
| 269478 |
+
"learning_rate": 7.874166666666667e-06,
|
| 269479 |
+
"loss": 1.1427,
|
| 269480 |
+
"step": 133705
|
| 269481 |
+
},
|
| 269482 |
+
{
|
| 269483 |
+
"epoch": 1077.44,
|
| 269484 |
+
"learning_rate": 7.87408653846154e-06,
|
| 269485 |
+
"loss": 0.2917,
|
| 269486 |
+
"step": 133710
|
| 269487 |
+
},
|
| 269488 |
+
{
|
| 269489 |
+
"epoch": 1077.48,
|
| 269490 |
+
"learning_rate": 7.874006410256411e-06,
|
| 269491 |
+
"loss": 0.3403,
|
| 269492 |
+
"step": 133715
|
| 269493 |
+
},
|
| 269494 |
+
{
|
| 269495 |
+
"epoch": 1077.52,
|
| 269496 |
+
"learning_rate": 7.873926282051282e-06,
|
| 269497 |
+
"loss": 0.4189,
|
| 269498 |
+
"step": 133720
|
| 269499 |
+
},
|
| 269500 |
+
{
|
| 269501 |
+
"epoch": 1077.56,
|
| 269502 |
+
"learning_rate": 7.873846153846154e-06,
|
| 269503 |
+
"loss": 0.4168,
|
| 269504 |
+
"step": 133725
|
| 269505 |
+
},
|
| 269506 |
+
{
|
| 269507 |
+
"epoch": 1077.6,
|
| 269508 |
+
"learning_rate": 7.873766025641027e-06,
|
| 269509 |
+
"loss": 1.1052,
|
| 269510 |
+
"step": 133730
|
| 269511 |
+
},
|
| 269512 |
+
{
|
| 269513 |
+
"epoch": 1077.64,
|
| 269514 |
+
"learning_rate": 7.873685897435898e-06,
|
| 269515 |
+
"loss": 0.3159,
|
| 269516 |
+
"step": 133735
|
| 269517 |
+
},
|
| 269518 |
+
{
|
| 269519 |
+
"epoch": 1077.68,
|
| 269520 |
+
"learning_rate": 7.87360576923077e-06,
|
| 269521 |
+
"loss": 0.2568,
|
| 269522 |
+
"step": 133740
|
| 269523 |
+
},
|
| 269524 |
+
{
|
| 269525 |
+
"epoch": 1077.72,
|
| 269526 |
+
"learning_rate": 7.873525641025642e-06,
|
| 269527 |
+
"loss": 0.3329,
|
| 269528 |
+
"step": 133745
|
| 269529 |
+
},
|
| 269530 |
+
{
|
| 269531 |
+
"epoch": 1077.76,
|
| 269532 |
+
"learning_rate": 7.873445512820514e-06,
|
| 269533 |
+
"loss": 0.4664,
|
| 269534 |
+
"step": 133750
|
| 269535 |
+
},
|
| 269536 |
+
{
|
| 269537 |
+
"epoch": 1077.8,
|
| 269538 |
+
"learning_rate": 7.873365384615385e-06,
|
| 269539 |
+
"loss": 1.05,
|
| 269540 |
+
"step": 133755
|
| 269541 |
+
},
|
| 269542 |
+
{
|
| 269543 |
+
"epoch": 1077.84,
|
| 269544 |
+
"learning_rate": 7.873285256410257e-06,
|
| 269545 |
+
"loss": 0.3629,
|
| 269546 |
+
"step": 133760
|
| 269547 |
+
},
|
| 269548 |
+
{
|
| 269549 |
+
"epoch": 1077.88,
|
| 269550 |
+
"learning_rate": 7.87320512820513e-06,
|
| 269551 |
+
"loss": 0.3166,
|
| 269552 |
+
"step": 133765
|
| 269553 |
+
},
|
| 269554 |
+
{
|
| 269555 |
+
"epoch": 1077.92,
|
| 269556 |
+
"learning_rate": 7.873125000000001e-06,
|
| 269557 |
+
"loss": 0.3112,
|
| 269558 |
+
"step": 133770
|
| 269559 |
+
},
|
| 269560 |
+
{
|
| 269561 |
+
"epoch": 1077.96,
|
| 269562 |
+
"learning_rate": 7.873044871794872e-06,
|
| 269563 |
+
"loss": 0.4494,
|
| 269564 |
+
"step": 133775
|
| 269565 |
+
},
|
| 269566 |
+
{
|
| 269567 |
+
"epoch": 1078.0,
|
| 269568 |
+
"eval_loss": 0.36935585737228394,
|
| 269569 |
+
"eval_runtime": 35.5506,
|
| 269570 |
+
"eval_samples_per_second": 23.488,
|
| 269571 |
+
"eval_steps_per_second": 0.759,
|
| 269572 |
+
"eval_wer": 0.18314665084192566,
|
| 269573 |
+
"step": 133779
|
| 269574 |
+
},
|
| 269575 |
+
{
|
| 269576 |
+
"epoch": 1078.01,
|
| 269577 |
+
"learning_rate": 7.872964743589744e-06,
|
| 269578 |
+
"loss": 0.3194,
|
| 269579 |
+
"step": 133780
|
| 269580 |
+
},
|
| 269581 |
+
{
|
| 269582 |
+
"epoch": 1078.05,
|
| 269583 |
+
"learning_rate": 7.872884615384617e-06,
|
| 269584 |
+
"loss": 0.2743,
|
| 269585 |
+
"step": 133785
|
| 269586 |
+
},
|
| 269587 |
+
{
|
| 269588 |
+
"epoch": 1078.09,
|
| 269589 |
+
"learning_rate": 7.872804487179488e-06,
|
| 269590 |
+
"loss": 0.249,
|
| 269591 |
+
"step": 133790
|
| 269592 |
+
},
|
| 269593 |
+
{
|
| 269594 |
+
"epoch": 1078.13,
|
| 269595 |
+
"learning_rate": 7.87272435897436e-06,
|
| 269596 |
+
"loss": 0.3227,
|
| 269597 |
+
"step": 133795
|
| 269598 |
+
},
|
| 269599 |
+
{
|
| 269600 |
+
"epoch": 1078.17,
|
| 269601 |
+
"learning_rate": 7.872644230769232e-06,
|
| 269602 |
+
"loss": 0.5637,
|
| 269603 |
+
"step": 133800
|
| 269604 |
+
},
|
| 269605 |
+
{
|
| 269606 |
+
"epoch": 1078.21,
|
| 269607 |
+
"learning_rate": 7.872564102564104e-06,
|
| 269608 |
+
"loss": 1.0401,
|
| 269609 |
+
"step": 133805
|
| 269610 |
+
},
|
| 269611 |
+
{
|
| 269612 |
+
"epoch": 1078.25,
|
| 269613 |
+
"learning_rate": 7.872483974358975e-06,
|
| 269614 |
+
"loss": 0.3121,
|
| 269615 |
+
"step": 133810
|
| 269616 |
+
},
|
| 269617 |
+
{
|
| 269618 |
+
"epoch": 1078.29,
|
| 269619 |
+
"learning_rate": 7.872403846153847e-06,
|
| 269620 |
+
"loss": 0.2549,
|
| 269621 |
+
"step": 133815
|
| 269622 |
+
},
|
| 269623 |
+
{
|
| 269624 |
+
"epoch": 1078.33,
|
| 269625 |
+
"learning_rate": 7.87232371794872e-06,
|
| 269626 |
+
"loss": 0.381,
|
| 269627 |
+
"step": 133820
|
| 269628 |
+
},
|
| 269629 |
+
{
|
| 269630 |
+
"epoch": 1078.37,
|
| 269631 |
+
"learning_rate": 7.87224358974359e-06,
|
| 269632 |
+
"loss": 0.4923,
|
| 269633 |
+
"step": 133825
|
| 269634 |
+
},
|
| 269635 |
+
{
|
| 269636 |
+
"epoch": 1078.41,
|
| 269637 |
+
"learning_rate": 7.872163461538462e-06,
|
| 269638 |
+
"loss": 1.0382,
|
| 269639 |
+
"step": 133830
|
| 269640 |
+
},
|
| 269641 |
+
{
|
| 269642 |
+
"epoch": 1078.45,
|
| 269643 |
+
"learning_rate": 7.872083333333334e-06,
|
| 269644 |
+
"loss": 0.2953,
|
| 269645 |
+
"step": 133835
|
| 269646 |
+
},
|
| 269647 |
+
{
|
| 269648 |
+
"epoch": 1078.49,
|
| 269649 |
+
"learning_rate": 7.872003205128205e-06,
|
| 269650 |
+
"loss": 0.3126,
|
| 269651 |
+
"step": 133840
|
| 269652 |
+
},
|
| 269653 |
+
{
|
| 269654 |
+
"epoch": 1078.53,
|
| 269655 |
+
"learning_rate": 7.871923076923078e-06,
|
| 269656 |
+
"loss": 0.3423,
|
| 269657 |
+
"step": 133845
|
| 269658 |
+
},
|
| 269659 |
+
{
|
| 269660 |
+
"epoch": 1078.57,
|
| 269661 |
+
"learning_rate": 7.87184294871795e-06,
|
| 269662 |
+
"loss": 0.5135,
|
| 269663 |
+
"step": 133850
|
| 269664 |
+
},
|
| 269665 |
+
{
|
| 269666 |
+
"epoch": 1078.61,
|
| 269667 |
+
"learning_rate": 7.87176282051282e-06,
|
| 269668 |
+
"loss": 0.8895,
|
| 269669 |
+
"step": 133855
|
| 269670 |
+
},
|
| 269671 |
+
{
|
| 269672 |
+
"epoch": 1078.65,
|
| 269673 |
+
"learning_rate": 7.871682692307692e-06,
|
| 269674 |
+
"loss": 0.2551,
|
| 269675 |
+
"step": 133860
|
| 269676 |
+
},
|
| 269677 |
+
{
|
| 269678 |
+
"epoch": 1078.69,
|
| 269679 |
+
"learning_rate": 7.871602564102565e-06,
|
| 269680 |
+
"loss": 0.2887,
|
| 269681 |
+
"step": 133865
|
| 269682 |
+
},
|
| 269683 |
+
{
|
| 269684 |
+
"epoch": 1078.73,
|
| 269685 |
+
"learning_rate": 7.871522435897437e-06,
|
| 269686 |
+
"loss": 0.3447,
|
| 269687 |
+
"step": 133870
|
| 269688 |
+
},
|
| 269689 |
+
{
|
| 269690 |
+
"epoch": 1078.77,
|
| 269691 |
+
"learning_rate": 7.871442307692308e-06,
|
| 269692 |
+
"loss": 0.4865,
|
| 269693 |
+
"step": 133875
|
| 269694 |
+
},
|
| 269695 |
+
{
|
| 269696 |
+
"epoch": 1078.81,
|
| 269697 |
+
"learning_rate": 7.87136217948718e-06,
|
| 269698 |
+
"loss": 1.1507,
|
| 269699 |
+
"step": 133880
|
| 269700 |
+
},
|
| 269701 |
+
{
|
| 269702 |
+
"epoch": 1078.85,
|
| 269703 |
+
"learning_rate": 7.871282051282052e-06,
|
| 269704 |
+
"loss": 0.3337,
|
| 269705 |
+
"step": 133885
|
| 269706 |
+
},
|
| 269707 |
+
{
|
| 269708 |
+
"epoch": 1078.89,
|
| 269709 |
+
"learning_rate": 7.871201923076924e-06,
|
| 269710 |
+
"loss": 0.3155,
|
| 269711 |
+
"step": 133890
|
| 269712 |
+
},
|
| 269713 |
+
{
|
| 269714 |
+
"epoch": 1078.93,
|
| 269715 |
+
"learning_rate": 7.871121794871795e-06,
|
| 269716 |
+
"loss": 0.3432,
|
| 269717 |
+
"step": 133895
|
| 269718 |
+
},
|
| 269719 |
+
{
|
| 269720 |
+
"epoch": 1078.97,
|
| 269721 |
+
"learning_rate": 7.871041666666668e-06,
|
| 269722 |
+
"loss": 0.5148,
|
| 269723 |
+
"step": 133900
|
| 269724 |
+
},
|
| 269725 |
+
{
|
| 269726 |
+
"epoch": 1079.0,
|
| 269727 |
+
"eval_loss": 0.41137564182281494,
|
| 269728 |
+
"eval_runtime": 35.9724,
|
| 269729 |
+
"eval_samples_per_second": 23.212,
|
| 269730 |
+
"eval_steps_per_second": 0.751,
|
| 269731 |
+
"eval_wer": 0.17041547277936964,
|
| 269732 |
+
"step": 133903
|
| 269733 |
+
},
|
| 269734 |
+
{
|
| 269735 |
+
"epoch": 1071.02,
|
| 269736 |
+
"learning_rate": 7.87096153846154e-06,
|
| 269737 |
+
"loss": 0.3643,
|
| 269738 |
+
"step": 133905
|
| 269739 |
+
},
|
| 269740 |
+
{
|
| 269741 |
+
"epoch": 1071.06,
|
| 269742 |
+
"learning_rate": 7.870881410256411e-06,
|
| 269743 |
+
"loss": 0.3349,
|
| 269744 |
+
"step": 133910
|
| 269745 |
+
},
|
| 269746 |
+
{
|
| 269747 |
+
"epoch": 1071.1,
|
| 269748 |
+
"learning_rate": 7.870801282051282e-06,
|
| 269749 |
+
"loss": 0.275,
|
| 269750 |
+
"step": 133915
|
| 269751 |
+
},
|
| 269752 |
+
{
|
| 269753 |
+
"epoch": 1071.14,
|
| 269754 |
+
"learning_rate": 7.870721153846155e-06,
|
| 269755 |
+
"loss": 0.303,
|
| 269756 |
+
"step": 133920
|
| 269757 |
+
},
|
| 269758 |
+
{
|
| 269759 |
+
"epoch": 1071.18,
|
| 269760 |
+
"learning_rate": 7.870641025641027e-06,
|
| 269761 |
+
"loss": 0.5914,
|
| 269762 |
+
"step": 133925
|
| 269763 |
+
},
|
| 269764 |
+
{
|
| 269765 |
+
"epoch": 1071.22,
|
| 269766 |
+
"learning_rate": 7.870560897435898e-06,
|
| 269767 |
+
"loss": 1.0265,
|
| 269768 |
+
"step": 133930
|
| 269769 |
+
},
|
| 269770 |
+
{
|
| 269771 |
+
"epoch": 1071.26,
|
| 269772 |
+
"learning_rate": 7.87048076923077e-06,
|
| 269773 |
+
"loss": 0.2939,
|
| 269774 |
+
"step": 133935
|
| 269775 |
+
},
|
| 269776 |
+
{
|
| 269777 |
+
"epoch": 1071.3,
|
| 269778 |
+
"learning_rate": 7.870400641025642e-06,
|
| 269779 |
+
"loss": 0.2298,
|
| 269780 |
+
"step": 133940
|
| 269781 |
+
},
|
| 269782 |
+
{
|
| 269783 |
+
"epoch": 1071.34,
|
| 269784 |
+
"learning_rate": 7.870320512820514e-06,
|
| 269785 |
+
"loss": 0.2893,
|
| 269786 |
+
"step": 133945
|
| 269787 |
+
},
|
| 269788 |
+
{
|
| 269789 |
+
"epoch": 1071.38,
|
| 269790 |
+
"learning_rate": 7.870240384615385e-06,
|
| 269791 |
+
"loss": 0.5808,
|
| 269792 |
+
"step": 133950
|
| 269793 |
+
},
|
| 269794 |
+
{
|
| 269795 |
+
"epoch": 1071.42,
|
| 269796 |
+
"learning_rate": 7.870160256410258e-06,
|
| 269797 |
+
"loss": 0.9685,
|
| 269798 |
+
"step": 133955
|
| 269799 |
+
},
|
| 269800 |
+
{
|
| 269801 |
+
"epoch": 1071.46,
|
| 269802 |
+
"learning_rate": 7.87008012820513e-06,
|
| 269803 |
+
"loss": 0.2692,
|
| 269804 |
+
"step": 133960
|
| 269805 |
+
},
|
| 269806 |
+
{
|
| 269807 |
+
"epoch": 1071.5,
|
| 269808 |
+
"learning_rate": 7.870000000000001e-06,
|
| 269809 |
+
"loss": 0.2342,
|
| 269810 |
+
"step": 133965
|
| 269811 |
+
},
|
| 269812 |
+
{
|
| 269813 |
+
"epoch": 1071.54,
|
| 269814 |
+
"learning_rate": 7.869919871794872e-06,
|
| 269815 |
+
"loss": 0.3344,
|
| 269816 |
+
"step": 133970
|
| 269817 |
+
},
|
| 269818 |
+
{
|
| 269819 |
+
"epoch": 1071.58,
|
| 269820 |
+
"learning_rate": 7.869839743589745e-06,
|
| 269821 |
+
"loss": 0.5782,
|
| 269822 |
+
"step": 133975
|
| 269823 |
+
},
|
| 269824 |
+
{
|
| 269825 |
+
"epoch": 1071.62,
|
| 269826 |
+
"learning_rate": 7.869759615384615e-06,
|
| 269827 |
+
"loss": 0.9415,
|
| 269828 |
+
"step": 133980
|
| 269829 |
+
},
|
| 269830 |
+
{
|
| 269831 |
+
"epoch": 1071.66,
|
| 269832 |
+
"learning_rate": 7.869679487179488e-06,
|
| 269833 |
+
"loss": 0.248,
|
| 269834 |
+
"step": 133985
|
| 269835 |
+
},
|
| 269836 |
+
{
|
| 269837 |
+
"epoch": 1071.7,
|
| 269838 |
+
"learning_rate": 7.86959935897436e-06,
|
| 269839 |
+
"loss": 0.2807,
|
| 269840 |
+
"step": 133990
|
| 269841 |
+
},
|
| 269842 |
+
{
|
| 269843 |
+
"epoch": 1071.74,
|
| 269844 |
+
"learning_rate": 7.86951923076923e-06,
|
| 269845 |
+
"loss": 0.3315,
|
| 269846 |
+
"step": 133995
|
| 269847 |
+
},
|
| 269848 |
+
{
|
| 269849 |
+
"epoch": 1071.78,
|
| 269850 |
+
"learning_rate": 7.869439102564104e-06,
|
| 269851 |
+
"loss": 0.6116,
|
| 269852 |
+
"step": 134000
|
| 269853 |
+
},
|
| 269854 |
+
{
|
| 269855 |
+
"epoch": 1071.82,
|
| 269856 |
+
"learning_rate": 7.869358974358975e-06,
|
| 269857 |
+
"loss": 0.9509,
|
| 269858 |
+
"step": 134005
|
| 269859 |
+
},
|
| 269860 |
+
{
|
| 269861 |
+
"epoch": 1071.86,
|
| 269862 |
+
"learning_rate": 7.869278846153846e-06,
|
| 269863 |
+
"loss": 0.2919,
|
| 269864 |
+
"step": 134010
|
| 269865 |
+
},
|
| 269866 |
+
{
|
| 269867 |
+
"epoch": 1071.9,
|
| 269868 |
+
"learning_rate": 7.869198717948718e-06,
|
| 269869 |
+
"loss": 0.296,
|
| 269870 |
+
"step": 134015
|
| 269871 |
+
},
|
| 269872 |
+
{
|
| 269873 |
+
"epoch": 1071.94,
|
| 269874 |
+
"learning_rate": 7.869118589743591e-06,
|
| 269875 |
+
"loss": 0.3532,
|
| 269876 |
+
"step": 134020
|
| 269877 |
+
},
|
| 269878 |
+
{
|
| 269879 |
+
"epoch": 1071.98,
|
| 269880 |
+
"learning_rate": 7.869038461538462e-06,
|
| 269881 |
+
"loss": 0.5742,
|
| 269882 |
+
"step": 134025
|
| 269883 |
+
},
|
| 269884 |
+
{
|
| 269885 |
+
"epoch": 1072.0,
|
| 269886 |
+
"eval_loss": 0.3423796594142914,
|
| 269887 |
+
"eval_runtime": 35.6088,
|
| 269888 |
+
"eval_samples_per_second": 23.449,
|
| 269889 |
+
"eval_steps_per_second": 0.758,
|
| 269890 |
+
"eval_wer": 0.17097562782428327,
|
| 269891 |
+
"step": 134028
|
| 269892 |
+
},
|
| 269893 |
+
{
|
| 269894 |
+
"epoch": 1080.02,
|
| 269895 |
+
"learning_rate": 7.868958333333334e-06,
|
| 269896 |
+
"loss": 0.2569,
|
| 269897 |
+
"step": 134030
|
| 269898 |
+
},
|
| 269899 |
+
{
|
| 269900 |
+
"epoch": 1080.06,
|
| 269901 |
+
"learning_rate": 7.868878205128205e-06,
|
| 269902 |
+
"loss": 0.3096,
|
| 269903 |
+
"step": 134035
|
| 269904 |
+
},
|
| 269905 |
+
{
|
| 269906 |
+
"epoch": 1080.1,
|
| 269907 |
+
"learning_rate": 7.868798076923078e-06,
|
| 269908 |
+
"loss": 0.2902,
|
| 269909 |
+
"step": 134040
|
| 269910 |
+
},
|
| 269911 |
+
{
|
| 269912 |
+
"epoch": 1080.14,
|
| 269913 |
+
"learning_rate": 7.86871794871795e-06,
|
| 269914 |
+
"loss": 0.3467,
|
| 269915 |
+
"step": 134045
|
| 269916 |
+
},
|
| 269917 |
+
{
|
| 269918 |
+
"epoch": 1080.18,
|
| 269919 |
+
"learning_rate": 7.86863782051282e-06,
|
| 269920 |
+
"loss": 0.5397,
|
| 269921 |
+
"step": 134050
|
| 269922 |
+
},
|
| 269923 |
+
{
|
| 269924 |
+
"epoch": 1080.22,
|
| 269925 |
+
"learning_rate": 7.868557692307694e-06,
|
| 269926 |
+
"loss": 1.0189,
|
| 269927 |
+
"step": 134055
|
| 269928 |
+
},
|
| 269929 |
+
{
|
| 269930 |
+
"epoch": 1080.26,
|
| 269931 |
+
"learning_rate": 7.868477564102565e-06,
|
| 269932 |
+
"loss": 0.2523,
|
| 269933 |
+
"step": 134060
|
| 269934 |
+
},
|
| 269935 |
+
{
|
| 269936 |
+
"epoch": 1080.3,
|
| 269937 |
+
"learning_rate": 7.868397435897437e-06,
|
| 269938 |
+
"loss": 0.2536,
|
| 269939 |
+
"step": 134065
|
| 269940 |
+
},
|
| 269941 |
+
{
|
| 269942 |
+
"epoch": 1080.34,
|
| 269943 |
+
"learning_rate": 7.868317307692308e-06,
|
| 269944 |
+
"loss": 0.3514,
|
| 269945 |
+
"step": 134070
|
| 269946 |
+
},
|
| 269947 |
+
{
|
| 269948 |
+
"epoch": 1080.38,
|
| 269949 |
+
"learning_rate": 7.868237179487181e-06,
|
| 269950 |
+
"loss": 0.7254,
|
| 269951 |
+
"step": 134075
|
| 269952 |
+
},
|
| 269953 |
+
{
|
| 269954 |
+
"epoch": 1080.42,
|
| 269955 |
+
"learning_rate": 7.868157051282052e-06,
|
| 269956 |
+
"loss": 1.1177,
|
| 269957 |
+
"step": 134080
|
| 269958 |
+
},
|
| 269959 |
+
{
|
| 269960 |
+
"epoch": 1080.46,
|
| 269961 |
+
"learning_rate": 7.868076923076924e-06,
|
| 269962 |
+
"loss": 0.2958,
|
| 269963 |
+
"step": 134085
|
| 269964 |
+
},
|
| 269965 |
+
{
|
| 269966 |
+
"epoch": 1080.5,
|
| 269967 |
+
"learning_rate": 7.867996794871795e-06,
|
| 269968 |
+
"loss": 0.2778,
|
| 269969 |
+
"step": 134090
|
| 269970 |
+
},
|
| 269971 |
+
{
|
| 269972 |
+
"epoch": 1080.54,
|
| 269973 |
+
"learning_rate": 7.867916666666668e-06,
|
| 269974 |
+
"loss": 0.3512,
|
| 269975 |
+
"step": 134095
|
| 269976 |
+
},
|
| 269977 |
+
{
|
| 269978 |
+
"epoch": 1080.58,
|
| 269979 |
+
"learning_rate": 7.86783653846154e-06,
|
| 269980 |
+
"loss": 0.555,
|
| 269981 |
+
"step": 134100
|
| 269982 |
+
},
|
| 269983 |
+
{
|
| 269984 |
+
"epoch": 1080.62,
|
| 269985 |
+
"learning_rate": 7.86775641025641e-06,
|
| 269986 |
+
"loss": 1.1554,
|
| 269987 |
+
"step": 134105
|
| 269988 |
+
},
|
| 269989 |
+
{
|
| 269990 |
+
"epoch": 1080.66,
|
| 269991 |
+
"learning_rate": 7.867676282051284e-06,
|
| 269992 |
+
"loss": 0.2772,
|
| 269993 |
+
"step": 134110
|
| 269994 |
+
},
|
| 269995 |
+
{
|
| 269996 |
+
"epoch": 1080.7,
|
| 269997 |
+
"learning_rate": 7.867596153846153e-06,
|
| 269998 |
+
"loss": 0.3057,
|
| 269999 |
+
"step": 134115
|
| 270000 |
+
},
|
| 270001 |
+
{
|
| 270002 |
+
"epoch": 1080.74,
|
| 270003 |
+
"learning_rate": 7.867516025641027e-06,
|
| 270004 |
+
"loss": 0.3324,
|
| 270005 |
+
"step": 134120
|
| 270006 |
+
},
|
| 270007 |
+
{
|
| 270008 |
+
"epoch": 1080.78,
|
| 270009 |
+
"learning_rate": 7.867435897435898e-06,
|
| 270010 |
+
"loss": 0.5588,
|
| 270011 |
+
"step": 134125
|
| 270012 |
+
},
|
| 270013 |
+
{
|
| 270014 |
+
"epoch": 1080.82,
|
| 270015 |
+
"learning_rate": 7.86735576923077e-06,
|
| 270016 |
+
"loss": 1.137,
|
| 270017 |
+
"step": 134130
|
| 270018 |
+
},
|
| 270019 |
+
{
|
| 270020 |
+
"epoch": 1080.86,
|
| 270021 |
+
"learning_rate": 7.86727564102564e-06,
|
| 270022 |
+
"loss": 0.293,
|
| 270023 |
+
"step": 134135
|
| 270024 |
+
},
|
| 270025 |
+
{
|
| 270026 |
+
"epoch": 1080.9,
|
| 270027 |
+
"learning_rate": 7.867195512820514e-06,
|
| 270028 |
+
"loss": 0.3207,
|
| 270029 |
+
"step": 134140
|
| 270030 |
+
},
|
| 270031 |
+
{
|
| 270032 |
+
"epoch": 1080.94,
|
| 270033 |
+
"learning_rate": 7.867115384615385e-06,
|
| 270034 |
+
"loss": 0.3501,
|
| 270035 |
+
"step": 134145
|
| 270036 |
+
},
|
| 270037 |
+
{
|
| 270038 |
+
"epoch": 1080.98,
|
| 270039 |
+
"learning_rate": 7.867035256410256e-06,
|
| 270040 |
+
"loss": 0.767,
|
| 270041 |
+
"step": 134150
|
| 270042 |
+
},
|
| 270043 |
+
{
|
| 270044 |
+
"epoch": 1081.0,
|
| 270045 |
+
"eval_loss": 0.3516015112400055,
|
| 270046 |
+
"eval_runtime": 36.4198,
|
| 270047 |
+
"eval_samples_per_second": 22.955,
|
| 270048 |
+
"eval_steps_per_second": 0.741,
|
| 270049 |
+
"eval_wer": 0.1819801839047687,
|
| 270050 |
+
"step": 134152
|
| 270051 |
}
|
| 270052 |
],
|
| 270053 |
"max_steps": 620000,
|
| 270054 |
"num_train_epochs": 5000,
|
| 270055 |
+
"total_flos": 3.774979260580794e+20,
|
| 270056 |
"trial_name": null,
|
| 270057 |
"trial_params": null
|
| 270058 |
}
|
model-bin/finetune/base/{checkpoint-133530 β checkpoint-134152}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1630173460.284948/events.out.tfevents.1630173460.86bb0ddabf9b.4092.251
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0663966003fa02b1812f8ba56338d9555f48211f3ecd478b8106d78daf474d5b
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630173847.4052565/events.out.tfevents.1630173847.86bb0ddabf9b.4092.253
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9cb6219d11e7906000bd3e22599a21df47651f33dcfbca94e516cf096d20a6a6
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630174238.155075/events.out.tfevents.1630174238.86bb0ddabf9b.4092.255
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:973f26703749c96d6fcdbb097e4c200d3f91a76c8ea263a3c5edbe0eda1642f6
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630174625.3833299/events.out.tfevents.1630174625.86bb0ddabf9b.4092.257
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:737ac480260f8043607fa3b7050dbcdf16b58f43eb25d387e4ae969f04731901
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630175012.849805/events.out.tfevents.1630175012.86bb0ddabf9b.4092.259
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d4b516171f2fe94023507ec5aecab580ebabc50b6f2dc8f6b69dd303220497e3
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630173460.86bb0ddabf9b.4092.250
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9901c68ba5c8ffdc6da733b931c8ce2c57b123cb5df451387131a533d1cac86e
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630173847.86bb0ddabf9b.4092.252
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6f14716706d90ec7f61cdff14c85a16441331bcfb3b998d2083754aa3ae8f48a
|
| 3 |
+
size 8462
|
model-bin/finetune/base/log/events.out.tfevents.1630174238.86bb0ddabf9b.4092.254
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4b2fcb555e8f4671789c82bc10264e8283e02e41a6b60b5998b91e24606e47fd
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630174625.86bb0ddabf9b.4092.256
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:78813b6ecc3938169f66820ae6701faff5aaa82a4c2eb6ea883c89ec54649874
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630175012.86bb0ddabf9b.4092.258
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c78bcf475dbf4a8089057e2b72d8e4ec4cc76b0860f78a34b849bf98a3c334b9
|
| 3 |
+
size 8622
|