"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-70809 β checkpoint-71429}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-70809 β checkpoint-71429}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-70809 β checkpoint-71429}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-70809 β checkpoint-71429}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-70809 β checkpoint-71429}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-70809 β checkpoint-71429}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-70809 β checkpoint-71429}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-70809 β checkpoint-71429}/trainer_state.json +792 -3
- model-bin/finetune/base/{checkpoint-70809 β checkpoint-71429}/training_args.bin +0 -0
- model-bin/finetune/base/log/1629858970.8441613/events.out.tfevents.1629858970.7e498afd5545.905.23 +3 -0
- model-bin/finetune/base/log/1629859624.9463222/events.out.tfevents.1629859624.7e498afd5545.905.25 +3 -0
- model-bin/finetune/base/log/1629860278.0745044/events.out.tfevents.1629860278.7e498afd5545.905.27 +3 -0
- model-bin/finetune/base/log/1629860932.5132222/events.out.tfevents.1629860932.7e498afd5545.905.29 +3 -0
- model-bin/finetune/base/log/1629861583.8800406/events.out.tfevents.1629861583.7e498afd5545.905.31 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629858970.7e498afd5545.905.22 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629859624.7e498afd5545.905.24 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629860278.7e498afd5545.905.26 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629860932.7e498afd5545.905.28 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629861583.7e498afd5545.905.30 +3 -0
model-bin/finetune/base/{checkpoint-70809 β checkpoint-71429}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-70809 β checkpoint-71429}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1c2cf74949ac6fbe57789fc93a1ba947fc6fea6315730c11480f522422c2cb32
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-70809 β checkpoint-71429}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-70809 β checkpoint-71429}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b94fe687400b28878fb9b1535714b9e18e88910b1486e085968c125fe165b608
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-70809 β checkpoint-71429}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a44b4a00afd33832285479c87b3d0330ccd6122276d09600764d244ca2e0d676
|
| 3 |
size 14503
|
model-bin/finetune/base/{checkpoint-70809 β checkpoint-71429}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fa5ffd5ce4a042224436871219db3b89c25b62d29b4a4605aafd1e3b9552f733
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-70809 β checkpoint-71429}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:34abd4dd1bca145f62df90529f947d27c07749481bac26350eaca06e62acd0fe
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-70809 β checkpoint-71429}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.18412114350410416,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-69565",
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -189453,11 +189453,800 @@
|
|
| 189453 |
"eval_steps_per_second": 0.673,
|
| 189454 |
"eval_wer": 0.18833597008485545,
|
| 189455 |
"step": 70809
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 189456 |
}
|
| 189457 |
],
|
| 189458 |
"max_steps": 620000,
|
| 189459 |
"num_train_epochs": 5000,
|
| 189460 |
-
"total_flos":
|
| 189461 |
"trial_name": null,
|
| 189462 |
"trial_params": null
|
| 189463 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.18412114350410416,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-69565",
|
| 4 |
+
"epoch": 575.995983935743,
|
| 5 |
+
"global_step": 71429,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 189453 |
"eval_steps_per_second": 0.673,
|
| 189454 |
"eval_wer": 0.18833597008485545,
|
| 189455 |
"step": 70809
|
| 189456 |
+
},
|
| 189457 |
+
{
|
| 189458 |
+
"epoch": 571.01,
|
| 189459 |
+
"learning_rate": 8.881666666666667e-06,
|
| 189460 |
+
"loss": 0.3513,
|
| 189461 |
+
"step": 70810
|
| 189462 |
+
},
|
| 189463 |
+
{
|
| 189464 |
+
"epoch": 571.05,
|
| 189465 |
+
"learning_rate": 8.88158653846154e-06,
|
| 189466 |
+
"loss": 0.2739,
|
| 189467 |
+
"step": 70815
|
| 189468 |
+
},
|
| 189469 |
+
{
|
| 189470 |
+
"epoch": 571.09,
|
| 189471 |
+
"learning_rate": 8.881506410256411e-06,
|
| 189472 |
+
"loss": 0.3412,
|
| 189473 |
+
"step": 70820
|
| 189474 |
+
},
|
| 189475 |
+
{
|
| 189476 |
+
"epoch": 571.13,
|
| 189477 |
+
"learning_rate": 8.881426282051282e-06,
|
| 189478 |
+
"loss": 0.3873,
|
| 189479 |
+
"step": 70825
|
| 189480 |
+
},
|
| 189481 |
+
{
|
| 189482 |
+
"epoch": 571.17,
|
| 189483 |
+
"learning_rate": 8.881346153846154e-06,
|
| 189484 |
+
"loss": 0.5661,
|
| 189485 |
+
"step": 70830
|
| 189486 |
+
},
|
| 189487 |
+
{
|
| 189488 |
+
"epoch": 571.21,
|
| 189489 |
+
"learning_rate": 8.881266025641027e-06,
|
| 189490 |
+
"loss": 1.1271,
|
| 189491 |
+
"step": 70835
|
| 189492 |
+
},
|
| 189493 |
+
{
|
| 189494 |
+
"epoch": 571.25,
|
| 189495 |
+
"learning_rate": 8.881185897435898e-06,
|
| 189496 |
+
"loss": 0.3545,
|
| 189497 |
+
"step": 70840
|
| 189498 |
+
},
|
| 189499 |
+
{
|
| 189500 |
+
"epoch": 571.29,
|
| 189501 |
+
"learning_rate": 8.88110576923077e-06,
|
| 189502 |
+
"loss": 0.2882,
|
| 189503 |
+
"step": 70845
|
| 189504 |
+
},
|
| 189505 |
+
{
|
| 189506 |
+
"epoch": 571.33,
|
| 189507 |
+
"learning_rate": 8.881025641025641e-06,
|
| 189508 |
+
"loss": 0.352,
|
| 189509 |
+
"step": 70850
|
| 189510 |
+
},
|
| 189511 |
+
{
|
| 189512 |
+
"epoch": 571.37,
|
| 189513 |
+
"learning_rate": 8.880945512820514e-06,
|
| 189514 |
+
"loss": 0.5543,
|
| 189515 |
+
"step": 70855
|
| 189516 |
+
},
|
| 189517 |
+
{
|
| 189518 |
+
"epoch": 571.41,
|
| 189519 |
+
"learning_rate": 8.880865384615385e-06,
|
| 189520 |
+
"loss": 1.203,
|
| 189521 |
+
"step": 70860
|
| 189522 |
+
},
|
| 189523 |
+
{
|
| 189524 |
+
"epoch": 571.45,
|
| 189525 |
+
"learning_rate": 8.880785256410257e-06,
|
| 189526 |
+
"loss": 0.3559,
|
| 189527 |
+
"step": 70865
|
| 189528 |
+
},
|
| 189529 |
+
{
|
| 189530 |
+
"epoch": 571.49,
|
| 189531 |
+
"learning_rate": 8.88070512820513e-06,
|
| 189532 |
+
"loss": 0.3581,
|
| 189533 |
+
"step": 70870
|
| 189534 |
+
},
|
| 189535 |
+
{
|
| 189536 |
+
"epoch": 571.53,
|
| 189537 |
+
"learning_rate": 8.880625000000001e-06,
|
| 189538 |
+
"loss": 0.3475,
|
| 189539 |
+
"step": 70875
|
| 189540 |
+
},
|
| 189541 |
+
{
|
| 189542 |
+
"epoch": 571.57,
|
| 189543 |
+
"learning_rate": 8.880544871794872e-06,
|
| 189544 |
+
"loss": 0.6108,
|
| 189545 |
+
"step": 70880
|
| 189546 |
+
},
|
| 189547 |
+
{
|
| 189548 |
+
"epoch": 571.61,
|
| 189549 |
+
"learning_rate": 8.880464743589744e-06,
|
| 189550 |
+
"loss": 1.0645,
|
| 189551 |
+
"step": 70885
|
| 189552 |
+
},
|
| 189553 |
+
{
|
| 189554 |
+
"epoch": 571.65,
|
| 189555 |
+
"learning_rate": 8.880384615384617e-06,
|
| 189556 |
+
"loss": 0.4175,
|
| 189557 |
+
"step": 70890
|
| 189558 |
+
},
|
| 189559 |
+
{
|
| 189560 |
+
"epoch": 571.69,
|
| 189561 |
+
"learning_rate": 8.880304487179488e-06,
|
| 189562 |
+
"loss": 0.3068,
|
| 189563 |
+
"step": 70895
|
| 189564 |
+
},
|
| 189565 |
+
{
|
| 189566 |
+
"epoch": 571.73,
|
| 189567 |
+
"learning_rate": 8.88022435897436e-06,
|
| 189568 |
+
"loss": 0.3434,
|
| 189569 |
+
"step": 70900
|
| 189570 |
+
},
|
| 189571 |
+
{
|
| 189572 |
+
"epoch": 571.77,
|
| 189573 |
+
"learning_rate": 8.880144230769233e-06,
|
| 189574 |
+
"loss": 0.5246,
|
| 189575 |
+
"step": 70905
|
| 189576 |
+
},
|
| 189577 |
+
{
|
| 189578 |
+
"epoch": 571.81,
|
| 189579 |
+
"learning_rate": 8.880064102564104e-06,
|
| 189580 |
+
"loss": 1.2047,
|
| 189581 |
+
"step": 70910
|
| 189582 |
+
},
|
| 189583 |
+
{
|
| 189584 |
+
"epoch": 571.85,
|
| 189585 |
+
"learning_rate": 8.879983974358975e-06,
|
| 189586 |
+
"loss": 0.3039,
|
| 189587 |
+
"step": 70915
|
| 189588 |
+
},
|
| 189589 |
+
{
|
| 189590 |
+
"epoch": 571.9,
|
| 189591 |
+
"learning_rate": 8.879903846153847e-06,
|
| 189592 |
+
"loss": 0.3366,
|
| 189593 |
+
"step": 70920
|
| 189594 |
+
},
|
| 189595 |
+
{
|
| 189596 |
+
"epoch": 571.94,
|
| 189597 |
+
"learning_rate": 8.87982371794872e-06,
|
| 189598 |
+
"loss": 0.412,
|
| 189599 |
+
"step": 70925
|
| 189600 |
+
},
|
| 189601 |
+
{
|
| 189602 |
+
"epoch": 571.98,
|
| 189603 |
+
"learning_rate": 8.87974358974359e-06,
|
| 189604 |
+
"loss": 0.6262,
|
| 189605 |
+
"step": 70930
|
| 189606 |
+
},
|
| 189607 |
+
{
|
| 189608 |
+
"epoch": 572.0,
|
| 189609 |
+
"eval_loss": 0.37808266282081604,
|
| 189610 |
+
"eval_runtime": 39.9883,
|
| 189611 |
+
"eval_samples_per_second": 21.006,
|
| 189612 |
+
"eval_steps_per_second": 0.675,
|
| 189613 |
+
"eval_wer": 0.1896881435716387,
|
| 189614 |
+
"step": 70933
|
| 189615 |
+
},
|
| 189616 |
+
{
|
| 189617 |
+
"epoch": 572.02,
|
| 189618 |
+
"learning_rate": 8.879663461538462e-06,
|
| 189619 |
+
"loss": 0.4067,
|
| 189620 |
+
"step": 70935
|
| 189621 |
+
},
|
| 189622 |
+
{
|
| 189623 |
+
"epoch": 572.06,
|
| 189624 |
+
"learning_rate": 8.879583333333334e-06,
|
| 189625 |
+
"loss": 0.3122,
|
| 189626 |
+
"step": 70940
|
| 189627 |
+
},
|
| 189628 |
+
{
|
| 189629 |
+
"epoch": 572.1,
|
| 189630 |
+
"learning_rate": 8.879503205128205e-06,
|
| 189631 |
+
"loss": 0.3144,
|
| 189632 |
+
"step": 70945
|
| 189633 |
+
},
|
| 189634 |
+
{
|
| 189635 |
+
"epoch": 572.14,
|
| 189636 |
+
"learning_rate": 8.879423076923077e-06,
|
| 189637 |
+
"loss": 0.3966,
|
| 189638 |
+
"step": 70950
|
| 189639 |
+
},
|
| 189640 |
+
{
|
| 189641 |
+
"epoch": 572.18,
|
| 189642 |
+
"learning_rate": 8.87934294871795e-06,
|
| 189643 |
+
"loss": 0.7134,
|
| 189644 |
+
"step": 70955
|
| 189645 |
+
},
|
| 189646 |
+
{
|
| 189647 |
+
"epoch": 572.22,
|
| 189648 |
+
"learning_rate": 8.879262820512821e-06,
|
| 189649 |
+
"loss": 1.0855,
|
| 189650 |
+
"step": 70960
|
| 189651 |
+
},
|
| 189652 |
+
{
|
| 189653 |
+
"epoch": 572.26,
|
| 189654 |
+
"learning_rate": 8.879182692307692e-06,
|
| 189655 |
+
"loss": 0.2845,
|
| 189656 |
+
"step": 70965
|
| 189657 |
+
},
|
| 189658 |
+
{
|
| 189659 |
+
"epoch": 572.3,
|
| 189660 |
+
"learning_rate": 8.879102564102565e-06,
|
| 189661 |
+
"loss": 0.3495,
|
| 189662 |
+
"step": 70970
|
| 189663 |
+
},
|
| 189664 |
+
{
|
| 189665 |
+
"epoch": 572.34,
|
| 189666 |
+
"learning_rate": 8.879022435897437e-06,
|
| 189667 |
+
"loss": 0.325,
|
| 189668 |
+
"step": 70975
|
| 189669 |
+
},
|
| 189670 |
+
{
|
| 189671 |
+
"epoch": 572.38,
|
| 189672 |
+
"learning_rate": 8.878942307692308e-06,
|
| 189673 |
+
"loss": 0.6465,
|
| 189674 |
+
"step": 70980
|
| 189675 |
+
},
|
| 189676 |
+
{
|
| 189677 |
+
"epoch": 572.42,
|
| 189678 |
+
"learning_rate": 8.87886217948718e-06,
|
| 189679 |
+
"loss": 1.0453,
|
| 189680 |
+
"step": 70985
|
| 189681 |
+
},
|
| 189682 |
+
{
|
| 189683 |
+
"epoch": 572.46,
|
| 189684 |
+
"learning_rate": 8.878782051282052e-06,
|
| 189685 |
+
"loss": 0.3642,
|
| 189686 |
+
"step": 70990
|
| 189687 |
+
},
|
| 189688 |
+
{
|
| 189689 |
+
"epoch": 572.5,
|
| 189690 |
+
"learning_rate": 8.878701923076924e-06,
|
| 189691 |
+
"loss": 0.3005,
|
| 189692 |
+
"step": 70995
|
| 189693 |
+
},
|
| 189694 |
+
{
|
| 189695 |
+
"epoch": 572.54,
|
| 189696 |
+
"learning_rate": 8.878621794871795e-06,
|
| 189697 |
+
"loss": 0.3992,
|
| 189698 |
+
"step": 71000
|
| 189699 |
+
},
|
| 189700 |
+
{
|
| 189701 |
+
"epoch": 572.58,
|
| 189702 |
+
"learning_rate": 8.878541666666668e-06,
|
| 189703 |
+
"loss": 0.6974,
|
| 189704 |
+
"step": 71005
|
| 189705 |
+
},
|
| 189706 |
+
{
|
| 189707 |
+
"epoch": 572.62,
|
| 189708 |
+
"learning_rate": 8.87846153846154e-06,
|
| 189709 |
+
"loss": 0.9939,
|
| 189710 |
+
"step": 71010
|
| 189711 |
+
},
|
| 189712 |
+
{
|
| 189713 |
+
"epoch": 572.66,
|
| 189714 |
+
"learning_rate": 8.878381410256411e-06,
|
| 189715 |
+
"loss": 0.3028,
|
| 189716 |
+
"step": 71015
|
| 189717 |
+
},
|
| 189718 |
+
{
|
| 189719 |
+
"epoch": 572.7,
|
| 189720 |
+
"learning_rate": 8.878301282051282e-06,
|
| 189721 |
+
"loss": 0.3952,
|
| 189722 |
+
"step": 71020
|
| 189723 |
+
},
|
| 189724 |
+
{
|
| 189725 |
+
"epoch": 572.74,
|
| 189726 |
+
"learning_rate": 8.878221153846155e-06,
|
| 189727 |
+
"loss": 0.3883,
|
| 189728 |
+
"step": 71025
|
| 189729 |
+
},
|
| 189730 |
+
{
|
| 189731 |
+
"epoch": 572.78,
|
| 189732 |
+
"learning_rate": 8.878141025641027e-06,
|
| 189733 |
+
"loss": 0.6341,
|
| 189734 |
+
"step": 71030
|
| 189735 |
+
},
|
| 189736 |
+
{
|
| 189737 |
+
"epoch": 572.82,
|
| 189738 |
+
"learning_rate": 8.878060897435898e-06,
|
| 189739 |
+
"loss": 1.0706,
|
| 189740 |
+
"step": 71035
|
| 189741 |
+
},
|
| 189742 |
+
{
|
| 189743 |
+
"epoch": 572.86,
|
| 189744 |
+
"learning_rate": 8.87798076923077e-06,
|
| 189745 |
+
"loss": 0.2861,
|
| 189746 |
+
"step": 71040
|
| 189747 |
+
},
|
| 189748 |
+
{
|
| 189749 |
+
"epoch": 572.9,
|
| 189750 |
+
"learning_rate": 8.877900641025642e-06,
|
| 189751 |
+
"loss": 0.4219,
|
| 189752 |
+
"step": 71045
|
| 189753 |
+
},
|
| 189754 |
+
{
|
| 189755 |
+
"epoch": 572.94,
|
| 189756 |
+
"learning_rate": 8.877820512820514e-06,
|
| 189757 |
+
"loss": 0.3828,
|
| 189758 |
+
"step": 71050
|
| 189759 |
+
},
|
| 189760 |
+
{
|
| 189761 |
+
"epoch": 572.98,
|
| 189762 |
+
"learning_rate": 8.877740384615385e-06,
|
| 189763 |
+
"loss": 0.6617,
|
| 189764 |
+
"step": 71055
|
| 189765 |
+
},
|
| 189766 |
+
{
|
| 189767 |
+
"epoch": 573.0,
|
| 189768 |
+
"eval_loss": 0.37939906120300293,
|
| 189769 |
+
"eval_runtime": 39.3467,
|
| 189770 |
+
"eval_samples_per_second": 21.349,
|
| 189771 |
+
"eval_steps_per_second": 0.686,
|
| 189772 |
+
"eval_wer": 0.19251534649804009,
|
| 189773 |
+
"step": 71057
|
| 189774 |
+
},
|
| 189775 |
+
{
|
| 189776 |
+
"epoch": 573.02,
|
| 189777 |
+
"learning_rate": 8.877660256410258e-06,
|
| 189778 |
+
"loss": 0.3238,
|
| 189779 |
+
"step": 71060
|
| 189780 |
+
},
|
| 189781 |
+
{
|
| 189782 |
+
"epoch": 573.06,
|
| 189783 |
+
"learning_rate": 8.87758012820513e-06,
|
| 189784 |
+
"loss": 0.3552,
|
| 189785 |
+
"step": 71065
|
| 189786 |
+
},
|
| 189787 |
+
{
|
| 189788 |
+
"epoch": 573.1,
|
| 189789 |
+
"learning_rate": 8.877500000000001e-06,
|
| 189790 |
+
"loss": 0.3396,
|
| 189791 |
+
"step": 71070
|
| 189792 |
+
},
|
| 189793 |
+
{
|
| 189794 |
+
"epoch": 573.14,
|
| 189795 |
+
"learning_rate": 8.877419871794872e-06,
|
| 189796 |
+
"loss": 0.3658,
|
| 189797 |
+
"step": 71075
|
| 189798 |
+
},
|
| 189799 |
+
{
|
| 189800 |
+
"epoch": 573.18,
|
| 189801 |
+
"learning_rate": 8.877339743589745e-06,
|
| 189802 |
+
"loss": 0.8553,
|
| 189803 |
+
"step": 71080
|
| 189804 |
+
},
|
| 189805 |
+
{
|
| 189806 |
+
"epoch": 573.22,
|
| 189807 |
+
"learning_rate": 8.877259615384615e-06,
|
| 189808 |
+
"loss": 0.8966,
|
| 189809 |
+
"step": 71085
|
| 189810 |
+
},
|
| 189811 |
+
{
|
| 189812 |
+
"epoch": 573.27,
|
| 189813 |
+
"learning_rate": 8.877179487179488e-06,
|
| 189814 |
+
"loss": 0.4122,
|
| 189815 |
+
"step": 71090
|
| 189816 |
+
},
|
| 189817 |
+
{
|
| 189818 |
+
"epoch": 573.31,
|
| 189819 |
+
"learning_rate": 8.87709935897436e-06,
|
| 189820 |
+
"loss": 0.3168,
|
| 189821 |
+
"step": 71095
|
| 189822 |
+
},
|
| 189823 |
+
{
|
| 189824 |
+
"epoch": 573.35,
|
| 189825 |
+
"learning_rate": 8.87701923076923e-06,
|
| 189826 |
+
"loss": 0.3781,
|
| 189827 |
+
"step": 71100
|
| 189828 |
+
},
|
| 189829 |
+
{
|
| 189830 |
+
"epoch": 573.39,
|
| 189831 |
+
"learning_rate": 8.876939102564102e-06,
|
| 189832 |
+
"loss": 0.7079,
|
| 189833 |
+
"step": 71105
|
| 189834 |
+
},
|
| 189835 |
+
{
|
| 189836 |
+
"epoch": 573.43,
|
| 189837 |
+
"learning_rate": 8.876858974358975e-06,
|
| 189838 |
+
"loss": 0.9013,
|
| 189839 |
+
"step": 71110
|
| 189840 |
+
},
|
| 189841 |
+
{
|
| 189842 |
+
"epoch": 573.47,
|
| 189843 |
+
"learning_rate": 8.876778846153847e-06,
|
| 189844 |
+
"loss": 0.3044,
|
| 189845 |
+
"step": 71115
|
| 189846 |
+
},
|
| 189847 |
+
{
|
| 189848 |
+
"epoch": 573.51,
|
| 189849 |
+
"learning_rate": 8.876698717948718e-06,
|
| 189850 |
+
"loss": 0.3062,
|
| 189851 |
+
"step": 71120
|
| 189852 |
+
},
|
| 189853 |
+
{
|
| 189854 |
+
"epoch": 573.55,
|
| 189855 |
+
"learning_rate": 8.876618589743591e-06,
|
| 189856 |
+
"loss": 0.4105,
|
| 189857 |
+
"step": 71125
|
| 189858 |
+
},
|
| 189859 |
+
{
|
| 189860 |
+
"epoch": 573.59,
|
| 189861 |
+
"learning_rate": 8.876538461538462e-06,
|
| 189862 |
+
"loss": 0.8214,
|
| 189863 |
+
"step": 71130
|
| 189864 |
+
},
|
| 189865 |
+
{
|
| 189866 |
+
"epoch": 573.63,
|
| 189867 |
+
"learning_rate": 8.876458333333334e-06,
|
| 189868 |
+
"loss": 0.739,
|
| 189869 |
+
"step": 71135
|
| 189870 |
+
},
|
| 189871 |
+
{
|
| 189872 |
+
"epoch": 573.67,
|
| 189873 |
+
"learning_rate": 8.876378205128205e-06,
|
| 189874 |
+
"loss": 0.2696,
|
| 189875 |
+
"step": 71140
|
| 189876 |
+
},
|
| 189877 |
+
{
|
| 189878 |
+
"epoch": 573.71,
|
| 189879 |
+
"learning_rate": 8.876298076923078e-06,
|
| 189880 |
+
"loss": 0.2784,
|
| 189881 |
+
"step": 71145
|
| 189882 |
+
},
|
| 189883 |
+
{
|
| 189884 |
+
"epoch": 573.75,
|
| 189885 |
+
"learning_rate": 8.87621794871795e-06,
|
| 189886 |
+
"loss": 0.3571,
|
| 189887 |
+
"step": 71150
|
| 189888 |
+
},
|
| 189889 |
+
{
|
| 189890 |
+
"epoch": 573.79,
|
| 189891 |
+
"learning_rate": 8.876137820512821e-06,
|
| 189892 |
+
"loss": 0.9113,
|
| 189893 |
+
"step": 71155
|
| 189894 |
+
},
|
| 189895 |
+
{
|
| 189896 |
+
"epoch": 573.83,
|
| 189897 |
+
"learning_rate": 8.876057692307694e-06,
|
| 189898 |
+
"loss": 0.9159,
|
| 189899 |
+
"step": 71160
|
| 189900 |
+
},
|
| 189901 |
+
{
|
| 189902 |
+
"epoch": 573.87,
|
| 189903 |
+
"learning_rate": 8.875977564102565e-06,
|
| 189904 |
+
"loss": 0.2806,
|
| 189905 |
+
"step": 71165
|
| 189906 |
+
},
|
| 189907 |
+
{
|
| 189908 |
+
"epoch": 573.91,
|
| 189909 |
+
"learning_rate": 8.875897435897437e-06,
|
| 189910 |
+
"loss": 0.3812,
|
| 189911 |
+
"step": 71170
|
| 189912 |
+
},
|
| 189913 |
+
{
|
| 189914 |
+
"epoch": 573.95,
|
| 189915 |
+
"learning_rate": 8.875817307692308e-06,
|
| 189916 |
+
"loss": 0.4332,
|
| 189917 |
+
"step": 71175
|
| 189918 |
+
},
|
| 189919 |
+
{
|
| 189920 |
+
"epoch": 573.99,
|
| 189921 |
+
"learning_rate": 8.875737179487181e-06,
|
| 189922 |
+
"loss": 0.8548,
|
| 189923 |
+
"step": 71180
|
| 189924 |
+
},
|
| 189925 |
+
{
|
| 189926 |
+
"epoch": 574.0,
|
| 189927 |
+
"eval_loss": 0.3877381682395935,
|
| 189928 |
+
"eval_runtime": 39.9108,
|
| 189929 |
+
"eval_samples_per_second": 21.047,
|
| 189930 |
+
"eval_steps_per_second": 0.677,
|
| 189931 |
+
"eval_wer": 0.19365056511410267,
|
| 189932 |
+
"step": 71181
|
| 189933 |
+
},
|
| 189934 |
+
{
|
| 189935 |
+
"epoch": 574.03,
|
| 189936 |
+
"learning_rate": 8.875657051282052e-06,
|
| 189937 |
+
"loss": 0.3677,
|
| 189938 |
+
"step": 71185
|
| 189939 |
+
},
|
| 189940 |
+
{
|
| 189941 |
+
"epoch": 574.07,
|
| 189942 |
+
"learning_rate": 8.875576923076924e-06,
|
| 189943 |
+
"loss": 0.3396,
|
| 189944 |
+
"step": 71190
|
| 189945 |
+
},
|
| 189946 |
+
{
|
| 189947 |
+
"epoch": 574.11,
|
| 189948 |
+
"learning_rate": 8.875496794871795e-06,
|
| 189949 |
+
"loss": 0.3444,
|
| 189950 |
+
"step": 71195
|
| 189951 |
+
},
|
| 189952 |
+
{
|
| 189953 |
+
"epoch": 574.15,
|
| 189954 |
+
"learning_rate": 8.875416666666668e-06,
|
| 189955 |
+
"loss": 0.3975,
|
| 189956 |
+
"step": 71200
|
| 189957 |
+
},
|
| 189958 |
+
{
|
| 189959 |
+
"epoch": 574.19,
|
| 189960 |
+
"learning_rate": 8.875336538461538e-06,
|
| 189961 |
+
"loss": 0.9141,
|
| 189962 |
+
"step": 71205
|
| 189963 |
+
},
|
| 189964 |
+
{
|
| 189965 |
+
"epoch": 574.23,
|
| 189966 |
+
"learning_rate": 8.875256410256411e-06,
|
| 189967 |
+
"loss": 0.7714,
|
| 189968 |
+
"step": 71210
|
| 189969 |
+
},
|
| 189970 |
+
{
|
| 189971 |
+
"epoch": 574.27,
|
| 189972 |
+
"learning_rate": 8.875176282051284e-06,
|
| 189973 |
+
"loss": 0.3316,
|
| 189974 |
+
"step": 71215
|
| 189975 |
+
},
|
| 189976 |
+
{
|
| 189977 |
+
"epoch": 574.31,
|
| 189978 |
+
"learning_rate": 8.875096153846154e-06,
|
| 189979 |
+
"loss": 0.3371,
|
| 189980 |
+
"step": 71220
|
| 189981 |
+
},
|
| 189982 |
+
{
|
| 189983 |
+
"epoch": 574.35,
|
| 189984 |
+
"learning_rate": 8.875016025641027e-06,
|
| 189985 |
+
"loss": 0.4133,
|
| 189986 |
+
"step": 71225
|
| 189987 |
+
},
|
| 189988 |
+
{
|
| 189989 |
+
"epoch": 574.39,
|
| 189990 |
+
"learning_rate": 8.874935897435898e-06,
|
| 189991 |
+
"loss": 1.0848,
|
| 189992 |
+
"step": 71230
|
| 189993 |
+
},
|
| 189994 |
+
{
|
| 189995 |
+
"epoch": 574.43,
|
| 189996 |
+
"learning_rate": 8.87485576923077e-06,
|
| 189997 |
+
"loss": 0.6723,
|
| 189998 |
+
"step": 71235
|
| 189999 |
+
},
|
| 190000 |
+
{
|
| 190001 |
+
"epoch": 574.47,
|
| 190002 |
+
"learning_rate": 8.87477564102564e-06,
|
| 190003 |
+
"loss": 0.2706,
|
| 190004 |
+
"step": 71240
|
| 190005 |
+
},
|
| 190006 |
+
{
|
| 190007 |
+
"epoch": 574.51,
|
| 190008 |
+
"learning_rate": 8.874695512820514e-06,
|
| 190009 |
+
"loss": 0.3278,
|
| 190010 |
+
"step": 71245
|
| 190011 |
+
},
|
| 190012 |
+
{
|
| 190013 |
+
"epoch": 574.55,
|
| 190014 |
+
"learning_rate": 8.874615384615385e-06,
|
| 190015 |
+
"loss": 0.3897,
|
| 190016 |
+
"step": 71250
|
| 190017 |
+
},
|
| 190018 |
+
{
|
| 190019 |
+
"epoch": 574.59,
|
| 190020 |
+
"learning_rate": 8.874535256410256e-06,
|
| 190021 |
+
"loss": 1.1266,
|
| 190022 |
+
"step": 71255
|
| 190023 |
+
},
|
| 190024 |
+
{
|
| 190025 |
+
"epoch": 574.63,
|
| 190026 |
+
"learning_rate": 8.87445512820513e-06,
|
| 190027 |
+
"loss": 0.8757,
|
| 190028 |
+
"step": 71260
|
| 190029 |
+
},
|
| 190030 |
+
{
|
| 190031 |
+
"epoch": 574.67,
|
| 190032 |
+
"learning_rate": 8.874375000000001e-06,
|
| 190033 |
+
"loss": 0.3173,
|
| 190034 |
+
"step": 71265
|
| 190035 |
+
},
|
| 190036 |
+
{
|
| 190037 |
+
"epoch": 574.71,
|
| 190038 |
+
"learning_rate": 8.874294871794872e-06,
|
| 190039 |
+
"loss": 0.2944,
|
| 190040 |
+
"step": 71270
|
| 190041 |
+
},
|
| 190042 |
+
{
|
| 190043 |
+
"epoch": 574.76,
|
| 190044 |
+
"learning_rate": 8.874214743589744e-06,
|
| 190045 |
+
"loss": 0.4019,
|
| 190046 |
+
"step": 71275
|
| 190047 |
+
},
|
| 190048 |
+
{
|
| 190049 |
+
"epoch": 574.8,
|
| 190050 |
+
"learning_rate": 8.874134615384617e-06,
|
| 190051 |
+
"loss": 1.0625,
|
| 190052 |
+
"step": 71280
|
| 190053 |
+
},
|
| 190054 |
+
{
|
| 190055 |
+
"epoch": 574.84,
|
| 190056 |
+
"learning_rate": 8.874054487179488e-06,
|
| 190057 |
+
"loss": 0.7614,
|
| 190058 |
+
"step": 71285
|
| 190059 |
+
},
|
| 190060 |
+
{
|
| 190061 |
+
"epoch": 574.88,
|
| 190062 |
+
"learning_rate": 8.87397435897436e-06,
|
| 190063 |
+
"loss": 0.2756,
|
| 190064 |
+
"step": 71290
|
| 190065 |
+
},
|
| 190066 |
+
{
|
| 190067 |
+
"epoch": 574.92,
|
| 190068 |
+
"learning_rate": 8.87389423076923e-06,
|
| 190069 |
+
"loss": 0.3798,
|
| 190070 |
+
"step": 71295
|
| 190071 |
+
},
|
| 190072 |
+
{
|
| 190073 |
+
"epoch": 574.96,
|
| 190074 |
+
"learning_rate": 8.873814102564104e-06,
|
| 190075 |
+
"loss": 0.4301,
|
| 190076 |
+
"step": 71300
|
| 190077 |
+
},
|
| 190078 |
+
{
|
| 190079 |
+
"epoch": 575.0,
|
| 190080 |
+
"learning_rate": 8.873733974358975e-06,
|
| 190081 |
+
"loss": 1.2421,
|
| 190082 |
+
"step": 71305
|
| 190083 |
+
},
|
| 190084 |
+
{
|
| 190085 |
+
"epoch": 575.0,
|
| 190086 |
+
"eval_loss": 0.4396674633026123,
|
| 190087 |
+
"eval_runtime": 40.1793,
|
| 190088 |
+
"eval_samples_per_second": 20.906,
|
| 190089 |
+
"eval_steps_per_second": 0.672,
|
| 190090 |
+
"eval_wer": 0.19375594063025517,
|
| 190091 |
+
"step": 71305
|
| 190092 |
+
},
|
| 190093 |
+
{
|
| 190094 |
+
"epoch": 575.04,
|
| 190095 |
+
"learning_rate": 8.873653846153847e-06,
|
| 190096 |
+
"loss": 0.3501,
|
| 190097 |
+
"step": 71310
|
| 190098 |
+
},
|
| 190099 |
+
{
|
| 190100 |
+
"epoch": 575.08,
|
| 190101 |
+
"learning_rate": 8.87357371794872e-06,
|
| 190102 |
+
"loss": 0.3542,
|
| 190103 |
+
"step": 71315
|
| 190104 |
+
},
|
| 190105 |
+
{
|
| 190106 |
+
"epoch": 575.12,
|
| 190107 |
+
"learning_rate": 8.873493589743591e-06,
|
| 190108 |
+
"loss": 0.3445,
|
| 190109 |
+
"step": 71320
|
| 190110 |
+
},
|
| 190111 |
+
{
|
| 190112 |
+
"epoch": 575.16,
|
| 190113 |
+
"learning_rate": 8.873413461538462e-06,
|
| 190114 |
+
"loss": 0.5054,
|
| 190115 |
+
"step": 71325
|
| 190116 |
+
},
|
| 190117 |
+
{
|
| 190118 |
+
"epoch": 575.2,
|
| 190119 |
+
"learning_rate": 8.873333333333334e-06,
|
| 190120 |
+
"loss": 1.209,
|
| 190121 |
+
"step": 71330
|
| 190122 |
+
},
|
| 190123 |
+
{
|
| 190124 |
+
"epoch": 575.24,
|
| 190125 |
+
"learning_rate": 8.873253205128207e-06,
|
| 190126 |
+
"loss": 0.3995,
|
| 190127 |
+
"step": 71335
|
| 190128 |
+
},
|
| 190129 |
+
{
|
| 190130 |
+
"epoch": 575.28,
|
| 190131 |
+
"learning_rate": 8.873173076923078e-06,
|
| 190132 |
+
"loss": 0.3047,
|
| 190133 |
+
"step": 71340
|
| 190134 |
+
},
|
| 190135 |
+
{
|
| 190136 |
+
"epoch": 575.32,
|
| 190137 |
+
"learning_rate": 8.87309294871795e-06,
|
| 190138 |
+
"loss": 0.3208,
|
| 190139 |
+
"step": 71345
|
| 190140 |
+
},
|
| 190141 |
+
{
|
| 190142 |
+
"epoch": 575.36,
|
| 190143 |
+
"learning_rate": 8.87301282051282e-06,
|
| 190144 |
+
"loss": 0.4865,
|
| 190145 |
+
"step": 71350
|
| 190146 |
+
},
|
| 190147 |
+
{
|
| 190148 |
+
"epoch": 575.4,
|
| 190149 |
+
"learning_rate": 8.872932692307694e-06,
|
| 190150 |
+
"loss": 1.2983,
|
| 190151 |
+
"step": 71355
|
| 190152 |
+
},
|
| 190153 |
+
{
|
| 190154 |
+
"epoch": 575.44,
|
| 190155 |
+
"learning_rate": 8.872852564102565e-06,
|
| 190156 |
+
"loss": 0.3453,
|
| 190157 |
+
"step": 71360
|
| 190158 |
+
},
|
| 190159 |
+
{
|
| 190160 |
+
"epoch": 575.48,
|
| 190161 |
+
"learning_rate": 8.872772435897437e-06,
|
| 190162 |
+
"loss": 0.3389,
|
| 190163 |
+
"step": 71365
|
| 190164 |
+
},
|
| 190165 |
+
{
|
| 190166 |
+
"epoch": 575.52,
|
| 190167 |
+
"learning_rate": 8.87269230769231e-06,
|
| 190168 |
+
"loss": 0.3255,
|
| 190169 |
+
"step": 71370
|
| 190170 |
+
},
|
| 190171 |
+
{
|
| 190172 |
+
"epoch": 575.56,
|
| 190173 |
+
"learning_rate": 8.87261217948718e-06,
|
| 190174 |
+
"loss": 0.4804,
|
| 190175 |
+
"step": 71375
|
| 190176 |
+
},
|
| 190177 |
+
{
|
| 190178 |
+
"epoch": 575.6,
|
| 190179 |
+
"learning_rate": 8.872532051282052e-06,
|
| 190180 |
+
"loss": 1.1065,
|
| 190181 |
+
"step": 71380
|
| 190182 |
+
},
|
| 190183 |
+
{
|
| 190184 |
+
"epoch": 575.64,
|
| 190185 |
+
"learning_rate": 8.872451923076924e-06,
|
| 190186 |
+
"loss": 0.3769,
|
| 190187 |
+
"step": 71385
|
| 190188 |
+
},
|
| 190189 |
+
{
|
| 190190 |
+
"epoch": 575.68,
|
| 190191 |
+
"learning_rate": 8.872371794871795e-06,
|
| 190192 |
+
"loss": 0.2841,
|
| 190193 |
+
"step": 71390
|
| 190194 |
+
},
|
| 190195 |
+
{
|
| 190196 |
+
"epoch": 575.72,
|
| 190197 |
+
"learning_rate": 8.872291666666666e-06,
|
| 190198 |
+
"loss": 0.3687,
|
| 190199 |
+
"step": 71395
|
| 190200 |
+
},
|
| 190201 |
+
{
|
| 190202 |
+
"epoch": 575.76,
|
| 190203 |
+
"learning_rate": 8.87221153846154e-06,
|
| 190204 |
+
"loss": 0.4895,
|
| 190205 |
+
"step": 71400
|
| 190206 |
+
},
|
| 190207 |
+
{
|
| 190208 |
+
"epoch": 575.8,
|
| 190209 |
+
"learning_rate": 8.87213141025641e-06,
|
| 190210 |
+
"loss": 1.1831,
|
| 190211 |
+
"step": 71405
|
| 190212 |
+
},
|
| 190213 |
+
{
|
| 190214 |
+
"epoch": 575.84,
|
| 190215 |
+
"learning_rate": 8.872051282051282e-06,
|
| 190216 |
+
"loss": 0.3517,
|
| 190217 |
+
"step": 71410
|
| 190218 |
+
},
|
| 190219 |
+
{
|
| 190220 |
+
"epoch": 575.88,
|
| 190221 |
+
"learning_rate": 8.871971153846155e-06,
|
| 190222 |
+
"loss": 0.3023,
|
| 190223 |
+
"step": 71415
|
| 190224 |
+
},
|
| 190225 |
+
{
|
| 190226 |
+
"epoch": 575.92,
|
| 190227 |
+
"learning_rate": 8.871891025641027e-06,
|
| 190228 |
+
"loss": 0.3109,
|
| 190229 |
+
"step": 71420
|
| 190230 |
+
},
|
| 190231 |
+
{
|
| 190232 |
+
"epoch": 575.96,
|
| 190233 |
+
"learning_rate": 8.871810897435898e-06,
|
| 190234 |
+
"loss": 0.5157,
|
| 190235 |
+
"step": 71425
|
| 190236 |
+
},
|
| 190237 |
+
{
|
| 190238 |
+
"epoch": 576.0,
|
| 190239 |
+
"eval_loss": 0.370493620634079,
|
| 190240 |
+
"eval_runtime": 40.7235,
|
| 190241 |
+
"eval_samples_per_second": 20.627,
|
| 190242 |
+
"eval_steps_per_second": 0.663,
|
| 190243 |
+
"eval_wer": 0.18538826318909307,
|
| 190244 |
+
"step": 71429
|
| 190245 |
}
|
| 190246 |
],
|
| 190247 |
"max_steps": 620000,
|
| 190248 |
"num_train_epochs": 5000,
|
| 190249 |
+
"total_flos": 2.0101646863500657e+20,
|
| 190250 |
"trial_name": null,
|
| 190251 |
"trial_params": null
|
| 190252 |
}
|
model-bin/finetune/base/{checkpoint-70809 β checkpoint-71429}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1629858970.8441613/events.out.tfevents.1629858970.7e498afd5545.905.23
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bdda49e59d41b85957a8cccc5577fcf0f644c5611ecd244a897f22720490ca05
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629859624.9463222/events.out.tfevents.1629859624.7e498afd5545.905.25
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f0de7580033efd4f6523ea6ea4d985866eb98165f7ce6cef08b659fd4f77df42
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629860278.0745044/events.out.tfevents.1629860278.7e498afd5545.905.27
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7f3764d5f0e85be9f251324fca32644721d02680b360ebd90ac23c0742e0a6f7
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629860932.5132222/events.out.tfevents.1629860932.7e498afd5545.905.29
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5e0087eb4fb9ddf06503c1a7b2305e2eb97a4422095bef32e827069366286069
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629861583.8800406/events.out.tfevents.1629861583.7e498afd5545.905.31
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:13c2e93be9000158ca59580c8a3592bd64fa86a07859832ca99e6402955c4981
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1629858970.7e498afd5545.905.22
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:95341d235f3561e1cc248e9f54713ad5d5c383fc0af5c13f55763c13de9e32d3
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629859624.7e498afd5545.905.24
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2c66a50c891088080220870c0899a1d489514ffd01056aad55cd7b982b2316b1
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629860278.7e498afd5545.905.26
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:49a2cad8d2fc8657e15e9a3c1596b3a90b0411c1665554e0fd6049efa13743ec
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629860932.7e498afd5545.905.28
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9e23d6fb0b7e0651fee49bbb66a50c47a62bbea9c783fb1b5f187de94c1bccca
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629861583.7e498afd5545.905.30
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:013cda62abdeb0bba2da058ab6dbf1e2e2fb143163b9e461ac2f41c2a39b07f1
|
| 3 |
+
size 8462
|