"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-73916 β checkpoint-74539}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-73916 β checkpoint-74539}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-73916 β checkpoint-74539}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-73916 β checkpoint-74539}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-73916 β checkpoint-74539}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-73916 β checkpoint-74539}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-73916 β checkpoint-74539}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-73916 β checkpoint-74539}/trainer_state.json +792 -3
- model-bin/finetune/base/{checkpoint-73916 β checkpoint-74539}/training_args.bin +0 -0
- model-bin/finetune/base/log/1629875783.5501642/events.out.tfevents.1629875783.7e498afd5545.905.73 +3 -0
- model-bin/finetune/base/log/1629876453.8615327/events.out.tfevents.1629876453.7e498afd5545.905.75 +3 -0
- model-bin/finetune/base/log/1629877099.8360853/events.out.tfevents.1629877099.7e498afd5545.905.77 +3 -0
- model-bin/finetune/base/log/1629877746.4587185/events.out.tfevents.1629877746.7e498afd5545.905.79 +3 -0
- model-bin/finetune/base/log/1629878403.3794868/events.out.tfevents.1629878403.7e498afd5545.905.81 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629875783.7e498afd5545.905.72 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629876453.7e498afd5545.905.74 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629877099.7e498afd5545.905.76 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629877746.7e498afd5545.905.78 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629878403.7e498afd5545.905.80 +3 -0
model-bin/finetune/base/{checkpoint-73916 β checkpoint-74539}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-73916 β checkpoint-74539}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:56aefc71d0980c995e93aee106a5ea7eb2b8637a7232ca5a4dcc1cdc4511335a
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-73916 β checkpoint-74539}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-73916 β checkpoint-74539}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9bd62e5038f5a12dc25600aabe73f692b4ca8f39c6d252fb9e3be2bbc5ed1145
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-73916 β checkpoint-74539}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a523a00f2b4a8aa255a532787f19739af80b95bca833548653c7ca44d7fe6056
|
| 3 |
size 14503
|
model-bin/finetune/base/{checkpoint-73916 β checkpoint-74539}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:89a798027e8a65425e8946906994aff28f6a07534f7b67f44af157705b99bf25
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-73916 β checkpoint-74539}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a5817db540e6b590b213245436521e08f3ff794616869a44c9a430a401dc99d6
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-73916 β checkpoint-74539}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.18412114350410416,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-69565",
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -193410,11 +193410,800 @@
|
|
| 193410 |
"eval_steps_per_second": 0.69,
|
| 193411 |
"eval_wer": 0.1899564928840056,
|
| 193412 |
"step": 73916
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 193413 |
}
|
| 193414 |
],
|
| 193415 |
"max_steps": 620000,
|
| 193416 |
"num_train_epochs": 5000,
|
| 193417 |
-
"total_flos": 2.
|
| 193418 |
"trial_name": null,
|
| 193419 |
"trial_params": null
|
| 193420 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.18412114350410416,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-69565",
|
| 4 |
+
"epoch": 600.995983935743,
|
| 5 |
+
"global_step": 74539,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 193410 |
"eval_steps_per_second": 0.69,
|
| 193411 |
"eval_wer": 0.1899564928840056,
|
| 193412 |
"step": 73916
|
| 193413 |
+
},
|
| 193414 |
+
{
|
| 193415 |
+
"epoch": 596.03,
|
| 193416 |
+
"learning_rate": 8.831875000000001e-06,
|
| 193417 |
+
"loss": 0.4505,
|
| 193418 |
+
"step": 73920
|
| 193419 |
+
},
|
| 193420 |
+
{
|
| 193421 |
+
"epoch": 596.07,
|
| 193422 |
+
"learning_rate": 8.831794871794872e-06,
|
| 193423 |
+
"loss": 0.3225,
|
| 193424 |
+
"step": 73925
|
| 193425 |
+
},
|
| 193426 |
+
{
|
| 193427 |
+
"epoch": 596.11,
|
| 193428 |
+
"learning_rate": 8.831714743589744e-06,
|
| 193429 |
+
"loss": 0.3288,
|
| 193430 |
+
"step": 73930
|
| 193431 |
+
},
|
| 193432 |
+
{
|
| 193433 |
+
"epoch": 596.15,
|
| 193434 |
+
"learning_rate": 8.831634615384617e-06,
|
| 193435 |
+
"loss": 0.413,
|
| 193436 |
+
"step": 73935
|
| 193437 |
+
},
|
| 193438 |
+
{
|
| 193439 |
+
"epoch": 596.19,
|
| 193440 |
+
"learning_rate": 8.831554487179488e-06,
|
| 193441 |
+
"loss": 0.8731,
|
| 193442 |
+
"step": 73940
|
| 193443 |
+
},
|
| 193444 |
+
{
|
| 193445 |
+
"epoch": 596.23,
|
| 193446 |
+
"learning_rate": 8.83147435897436e-06,
|
| 193447 |
+
"loss": 0.6002,
|
| 193448 |
+
"step": 73945
|
| 193449 |
+
},
|
| 193450 |
+
{
|
| 193451 |
+
"epoch": 596.27,
|
| 193452 |
+
"learning_rate": 8.83139423076923e-06,
|
| 193453 |
+
"loss": 0.3424,
|
| 193454 |
+
"step": 73950
|
| 193455 |
+
},
|
| 193456 |
+
{
|
| 193457 |
+
"epoch": 596.31,
|
| 193458 |
+
"learning_rate": 8.831314102564104e-06,
|
| 193459 |
+
"loss": 0.3209,
|
| 193460 |
+
"step": 73955
|
| 193461 |
+
},
|
| 193462 |
+
{
|
| 193463 |
+
"epoch": 596.35,
|
| 193464 |
+
"learning_rate": 8.831233974358975e-06,
|
| 193465 |
+
"loss": 0.4331,
|
| 193466 |
+
"step": 73960
|
| 193467 |
+
},
|
| 193468 |
+
{
|
| 193469 |
+
"epoch": 596.39,
|
| 193470 |
+
"learning_rate": 8.831153846153846e-06,
|
| 193471 |
+
"loss": 0.8799,
|
| 193472 |
+
"step": 73965
|
| 193473 |
+
},
|
| 193474 |
+
{
|
| 193475 |
+
"epoch": 596.43,
|
| 193476 |
+
"learning_rate": 8.831073717948718e-06,
|
| 193477 |
+
"loss": 0.7031,
|
| 193478 |
+
"step": 73970
|
| 193479 |
+
},
|
| 193480 |
+
{
|
| 193481 |
+
"epoch": 596.47,
|
| 193482 |
+
"learning_rate": 8.830993589743591e-06,
|
| 193483 |
+
"loss": 0.3835,
|
| 193484 |
+
"step": 73975
|
| 193485 |
+
},
|
| 193486 |
+
{
|
| 193487 |
+
"epoch": 596.51,
|
| 193488 |
+
"learning_rate": 8.830913461538462e-06,
|
| 193489 |
+
"loss": 0.2978,
|
| 193490 |
+
"step": 73980
|
| 193491 |
+
},
|
| 193492 |
+
{
|
| 193493 |
+
"epoch": 596.55,
|
| 193494 |
+
"learning_rate": 8.830833333333334e-06,
|
| 193495 |
+
"loss": 0.3954,
|
| 193496 |
+
"step": 73985
|
| 193497 |
+
},
|
| 193498 |
+
{
|
| 193499 |
+
"epoch": 596.59,
|
| 193500 |
+
"learning_rate": 8.830753205128207e-06,
|
| 193501 |
+
"loss": 0.8978,
|
| 193502 |
+
"step": 73990
|
| 193503 |
+
},
|
| 193504 |
+
{
|
| 193505 |
+
"epoch": 596.63,
|
| 193506 |
+
"learning_rate": 8.830673076923078e-06,
|
| 193507 |
+
"loss": 0.6776,
|
| 193508 |
+
"step": 73995
|
| 193509 |
+
},
|
| 193510 |
+
{
|
| 193511 |
+
"epoch": 596.67,
|
| 193512 |
+
"learning_rate": 8.83059294871795e-06,
|
| 193513 |
+
"loss": 0.3265,
|
| 193514 |
+
"step": 74000
|
| 193515 |
+
},
|
| 193516 |
+
{
|
| 193517 |
+
"epoch": 596.71,
|
| 193518 |
+
"learning_rate": 8.83051282051282e-06,
|
| 193519 |
+
"loss": 0.3834,
|
| 193520 |
+
"step": 74005
|
| 193521 |
+
},
|
| 193522 |
+
{
|
| 193523 |
+
"epoch": 596.76,
|
| 193524 |
+
"learning_rate": 8.830432692307694e-06,
|
| 193525 |
+
"loss": 0.366,
|
| 193526 |
+
"step": 74010
|
| 193527 |
+
},
|
| 193528 |
+
{
|
| 193529 |
+
"epoch": 596.8,
|
| 193530 |
+
"learning_rate": 8.830352564102565e-06,
|
| 193531 |
+
"loss": 0.9071,
|
| 193532 |
+
"step": 74015
|
| 193533 |
+
},
|
| 193534 |
+
{
|
| 193535 |
+
"epoch": 596.84,
|
| 193536 |
+
"learning_rate": 8.830272435897437e-06,
|
| 193537 |
+
"loss": 0.6363,
|
| 193538 |
+
"step": 74020
|
| 193539 |
+
},
|
| 193540 |
+
{
|
| 193541 |
+
"epoch": 596.88,
|
| 193542 |
+
"learning_rate": 8.830192307692308e-06,
|
| 193543 |
+
"loss": 0.3411,
|
| 193544 |
+
"step": 74025
|
| 193545 |
+
},
|
| 193546 |
+
{
|
| 193547 |
+
"epoch": 596.92,
|
| 193548 |
+
"learning_rate": 8.830112179487181e-06,
|
| 193549 |
+
"loss": 0.3157,
|
| 193550 |
+
"step": 74030
|
| 193551 |
+
},
|
| 193552 |
+
{
|
| 193553 |
+
"epoch": 596.96,
|
| 193554 |
+
"learning_rate": 8.830032051282052e-06,
|
| 193555 |
+
"loss": 0.3928,
|
| 193556 |
+
"step": 74035
|
| 193557 |
+
},
|
| 193558 |
+
{
|
| 193559 |
+
"epoch": 597.0,
|
| 193560 |
+
"learning_rate": 8.829951923076924e-06,
|
| 193561 |
+
"loss": 1.1771,
|
| 193562 |
+
"step": 74040
|
| 193563 |
+
},
|
| 193564 |
+
{
|
| 193565 |
+
"epoch": 597.0,
|
| 193566 |
+
"eval_loss": 0.5054441690444946,
|
| 193567 |
+
"eval_runtime": 40.5036,
|
| 193568 |
+
"eval_samples_per_second": 20.739,
|
| 193569 |
+
"eval_steps_per_second": 0.667,
|
| 193570 |
+
"eval_wer": 0.20609675281643472,
|
| 193571 |
+
"step": 74040
|
| 193572 |
+
},
|
| 193573 |
+
{
|
| 193574 |
+
"epoch": 592.04,
|
| 193575 |
+
"learning_rate": 8.829871794871797e-06,
|
| 193576 |
+
"loss": 0.3384,
|
| 193577 |
+
"step": 74045
|
| 193578 |
+
},
|
| 193579 |
+
{
|
| 193580 |
+
"epoch": 592.08,
|
| 193581 |
+
"learning_rate": 8.829791666666666e-06,
|
| 193582 |
+
"loss": 0.3687,
|
| 193583 |
+
"step": 74050
|
| 193584 |
+
},
|
| 193585 |
+
{
|
| 193586 |
+
"epoch": 592.12,
|
| 193587 |
+
"learning_rate": 8.82971153846154e-06,
|
| 193588 |
+
"loss": 0.3776,
|
| 193589 |
+
"step": 74055
|
| 193590 |
+
},
|
| 193591 |
+
{
|
| 193592 |
+
"epoch": 592.16,
|
| 193593 |
+
"learning_rate": 8.82963141025641e-06,
|
| 193594 |
+
"loss": 0.4675,
|
| 193595 |
+
"step": 74060
|
| 193596 |
+
},
|
| 193597 |
+
{
|
| 193598 |
+
"epoch": 592.2,
|
| 193599 |
+
"learning_rate": 8.829551282051282e-06,
|
| 193600 |
+
"loss": 1.0907,
|
| 193601 |
+
"step": 74065
|
| 193602 |
+
},
|
| 193603 |
+
{
|
| 193604 |
+
"epoch": 592.24,
|
| 193605 |
+
"learning_rate": 8.829471153846153e-06,
|
| 193606 |
+
"loss": 0.4606,
|
| 193607 |
+
"step": 74070
|
| 193608 |
+
},
|
| 193609 |
+
{
|
| 193610 |
+
"epoch": 592.28,
|
| 193611 |
+
"learning_rate": 8.829391025641027e-06,
|
| 193612 |
+
"loss": 0.2879,
|
| 193613 |
+
"step": 74075
|
| 193614 |
+
},
|
| 193615 |
+
{
|
| 193616 |
+
"epoch": 592.32,
|
| 193617 |
+
"learning_rate": 8.829310897435898e-06,
|
| 193618 |
+
"loss": 0.3348,
|
| 193619 |
+
"step": 74080
|
| 193620 |
+
},
|
| 193621 |
+
{
|
| 193622 |
+
"epoch": 592.36,
|
| 193623 |
+
"learning_rate": 8.82923076923077e-06,
|
| 193624 |
+
"loss": 0.5071,
|
| 193625 |
+
"step": 74085
|
| 193626 |
+
},
|
| 193627 |
+
{
|
| 193628 |
+
"epoch": 592.4,
|
| 193629 |
+
"learning_rate": 8.829150641025642e-06,
|
| 193630 |
+
"loss": 1.1857,
|
| 193631 |
+
"step": 74090
|
| 193632 |
+
},
|
| 193633 |
+
{
|
| 193634 |
+
"epoch": 592.44,
|
| 193635 |
+
"learning_rate": 8.829070512820514e-06,
|
| 193636 |
+
"loss": 0.3519,
|
| 193637 |
+
"step": 74095
|
| 193638 |
+
},
|
| 193639 |
+
{
|
| 193640 |
+
"epoch": 592.48,
|
| 193641 |
+
"learning_rate": 8.828990384615385e-06,
|
| 193642 |
+
"loss": 0.2731,
|
| 193643 |
+
"step": 74100
|
| 193644 |
+
},
|
| 193645 |
+
{
|
| 193646 |
+
"epoch": 592.52,
|
| 193647 |
+
"learning_rate": 8.828910256410256e-06,
|
| 193648 |
+
"loss": 0.3082,
|
| 193649 |
+
"step": 74105
|
| 193650 |
+
},
|
| 193651 |
+
{
|
| 193652 |
+
"epoch": 592.56,
|
| 193653 |
+
"learning_rate": 8.82883012820513e-06,
|
| 193654 |
+
"loss": 0.4305,
|
| 193655 |
+
"step": 74110
|
| 193656 |
+
},
|
| 193657 |
+
{
|
| 193658 |
+
"epoch": 592.6,
|
| 193659 |
+
"learning_rate": 8.82875e-06,
|
| 193660 |
+
"loss": 1.1372,
|
| 193661 |
+
"step": 74115
|
| 193662 |
+
},
|
| 193663 |
+
{
|
| 193664 |
+
"epoch": 592.64,
|
| 193665 |
+
"learning_rate": 8.828669871794872e-06,
|
| 193666 |
+
"loss": 0.3594,
|
| 193667 |
+
"step": 74120
|
| 193668 |
+
},
|
| 193669 |
+
{
|
| 193670 |
+
"epoch": 592.68,
|
| 193671 |
+
"learning_rate": 8.828589743589744e-06,
|
| 193672 |
+
"loss": 0.3154,
|
| 193673 |
+
"step": 74125
|
| 193674 |
+
},
|
| 193675 |
+
{
|
| 193676 |
+
"epoch": 592.72,
|
| 193677 |
+
"learning_rate": 8.828509615384617e-06,
|
| 193678 |
+
"loss": 0.3338,
|
| 193679 |
+
"step": 74130
|
| 193680 |
+
},
|
| 193681 |
+
{
|
| 193682 |
+
"epoch": 592.76,
|
| 193683 |
+
"learning_rate": 8.828429487179488e-06,
|
| 193684 |
+
"loss": 0.5104,
|
| 193685 |
+
"step": 74135
|
| 193686 |
+
},
|
| 193687 |
+
{
|
| 193688 |
+
"epoch": 592.8,
|
| 193689 |
+
"learning_rate": 8.82834935897436e-06,
|
| 193690 |
+
"loss": 1.0787,
|
| 193691 |
+
"step": 74140
|
| 193692 |
+
},
|
| 193693 |
+
{
|
| 193694 |
+
"epoch": 592.84,
|
| 193695 |
+
"learning_rate": 8.828269230769232e-06,
|
| 193696 |
+
"loss": 0.3784,
|
| 193697 |
+
"step": 74145
|
| 193698 |
+
},
|
| 193699 |
+
{
|
| 193700 |
+
"epoch": 592.88,
|
| 193701 |
+
"learning_rate": 8.828189102564104e-06,
|
| 193702 |
+
"loss": 0.2921,
|
| 193703 |
+
"step": 74150
|
| 193704 |
+
},
|
| 193705 |
+
{
|
| 193706 |
+
"epoch": 592.92,
|
| 193707 |
+
"learning_rate": 8.828108974358975e-06,
|
| 193708 |
+
"loss": 0.3931,
|
| 193709 |
+
"step": 74155
|
| 193710 |
+
},
|
| 193711 |
+
{
|
| 193712 |
+
"epoch": 592.96,
|
| 193713 |
+
"learning_rate": 8.828028846153846e-06,
|
| 193714 |
+
"loss": 0.5299,
|
| 193715 |
+
"step": 74160
|
| 193716 |
+
},
|
| 193717 |
+
{
|
| 193718 |
+
"epoch": 593.0,
|
| 193719 |
+
"learning_rate": 8.82794871794872e-06,
|
| 193720 |
+
"loss": 1.2197,
|
| 193721 |
+
"step": 74165
|
| 193722 |
+
},
|
| 193723 |
+
{
|
| 193724 |
+
"epoch": 593.0,
|
| 193725 |
+
"eval_loss": 0.4171510636806488,
|
| 193726 |
+
"eval_runtime": 40.75,
|
| 193727 |
+
"eval_samples_per_second": 20.638,
|
| 193728 |
+
"eval_steps_per_second": 0.663,
|
| 193729 |
+
"eval_wer": 0.20073123521399383,
|
| 193730 |
+
"step": 74165
|
| 193731 |
+
},
|
| 193732 |
+
{
|
| 193733 |
+
"epoch": 593.04,
|
| 193734 |
+
"learning_rate": 8.82786858974359e-06,
|
| 193735 |
+
"loss": 0.3557,
|
| 193736 |
+
"step": 74170
|
| 193737 |
+
},
|
| 193738 |
+
{
|
| 193739 |
+
"epoch": 593.08,
|
| 193740 |
+
"learning_rate": 8.827788461538462e-06,
|
| 193741 |
+
"loss": 0.3218,
|
| 193742 |
+
"step": 74175
|
| 193743 |
+
},
|
| 193744 |
+
{
|
| 193745 |
+
"epoch": 593.12,
|
| 193746 |
+
"learning_rate": 8.827708333333334e-06,
|
| 193747 |
+
"loss": 0.3495,
|
| 193748 |
+
"step": 74180
|
| 193749 |
+
},
|
| 193750 |
+
{
|
| 193751 |
+
"epoch": 593.16,
|
| 193752 |
+
"learning_rate": 8.827628205128207e-06,
|
| 193753 |
+
"loss": 0.5121,
|
| 193754 |
+
"step": 74185
|
| 193755 |
+
},
|
| 193756 |
+
{
|
| 193757 |
+
"epoch": 593.2,
|
| 193758 |
+
"learning_rate": 8.827548076923078e-06,
|
| 193759 |
+
"loss": 1.1842,
|
| 193760 |
+
"step": 74190
|
| 193761 |
+
},
|
| 193762 |
+
{
|
| 193763 |
+
"epoch": 593.24,
|
| 193764 |
+
"learning_rate": 8.82746794871795e-06,
|
| 193765 |
+
"loss": 0.3834,
|
| 193766 |
+
"step": 74195
|
| 193767 |
+
},
|
| 193768 |
+
{
|
| 193769 |
+
"epoch": 593.28,
|
| 193770 |
+
"learning_rate": 8.827387820512822e-06,
|
| 193771 |
+
"loss": 0.3191,
|
| 193772 |
+
"step": 74200
|
| 193773 |
+
},
|
| 193774 |
+
{
|
| 193775 |
+
"epoch": 593.32,
|
| 193776 |
+
"learning_rate": 8.827307692307692e-06,
|
| 193777 |
+
"loss": 0.3646,
|
| 193778 |
+
"step": 74205
|
| 193779 |
+
},
|
| 193780 |
+
{
|
| 193781 |
+
"epoch": 593.36,
|
| 193782 |
+
"learning_rate": 8.827227564102565e-06,
|
| 193783 |
+
"loss": 0.4677,
|
| 193784 |
+
"step": 74210
|
| 193785 |
+
},
|
| 193786 |
+
{
|
| 193787 |
+
"epoch": 593.4,
|
| 193788 |
+
"learning_rate": 8.827147435897436e-06,
|
| 193789 |
+
"loss": 1.2346,
|
| 193790 |
+
"step": 74215
|
| 193791 |
+
},
|
| 193792 |
+
{
|
| 193793 |
+
"epoch": 593.44,
|
| 193794 |
+
"learning_rate": 8.827067307692308e-06,
|
| 193795 |
+
"loss": 0.3663,
|
| 193796 |
+
"step": 74220
|
| 193797 |
+
},
|
| 193798 |
+
{
|
| 193799 |
+
"epoch": 593.48,
|
| 193800 |
+
"learning_rate": 8.82698717948718e-06,
|
| 193801 |
+
"loss": 0.2893,
|
| 193802 |
+
"step": 74225
|
| 193803 |
+
},
|
| 193804 |
+
{
|
| 193805 |
+
"epoch": 593.52,
|
| 193806 |
+
"learning_rate": 8.826907051282052e-06,
|
| 193807 |
+
"loss": 0.334,
|
| 193808 |
+
"step": 74230
|
| 193809 |
+
},
|
| 193810 |
+
{
|
| 193811 |
+
"epoch": 593.56,
|
| 193812 |
+
"learning_rate": 8.826826923076924e-06,
|
| 193813 |
+
"loss": 0.5513,
|
| 193814 |
+
"step": 74235
|
| 193815 |
+
},
|
| 193816 |
+
{
|
| 193817 |
+
"epoch": 593.6,
|
| 193818 |
+
"learning_rate": 8.826746794871795e-06,
|
| 193819 |
+
"loss": 1.281,
|
| 193820 |
+
"step": 74240
|
| 193821 |
+
},
|
| 193822 |
+
{
|
| 193823 |
+
"epoch": 593.64,
|
| 193824 |
+
"learning_rate": 8.826666666666668e-06,
|
| 193825 |
+
"loss": 0.3202,
|
| 193826 |
+
"step": 74245
|
| 193827 |
+
},
|
| 193828 |
+
{
|
| 193829 |
+
"epoch": 593.68,
|
| 193830 |
+
"learning_rate": 8.82658653846154e-06,
|
| 193831 |
+
"loss": 0.3076,
|
| 193832 |
+
"step": 74250
|
| 193833 |
+
},
|
| 193834 |
+
{
|
| 193835 |
+
"epoch": 593.72,
|
| 193836 |
+
"learning_rate": 8.82650641025641e-06,
|
| 193837 |
+
"loss": 0.3823,
|
| 193838 |
+
"step": 74255
|
| 193839 |
+
},
|
| 193840 |
+
{
|
| 193841 |
+
"epoch": 593.76,
|
| 193842 |
+
"learning_rate": 8.826426282051282e-06,
|
| 193843 |
+
"loss": 0.4498,
|
| 193844 |
+
"step": 74260
|
| 193845 |
+
},
|
| 193846 |
+
{
|
| 193847 |
+
"epoch": 593.8,
|
| 193848 |
+
"learning_rate": 8.826346153846155e-06,
|
| 193849 |
+
"loss": 1.3628,
|
| 193850 |
+
"step": 74265
|
| 193851 |
+
},
|
| 193852 |
+
{
|
| 193853 |
+
"epoch": 593.84,
|
| 193854 |
+
"learning_rate": 8.826266025641026e-06,
|
| 193855 |
+
"loss": 0.3403,
|
| 193856 |
+
"step": 74270
|
| 193857 |
+
},
|
| 193858 |
+
{
|
| 193859 |
+
"epoch": 593.88,
|
| 193860 |
+
"learning_rate": 8.826185897435898e-06,
|
| 193861 |
+
"loss": 0.3188,
|
| 193862 |
+
"step": 74275
|
| 193863 |
+
},
|
| 193864 |
+
{
|
| 193865 |
+
"epoch": 593.92,
|
| 193866 |
+
"learning_rate": 8.82610576923077e-06,
|
| 193867 |
+
"loss": 0.3402,
|
| 193868 |
+
"step": 74280
|
| 193869 |
+
},
|
| 193870 |
+
{
|
| 193871 |
+
"epoch": 593.96,
|
| 193872 |
+
"learning_rate": 8.826025641025642e-06,
|
| 193873 |
+
"loss": 0.486,
|
| 193874 |
+
"step": 74285
|
| 193875 |
+
},
|
| 193876 |
+
{
|
| 193877 |
+
"epoch": 594.0,
|
| 193878 |
+
"learning_rate": 8.825945512820514e-06,
|
| 193879 |
+
"loss": 1.311,
|
| 193880 |
+
"step": 74290
|
| 193881 |
+
},
|
| 193882 |
+
{
|
| 193883 |
+
"epoch": 594.0,
|
| 193884 |
+
"eval_loss": 0.35683706402778625,
|
| 193885 |
+
"eval_runtime": 40.5555,
|
| 193886 |
+
"eval_samples_per_second": 20.737,
|
| 193887 |
+
"eval_steps_per_second": 0.666,
|
| 193888 |
+
"eval_wer": 0.19015243684248193,
|
| 193889 |
+
"step": 74290
|
| 193890 |
+
},
|
| 193891 |
+
{
|
| 193892 |
+
"epoch": 594.04,
|
| 193893 |
+
"learning_rate": 8.825865384615385e-06,
|
| 193894 |
+
"loss": 0.358,
|
| 193895 |
+
"step": 74295
|
| 193896 |
+
},
|
| 193897 |
+
{
|
| 193898 |
+
"epoch": 594.08,
|
| 193899 |
+
"learning_rate": 8.825785256410258e-06,
|
| 193900 |
+
"loss": 0.2652,
|
| 193901 |
+
"step": 74300
|
| 193902 |
+
},
|
| 193903 |
+
{
|
| 193904 |
+
"epoch": 594.12,
|
| 193905 |
+
"learning_rate": 8.82570512820513e-06,
|
| 193906 |
+
"loss": 0.3892,
|
| 193907 |
+
"step": 74305
|
| 193908 |
+
},
|
| 193909 |
+
{
|
| 193910 |
+
"epoch": 594.16,
|
| 193911 |
+
"learning_rate": 8.825625e-06,
|
| 193912 |
+
"loss": 0.4418,
|
| 193913 |
+
"step": 74310
|
| 193914 |
+
},
|
| 193915 |
+
{
|
| 193916 |
+
"epoch": 594.2,
|
| 193917 |
+
"learning_rate": 8.825544871794872e-06,
|
| 193918 |
+
"loss": 1.2462,
|
| 193919 |
+
"step": 74315
|
| 193920 |
+
},
|
| 193921 |
+
{
|
| 193922 |
+
"epoch": 594.24,
|
| 193923 |
+
"learning_rate": 8.825464743589745e-06,
|
| 193924 |
+
"loss": 0.3345,
|
| 193925 |
+
"step": 74320
|
| 193926 |
+
},
|
| 193927 |
+
{
|
| 193928 |
+
"epoch": 594.28,
|
| 193929 |
+
"learning_rate": 8.825384615384617e-06,
|
| 193930 |
+
"loss": 0.2987,
|
| 193931 |
+
"step": 74325
|
| 193932 |
+
},
|
| 193933 |
+
{
|
| 193934 |
+
"epoch": 594.32,
|
| 193935 |
+
"learning_rate": 8.825304487179488e-06,
|
| 193936 |
+
"loss": 0.2656,
|
| 193937 |
+
"step": 74330
|
| 193938 |
+
},
|
| 193939 |
+
{
|
| 193940 |
+
"epoch": 594.36,
|
| 193941 |
+
"learning_rate": 8.825224358974361e-06,
|
| 193942 |
+
"loss": 0.5242,
|
| 193943 |
+
"step": 74335
|
| 193944 |
+
},
|
| 193945 |
+
{
|
| 193946 |
+
"epoch": 594.4,
|
| 193947 |
+
"learning_rate": 8.825144230769232e-06,
|
| 193948 |
+
"loss": 1.2575,
|
| 193949 |
+
"step": 74340
|
| 193950 |
+
},
|
| 193951 |
+
{
|
| 193952 |
+
"epoch": 594.44,
|
| 193953 |
+
"learning_rate": 8.825064102564104e-06,
|
| 193954 |
+
"loss": 0.3369,
|
| 193955 |
+
"step": 74345
|
| 193956 |
+
},
|
| 193957 |
+
{
|
| 193958 |
+
"epoch": 594.48,
|
| 193959 |
+
"learning_rate": 8.824983974358975e-06,
|
| 193960 |
+
"loss": 0.305,
|
| 193961 |
+
"step": 74350
|
| 193962 |
+
},
|
| 193963 |
+
{
|
| 193964 |
+
"epoch": 594.52,
|
| 193965 |
+
"learning_rate": 8.824903846153848e-06,
|
| 193966 |
+
"loss": 0.3236,
|
| 193967 |
+
"step": 74355
|
| 193968 |
+
},
|
| 193969 |
+
{
|
| 193970 |
+
"epoch": 594.56,
|
| 193971 |
+
"learning_rate": 8.824823717948718e-06,
|
| 193972 |
+
"loss": 0.4855,
|
| 193973 |
+
"step": 74360
|
| 193974 |
+
},
|
| 193975 |
+
{
|
| 193976 |
+
"epoch": 594.6,
|
| 193977 |
+
"learning_rate": 8.82474358974359e-06,
|
| 193978 |
+
"loss": 1.2836,
|
| 193979 |
+
"step": 74365
|
| 193980 |
+
},
|
| 193981 |
+
{
|
| 193982 |
+
"epoch": 594.64,
|
| 193983 |
+
"learning_rate": 8.824663461538462e-06,
|
| 193984 |
+
"loss": 0.3867,
|
| 193985 |
+
"step": 74370
|
| 193986 |
+
},
|
| 193987 |
+
{
|
| 193988 |
+
"epoch": 594.68,
|
| 193989 |
+
"learning_rate": 8.824583333333333e-06,
|
| 193990 |
+
"loss": 0.3239,
|
| 193991 |
+
"step": 74375
|
| 193992 |
+
},
|
| 193993 |
+
{
|
| 193994 |
+
"epoch": 594.72,
|
| 193995 |
+
"learning_rate": 8.824503205128205e-06,
|
| 193996 |
+
"loss": 0.4218,
|
| 193997 |
+
"step": 74380
|
| 193998 |
+
},
|
| 193999 |
+
{
|
| 194000 |
+
"epoch": 594.76,
|
| 194001 |
+
"learning_rate": 8.824423076923078e-06,
|
| 194002 |
+
"loss": 0.5141,
|
| 194003 |
+
"step": 74385
|
| 194004 |
+
},
|
| 194005 |
+
{
|
| 194006 |
+
"epoch": 594.8,
|
| 194007 |
+
"learning_rate": 8.82434294871795e-06,
|
| 194008 |
+
"loss": 1.3074,
|
| 194009 |
+
"step": 74390
|
| 194010 |
+
},
|
| 194011 |
+
{
|
| 194012 |
+
"epoch": 594.84,
|
| 194013 |
+
"learning_rate": 8.82426282051282e-06,
|
| 194014 |
+
"loss": 0.3945,
|
| 194015 |
+
"step": 74395
|
| 194016 |
+
},
|
| 194017 |
+
{
|
| 194018 |
+
"epoch": 594.88,
|
| 194019 |
+
"learning_rate": 8.824182692307694e-06,
|
| 194020 |
+
"loss": 0.3776,
|
| 194021 |
+
"step": 74400
|
| 194022 |
+
},
|
| 194023 |
+
{
|
| 194024 |
+
"epoch": 594.92,
|
| 194025 |
+
"learning_rate": 8.824102564102565e-06,
|
| 194026 |
+
"loss": 0.3246,
|
| 194027 |
+
"step": 74405
|
| 194028 |
+
},
|
| 194029 |
+
{
|
| 194030 |
+
"epoch": 594.96,
|
| 194031 |
+
"learning_rate": 8.824022435897436e-06,
|
| 194032 |
+
"loss": 0.4568,
|
| 194033 |
+
"step": 74410
|
| 194034 |
+
},
|
| 194035 |
+
{
|
| 194036 |
+
"epoch": 595.0,
|
| 194037 |
+
"learning_rate": 8.823942307692308e-06,
|
| 194038 |
+
"loss": 1.5204,
|
| 194039 |
+
"step": 74415
|
| 194040 |
+
},
|
| 194041 |
+
{
|
| 194042 |
+
"epoch": 595.0,
|
| 194043 |
+
"eval_loss": 0.437023401260376,
|
| 194044 |
+
"eval_runtime": 39.7077,
|
| 194045 |
+
"eval_samples_per_second": 21.18,
|
| 194046 |
+
"eval_steps_per_second": 0.68,
|
| 194047 |
+
"eval_wer": 0.1853523587652883,
|
| 194048 |
+
"step": 74415
|
| 194049 |
+
},
|
| 194050 |
+
{
|
| 194051 |
+
"epoch": 600.04,
|
| 194052 |
+
"learning_rate": 8.82386217948718e-06,
|
| 194053 |
+
"loss": 0.2979,
|
| 194054 |
+
"step": 74420
|
| 194055 |
+
},
|
| 194056 |
+
{
|
| 194057 |
+
"epoch": 600.08,
|
| 194058 |
+
"learning_rate": 8.823782051282052e-06,
|
| 194059 |
+
"loss": 0.2777,
|
| 194060 |
+
"step": 74425
|
| 194061 |
+
},
|
| 194062 |
+
{
|
| 194063 |
+
"epoch": 600.12,
|
| 194064 |
+
"learning_rate": 8.823701923076924e-06,
|
| 194065 |
+
"loss": 0.3034,
|
| 194066 |
+
"step": 74430
|
| 194067 |
+
},
|
| 194068 |
+
{
|
| 194069 |
+
"epoch": 600.16,
|
| 194070 |
+
"learning_rate": 8.823621794871795e-06,
|
| 194071 |
+
"loss": 0.4698,
|
| 194072 |
+
"step": 74435
|
| 194073 |
+
},
|
| 194074 |
+
{
|
| 194075 |
+
"epoch": 600.2,
|
| 194076 |
+
"learning_rate": 8.823541666666668e-06,
|
| 194077 |
+
"loss": 1.2657,
|
| 194078 |
+
"step": 74440
|
| 194079 |
+
},
|
| 194080 |
+
{
|
| 194081 |
+
"epoch": 600.24,
|
| 194082 |
+
"learning_rate": 8.82346153846154e-06,
|
| 194083 |
+
"loss": 0.3598,
|
| 194084 |
+
"step": 74445
|
| 194085 |
+
},
|
| 194086 |
+
{
|
| 194087 |
+
"epoch": 600.28,
|
| 194088 |
+
"learning_rate": 8.82338141025641e-06,
|
| 194089 |
+
"loss": 0.3204,
|
| 194090 |
+
"step": 74450
|
| 194091 |
+
},
|
| 194092 |
+
{
|
| 194093 |
+
"epoch": 600.32,
|
| 194094 |
+
"learning_rate": 8.823301282051284e-06,
|
| 194095 |
+
"loss": 0.3404,
|
| 194096 |
+
"step": 74455
|
| 194097 |
+
},
|
| 194098 |
+
{
|
| 194099 |
+
"epoch": 600.36,
|
| 194100 |
+
"learning_rate": 8.823221153846155e-06,
|
| 194101 |
+
"loss": 0.4718,
|
| 194102 |
+
"step": 74460
|
| 194103 |
+
},
|
| 194104 |
+
{
|
| 194105 |
+
"epoch": 600.4,
|
| 194106 |
+
"learning_rate": 8.823141025641026e-06,
|
| 194107 |
+
"loss": 1.1487,
|
| 194108 |
+
"step": 74465
|
| 194109 |
+
},
|
| 194110 |
+
{
|
| 194111 |
+
"epoch": 600.44,
|
| 194112 |
+
"learning_rate": 8.823060897435898e-06,
|
| 194113 |
+
"loss": 0.337,
|
| 194114 |
+
"step": 74470
|
| 194115 |
+
},
|
| 194116 |
+
{
|
| 194117 |
+
"epoch": 600.48,
|
| 194118 |
+
"learning_rate": 8.82298076923077e-06,
|
| 194119 |
+
"loss": 0.2972,
|
| 194120 |
+
"step": 74475
|
| 194121 |
+
},
|
| 194122 |
+
{
|
| 194123 |
+
"epoch": 600.52,
|
| 194124 |
+
"learning_rate": 8.82290064102564e-06,
|
| 194125 |
+
"loss": 0.3133,
|
| 194126 |
+
"step": 74480
|
| 194127 |
+
},
|
| 194128 |
+
{
|
| 194129 |
+
"epoch": 600.56,
|
| 194130 |
+
"learning_rate": 8.822820512820514e-06,
|
| 194131 |
+
"loss": 0.5256,
|
| 194132 |
+
"step": 74485
|
| 194133 |
+
},
|
| 194134 |
+
{
|
| 194135 |
+
"epoch": 600.6,
|
| 194136 |
+
"learning_rate": 8.822740384615387e-06,
|
| 194137 |
+
"loss": 1.1288,
|
| 194138 |
+
"step": 74490
|
| 194139 |
+
},
|
| 194140 |
+
{
|
| 194141 |
+
"epoch": 600.64,
|
| 194142 |
+
"learning_rate": 8.822660256410256e-06,
|
| 194143 |
+
"loss": 0.3383,
|
| 194144 |
+
"step": 74495
|
| 194145 |
+
},
|
| 194146 |
+
{
|
| 194147 |
+
"epoch": 600.68,
|
| 194148 |
+
"learning_rate": 8.82258012820513e-06,
|
| 194149 |
+
"loss": 0.3663,
|
| 194150 |
+
"step": 74500
|
| 194151 |
+
},
|
| 194152 |
+
{
|
| 194153 |
+
"epoch": 600.72,
|
| 194154 |
+
"learning_rate": 8.8225e-06,
|
| 194155 |
+
"loss": 0.5953,
|
| 194156 |
+
"step": 74505
|
| 194157 |
+
},
|
| 194158 |
+
{
|
| 194159 |
+
"epoch": 600.76,
|
| 194160 |
+
"learning_rate": 8.822419871794872e-06,
|
| 194161 |
+
"loss": 0.4843,
|
| 194162 |
+
"step": 74510
|
| 194163 |
+
},
|
| 194164 |
+
{
|
| 194165 |
+
"epoch": 600.8,
|
| 194166 |
+
"learning_rate": 8.822339743589743e-06,
|
| 194167 |
+
"loss": 1.1774,
|
| 194168 |
+
"step": 74515
|
| 194169 |
+
},
|
| 194170 |
+
{
|
| 194171 |
+
"epoch": 600.84,
|
| 194172 |
+
"learning_rate": 8.822259615384616e-06,
|
| 194173 |
+
"loss": 0.3648,
|
| 194174 |
+
"step": 74520
|
| 194175 |
+
},
|
| 194176 |
+
{
|
| 194177 |
+
"epoch": 600.88,
|
| 194178 |
+
"learning_rate": 8.822179487179488e-06,
|
| 194179 |
+
"loss": 0.312,
|
| 194180 |
+
"step": 74525
|
| 194181 |
+
},
|
| 194182 |
+
{
|
| 194183 |
+
"epoch": 600.92,
|
| 194184 |
+
"learning_rate": 8.822099358974359e-06,
|
| 194185 |
+
"loss": 0.3538,
|
| 194186 |
+
"step": 74530
|
| 194187 |
+
},
|
| 194188 |
+
{
|
| 194189 |
+
"epoch": 600.96,
|
| 194190 |
+
"learning_rate": 8.82201923076923e-06,
|
| 194191 |
+
"loss": 0.4753,
|
| 194192 |
+
"step": 74535
|
| 194193 |
+
},
|
| 194194 |
+
{
|
| 194195 |
+
"epoch": 601.0,
|
| 194196 |
+
"eval_loss": 0.4208657741546631,
|
| 194197 |
+
"eval_runtime": 40.5395,
|
| 194198 |
+
"eval_samples_per_second": 20.745,
|
| 194199 |
+
"eval_steps_per_second": 0.666,
|
| 194200 |
+
"eval_wer": 0.18913761467889909,
|
| 194201 |
+
"step": 74539
|
| 194202 |
}
|
| 194203 |
],
|
| 194204 |
"max_steps": 620000,
|
| 194205 |
"num_train_epochs": 5000,
|
| 194206 |
+
"total_flos": 2.097692339430875e+20,
|
| 194207 |
"trial_name": null,
|
| 194208 |
"trial_params": null
|
| 194209 |
}
|
model-bin/finetune/base/{checkpoint-73916 β checkpoint-74539}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1629875783.5501642/events.out.tfevents.1629875783.7e498afd5545.905.73
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:56464a81e08d7f1a82ade83b736f0756eb87693e0ddad64d889d382198e563fd
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629876453.8615327/events.out.tfevents.1629876453.7e498afd5545.905.75
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:be4bfbaec62c29f516a7e9a9d3917649f8fe892178caed42353cf30289db3e17
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629877099.8360853/events.out.tfevents.1629877099.7e498afd5545.905.77
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:645beaf58ccf8e567b5e37c2c291607674188f19f50f3e7ee43210e6407638dd
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629877746.4587185/events.out.tfevents.1629877746.7e498afd5545.905.79
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f60338304f297ccf1a1c0d050fa5246f8e8acbd98bdea98b1eef75c28f7d1170
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629878403.3794868/events.out.tfevents.1629878403.7e498afd5545.905.81
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:df1922471be3de7f9dd9e4f2a4714803af3f9197b17624b7f01328af60d370dc
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1629875783.7e498afd5545.905.72
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:77271c3153a042764056f6227bdbbc65ec6f3766b6800dfffce708dfdd9b7f0c
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629876453.7e498afd5545.905.74
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:30d0a63e279858ca5f738732cff0f1f32c7986714dad24fba8428eb4bd16a122
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629877099.7e498afd5545.905.76
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:709a794ebfbf29355911f836c4c1181d57b07ab751c07e59e543d2171a1932a6
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629877746.7e498afd5545.905.78
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:32769bd5c161b9f9e3573e58a8963e15f99507303208c12bcc0e1745c936edc3
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629878403.7e498afd5545.905.80
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9de9c1b83cc90e427f15e980270464cf53514719155c0e57b3b2b6f8df1b48f1
|
| 3 |
+
size 8462
|