"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-75162 β checkpoint-75787}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-75162 β checkpoint-75787}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-75162 β checkpoint-75787}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-75162 β checkpoint-75787}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-75162 β checkpoint-75787}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-75162 β checkpoint-75787}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-75162 β checkpoint-75787}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-75162 β checkpoint-75787}/trainer_state.json +798 -3
- model-bin/finetune/base/{checkpoint-75162 β checkpoint-75787}/training_args.bin +0 -0
- model-bin/finetune/base/log/1629882392.1236827/events.out.tfevents.1629882392.7e498afd5545.905.93 +3 -0
- model-bin/finetune/base/log/1629883056.4519293/events.out.tfevents.1629883056.7e498afd5545.905.95 +3 -0
- model-bin/finetune/base/log/1629883697.070274/events.out.tfevents.1629883697.7e498afd5545.905.97 +3 -0
- model-bin/finetune/base/log/1629884340.6409886/events.out.tfevents.1629884340.7e498afd5545.905.99 +3 -0
- model-bin/finetune/base/log/1629885006.4519074/events.out.tfevents.1629885006.7e498afd5545.905.101 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629882392.7e498afd5545.905.92 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629883056.7e498afd5545.905.94 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629883697.7e498afd5545.905.96 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629884340.7e498afd5545.905.98 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629885006.7e498afd5545.905.100 +3 -0
model-bin/finetune/base/{checkpoint-75162 β checkpoint-75787}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-75162 β checkpoint-75787}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:241cf9a0d25f6dfe6c4b09f0250d75c7b8556e0708ad5a758a4ae728e99ed97f
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-75162 β checkpoint-75787}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-75162 β checkpoint-75787}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:07eddd7e3b2c639376f65debd5c6e71a40abb39191deddef03737b5becdd9cb4
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-75162 β checkpoint-75787}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ff20bb0fbeb061a0d474c60bb2f68154a4d0944df8867df4e634564652ff6a60
|
| 3 |
size 14503
|
model-bin/finetune/base/{checkpoint-75162 β checkpoint-75787}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c6aaef4b9249c4bb80c33d2e747e6a2aee3a93d57119debf58fe887e8e98126a
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-75162 β checkpoint-75787}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8d36509b0709cb1c6a514d0b59c68cd5a3a2275d7fb28dc5813d880c7d4a4f0e
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-75162 β checkpoint-75787}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.18412114350410416,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-69565",
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -194994,11 +194994,806 @@
|
|
| 194994 |
"eval_steps_per_second": 0.665,
|
| 194995 |
"eval_wer": 0.2013764427557531,
|
| 194996 |
"step": 75162
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 194997 |
}
|
| 194998 |
],
|
| 194999 |
"max_steps": 625000,
|
| 195000 |
"num_train_epochs": 5000,
|
| 195001 |
-
"total_flos": 2.
|
| 195002 |
"trial_name": null,
|
| 195003 |
"trial_params": null
|
| 195004 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.18412114350410416,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-69565",
|
| 4 |
+
"epoch": 605.9960159362549,
|
| 5 |
+
"global_step": 75787,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 194994 |
"eval_steps_per_second": 0.665,
|
| 194995 |
"eval_wer": 0.2013764427557531,
|
| 194996 |
"step": 75162
|
| 194997 |
+
},
|
| 194998 |
+
{
|
| 194999 |
+
"epoch": 601.02,
|
| 195000 |
+
"learning_rate": 8.811923076923078e-06,
|
| 195001 |
+
"loss": 0.3253,
|
| 195002 |
+
"step": 75165
|
| 195003 |
+
},
|
| 195004 |
+
{
|
| 195005 |
+
"epoch": 601.06,
|
| 195006 |
+
"learning_rate": 8.811842948717949e-06,
|
| 195007 |
+
"loss": 0.3342,
|
| 195008 |
+
"step": 75170
|
| 195009 |
+
},
|
| 195010 |
+
{
|
| 195011 |
+
"epoch": 601.1,
|
| 195012 |
+
"learning_rate": 8.811762820512822e-06,
|
| 195013 |
+
"loss": 0.383,
|
| 195014 |
+
"step": 75175
|
| 195015 |
+
},
|
| 195016 |
+
{
|
| 195017 |
+
"epoch": 601.14,
|
| 195018 |
+
"learning_rate": 8.811682692307693e-06,
|
| 195019 |
+
"loss": 0.4493,
|
| 195020 |
+
"step": 75180
|
| 195021 |
+
},
|
| 195022 |
+
{
|
| 195023 |
+
"epoch": 601.18,
|
| 195024 |
+
"learning_rate": 8.811602564102565e-06,
|
| 195025 |
+
"loss": 0.7843,
|
| 195026 |
+
"step": 75185
|
| 195027 |
+
},
|
| 195028 |
+
{
|
| 195029 |
+
"epoch": 601.22,
|
| 195030 |
+
"learning_rate": 8.811522435897438e-06,
|
| 195031 |
+
"loss": 0.8807,
|
| 195032 |
+
"step": 75190
|
| 195033 |
+
},
|
| 195034 |
+
{
|
| 195035 |
+
"epoch": 601.26,
|
| 195036 |
+
"learning_rate": 8.811442307692309e-06,
|
| 195037 |
+
"loss": 0.2823,
|
| 195038 |
+
"step": 75195
|
| 195039 |
+
},
|
| 195040 |
+
{
|
| 195041 |
+
"epoch": 601.3,
|
| 195042 |
+
"learning_rate": 8.81136217948718e-06,
|
| 195043 |
+
"loss": 0.387,
|
| 195044 |
+
"step": 75200
|
| 195045 |
+
},
|
| 195046 |
+
{
|
| 195047 |
+
"epoch": 601.34,
|
| 195048 |
+
"learning_rate": 8.811282051282052e-06,
|
| 195049 |
+
"loss": 0.3841,
|
| 195050 |
+
"step": 75205
|
| 195051 |
+
},
|
| 195052 |
+
{
|
| 195053 |
+
"epoch": 601.38,
|
| 195054 |
+
"learning_rate": 8.811201923076925e-06,
|
| 195055 |
+
"loss": 0.7324,
|
| 195056 |
+
"step": 75210
|
| 195057 |
+
},
|
| 195058 |
+
{
|
| 195059 |
+
"epoch": 601.42,
|
| 195060 |
+
"learning_rate": 8.811121794871795e-06,
|
| 195061 |
+
"loss": 0.8773,
|
| 195062 |
+
"step": 75215
|
| 195063 |
+
},
|
| 195064 |
+
{
|
| 195065 |
+
"epoch": 601.46,
|
| 195066 |
+
"learning_rate": 8.811041666666668e-06,
|
| 195067 |
+
"loss": 0.2424,
|
| 195068 |
+
"step": 75220
|
| 195069 |
+
},
|
| 195070 |
+
{
|
| 195071 |
+
"epoch": 601.5,
|
| 195072 |
+
"learning_rate": 8.810961538461539e-06,
|
| 195073 |
+
"loss": 0.366,
|
| 195074 |
+
"step": 75225
|
| 195075 |
+
},
|
| 195076 |
+
{
|
| 195077 |
+
"epoch": 601.54,
|
| 195078 |
+
"learning_rate": 8.81088141025641e-06,
|
| 195079 |
+
"loss": 0.5086,
|
| 195080 |
+
"step": 75230
|
| 195081 |
+
},
|
| 195082 |
+
{
|
| 195083 |
+
"epoch": 601.58,
|
| 195084 |
+
"learning_rate": 8.810801282051282e-06,
|
| 195085 |
+
"loss": 0.8468,
|
| 195086 |
+
"step": 75235
|
| 195087 |
+
},
|
| 195088 |
+
{
|
| 195089 |
+
"epoch": 601.62,
|
| 195090 |
+
"learning_rate": 8.810721153846155e-06,
|
| 195091 |
+
"loss": 1.0401,
|
| 195092 |
+
"step": 75240
|
| 195093 |
+
},
|
| 195094 |
+
{
|
| 195095 |
+
"epoch": 601.66,
|
| 195096 |
+
"learning_rate": 8.810641025641026e-06,
|
| 195097 |
+
"loss": 0.3034,
|
| 195098 |
+
"step": 75245
|
| 195099 |
+
},
|
| 195100 |
+
{
|
| 195101 |
+
"epoch": 601.7,
|
| 195102 |
+
"learning_rate": 8.810560897435897e-06,
|
| 195103 |
+
"loss": 0.3393,
|
| 195104 |
+
"step": 75250
|
| 195105 |
+
},
|
| 195106 |
+
{
|
| 195107 |
+
"epoch": 601.74,
|
| 195108 |
+
"learning_rate": 8.81048076923077e-06,
|
| 195109 |
+
"loss": 0.3378,
|
| 195110 |
+
"step": 75255
|
| 195111 |
+
},
|
| 195112 |
+
{
|
| 195113 |
+
"epoch": 601.78,
|
| 195114 |
+
"learning_rate": 8.810400641025642e-06,
|
| 195115 |
+
"loss": 0.7611,
|
| 195116 |
+
"step": 75260
|
| 195117 |
+
},
|
| 195118 |
+
{
|
| 195119 |
+
"epoch": 601.82,
|
| 195120 |
+
"learning_rate": 8.810320512820513e-06,
|
| 195121 |
+
"loss": 1.0786,
|
| 195122 |
+
"step": 75265
|
| 195123 |
+
},
|
| 195124 |
+
{
|
| 195125 |
+
"epoch": 601.86,
|
| 195126 |
+
"learning_rate": 8.810240384615385e-06,
|
| 195127 |
+
"loss": 0.3302,
|
| 195128 |
+
"step": 75270
|
| 195129 |
+
},
|
| 195130 |
+
{
|
| 195131 |
+
"epoch": 601.9,
|
| 195132 |
+
"learning_rate": 8.810160256410258e-06,
|
| 195133 |
+
"loss": 0.2704,
|
| 195134 |
+
"step": 75275
|
| 195135 |
+
},
|
| 195136 |
+
{
|
| 195137 |
+
"epoch": 601.94,
|
| 195138 |
+
"learning_rate": 8.810080128205129e-06,
|
| 195139 |
+
"loss": 0.4022,
|
| 195140 |
+
"step": 75280
|
| 195141 |
+
},
|
| 195142 |
+
{
|
| 195143 |
+
"epoch": 601.98,
|
| 195144 |
+
"learning_rate": 8.81e-06,
|
| 195145 |
+
"loss": 0.8502,
|
| 195146 |
+
"step": 75285
|
| 195147 |
+
},
|
| 195148 |
+
{
|
| 195149 |
+
"epoch": 602.0,
|
| 195150 |
+
"eval_loss": 0.3613823354244232,
|
| 195151 |
+
"eval_runtime": 40.6158,
|
| 195152 |
+
"eval_samples_per_second": 20.706,
|
| 195153 |
+
"eval_steps_per_second": 0.665,
|
| 195154 |
+
"eval_wer": 0.1880054390610463,
|
| 195155 |
+
"step": 75287
|
| 195156 |
+
},
|
| 195157 |
+
{
|
| 195158 |
+
"epoch": 602.02,
|
| 195159 |
+
"learning_rate": 8.809919871794873e-06,
|
| 195160 |
+
"loss": 0.3835,
|
| 195161 |
+
"step": 75290
|
| 195162 |
+
},
|
| 195163 |
+
{
|
| 195164 |
+
"epoch": 602.06,
|
| 195165 |
+
"learning_rate": 8.809839743589745e-06,
|
| 195166 |
+
"loss": 0.2873,
|
| 195167 |
+
"step": 75295
|
| 195168 |
+
},
|
| 195169 |
+
{
|
| 195170 |
+
"epoch": 602.1,
|
| 195171 |
+
"learning_rate": 8.809759615384616e-06,
|
| 195172 |
+
"loss": 0.2944,
|
| 195173 |
+
"step": 75300
|
| 195174 |
+
},
|
| 195175 |
+
{
|
| 195176 |
+
"epoch": 602.14,
|
| 195177 |
+
"learning_rate": 8.809679487179487e-06,
|
| 195178 |
+
"loss": 0.3795,
|
| 195179 |
+
"step": 75305
|
| 195180 |
+
},
|
| 195181 |
+
{
|
| 195182 |
+
"epoch": 602.18,
|
| 195183 |
+
"learning_rate": 8.80959935897436e-06,
|
| 195184 |
+
"loss": 0.8076,
|
| 195185 |
+
"step": 75310
|
| 195186 |
+
},
|
| 195187 |
+
{
|
| 195188 |
+
"epoch": 602.22,
|
| 195189 |
+
"learning_rate": 8.809519230769232e-06,
|
| 195190 |
+
"loss": 0.8927,
|
| 195191 |
+
"step": 75315
|
| 195192 |
+
},
|
| 195193 |
+
{
|
| 195194 |
+
"epoch": 602.26,
|
| 195195 |
+
"learning_rate": 8.809439102564103e-06,
|
| 195196 |
+
"loss": 0.3271,
|
| 195197 |
+
"step": 75320
|
| 195198 |
+
},
|
| 195199 |
+
{
|
| 195200 |
+
"epoch": 602.3,
|
| 195201 |
+
"learning_rate": 8.809358974358975e-06,
|
| 195202 |
+
"loss": 0.3014,
|
| 195203 |
+
"step": 75325
|
| 195204 |
+
},
|
| 195205 |
+
{
|
| 195206 |
+
"epoch": 602.34,
|
| 195207 |
+
"learning_rate": 8.809278846153848e-06,
|
| 195208 |
+
"loss": 0.449,
|
| 195209 |
+
"step": 75330
|
| 195210 |
+
},
|
| 195211 |
+
{
|
| 195212 |
+
"epoch": 602.38,
|
| 195213 |
+
"learning_rate": 8.809198717948717e-06,
|
| 195214 |
+
"loss": 0.9804,
|
| 195215 |
+
"step": 75335
|
| 195216 |
+
},
|
| 195217 |
+
{
|
| 195218 |
+
"epoch": 602.42,
|
| 195219 |
+
"learning_rate": 8.80911858974359e-06,
|
| 195220 |
+
"loss": 0.8483,
|
| 195221 |
+
"step": 75340
|
| 195222 |
+
},
|
| 195223 |
+
{
|
| 195224 |
+
"epoch": 602.46,
|
| 195225 |
+
"learning_rate": 8.809038461538463e-06,
|
| 195226 |
+
"loss": 0.307,
|
| 195227 |
+
"step": 75345
|
| 195228 |
+
},
|
| 195229 |
+
{
|
| 195230 |
+
"epoch": 602.5,
|
| 195231 |
+
"learning_rate": 8.808958333333335e-06,
|
| 195232 |
+
"loss": 0.3112,
|
| 195233 |
+
"step": 75350
|
| 195234 |
+
},
|
| 195235 |
+
{
|
| 195236 |
+
"epoch": 602.54,
|
| 195237 |
+
"learning_rate": 8.808878205128206e-06,
|
| 195238 |
+
"loss": 0.4685,
|
| 195239 |
+
"step": 75355
|
| 195240 |
+
},
|
| 195241 |
+
{
|
| 195242 |
+
"epoch": 602.58,
|
| 195243 |
+
"learning_rate": 8.808798076923078e-06,
|
| 195244 |
+
"loss": 0.8196,
|
| 195245 |
+
"step": 75360
|
| 195246 |
+
},
|
| 195247 |
+
{
|
| 195248 |
+
"epoch": 602.62,
|
| 195249 |
+
"learning_rate": 8.80871794871795e-06,
|
| 195250 |
+
"loss": 0.7176,
|
| 195251 |
+
"step": 75365
|
| 195252 |
+
},
|
| 195253 |
+
{
|
| 195254 |
+
"epoch": 602.66,
|
| 195255 |
+
"learning_rate": 8.80863782051282e-06,
|
| 195256 |
+
"loss": 0.3299,
|
| 195257 |
+
"step": 75370
|
| 195258 |
+
},
|
| 195259 |
+
{
|
| 195260 |
+
"epoch": 602.7,
|
| 195261 |
+
"learning_rate": 8.808557692307693e-06,
|
| 195262 |
+
"loss": 0.3395,
|
| 195263 |
+
"step": 75375
|
| 195264 |
+
},
|
| 195265 |
+
{
|
| 195266 |
+
"epoch": 602.74,
|
| 195267 |
+
"learning_rate": 8.808477564102565e-06,
|
| 195268 |
+
"loss": 0.482,
|
| 195269 |
+
"step": 75380
|
| 195270 |
+
},
|
| 195271 |
+
{
|
| 195272 |
+
"epoch": 602.78,
|
| 195273 |
+
"learning_rate": 8.808397435897436e-06,
|
| 195274 |
+
"loss": 0.7871,
|
| 195275 |
+
"step": 75385
|
| 195276 |
+
},
|
| 195277 |
+
{
|
| 195278 |
+
"epoch": 602.82,
|
| 195279 |
+
"learning_rate": 8.808317307692307e-06,
|
| 195280 |
+
"loss": 1.0059,
|
| 195281 |
+
"step": 75390
|
| 195282 |
+
},
|
| 195283 |
+
{
|
| 195284 |
+
"epoch": 602.86,
|
| 195285 |
+
"learning_rate": 8.80823717948718e-06,
|
| 195286 |
+
"loss": 0.289,
|
| 195287 |
+
"step": 75395
|
| 195288 |
+
},
|
| 195289 |
+
{
|
| 195290 |
+
"epoch": 602.9,
|
| 195291 |
+
"learning_rate": 8.808157051282052e-06,
|
| 195292 |
+
"loss": 0.3244,
|
| 195293 |
+
"step": 75400
|
| 195294 |
+
},
|
| 195295 |
+
{
|
| 195296 |
+
"epoch": 602.94,
|
| 195297 |
+
"learning_rate": 8.808076923076923e-06,
|
| 195298 |
+
"loss": 0.4663,
|
| 195299 |
+
"step": 75405
|
| 195300 |
+
},
|
| 195301 |
+
{
|
| 195302 |
+
"epoch": 602.98,
|
| 195303 |
+
"learning_rate": 8.807996794871796e-06,
|
| 195304 |
+
"loss": 0.7288,
|
| 195305 |
+
"step": 75410
|
| 195306 |
+
},
|
| 195307 |
+
{
|
| 195308 |
+
"epoch": 603.0,
|
| 195309 |
+
"eval_loss": 0.4494644105434418,
|
| 195310 |
+
"eval_runtime": 40.3969,
|
| 195311 |
+
"eval_samples_per_second": 20.818,
|
| 195312 |
+
"eval_steps_per_second": 0.668,
|
| 195313 |
+
"eval_wer": 0.18498835177635412,
|
| 195314 |
+
"step": 75412
|
| 195315 |
+
},
|
| 195316 |
+
{
|
| 195317 |
+
"epoch": 603.02,
|
| 195318 |
+
"learning_rate": 8.807916666666668e-06,
|
| 195319 |
+
"loss": 0.4818,
|
| 195320 |
+
"step": 75415
|
| 195321 |
+
},
|
| 195322 |
+
{
|
| 195323 |
+
"epoch": 603.06,
|
| 195324 |
+
"learning_rate": 8.807836538461539e-06,
|
| 195325 |
+
"loss": 0.2858,
|
| 195326 |
+
"step": 75420
|
| 195327 |
+
},
|
| 195328 |
+
{
|
| 195329 |
+
"epoch": 603.1,
|
| 195330 |
+
"learning_rate": 8.80775641025641e-06,
|
| 195331 |
+
"loss": 0.3195,
|
| 195332 |
+
"step": 75425
|
| 195333 |
+
},
|
| 195334 |
+
{
|
| 195335 |
+
"epoch": 603.14,
|
| 195336 |
+
"learning_rate": 8.807676282051283e-06,
|
| 195337 |
+
"loss": 0.3992,
|
| 195338 |
+
"step": 75430
|
| 195339 |
+
},
|
| 195340 |
+
{
|
| 195341 |
+
"epoch": 603.18,
|
| 195342 |
+
"learning_rate": 8.807596153846155e-06,
|
| 195343 |
+
"loss": 0.8451,
|
| 195344 |
+
"step": 75435
|
| 195345 |
+
},
|
| 195346 |
+
{
|
| 195347 |
+
"epoch": 603.22,
|
| 195348 |
+
"learning_rate": 8.807516025641026e-06,
|
| 195349 |
+
"loss": 0.8592,
|
| 195350 |
+
"step": 75440
|
| 195351 |
+
},
|
| 195352 |
+
{
|
| 195353 |
+
"epoch": 603.26,
|
| 195354 |
+
"learning_rate": 8.807435897435899e-06,
|
| 195355 |
+
"loss": 0.3101,
|
| 195356 |
+
"step": 75445
|
| 195357 |
+
},
|
| 195358 |
+
{
|
| 195359 |
+
"epoch": 603.3,
|
| 195360 |
+
"learning_rate": 8.80735576923077e-06,
|
| 195361 |
+
"loss": 0.3108,
|
| 195362 |
+
"step": 75450
|
| 195363 |
+
},
|
| 195364 |
+
{
|
| 195365 |
+
"epoch": 603.34,
|
| 195366 |
+
"learning_rate": 8.807275641025642e-06,
|
| 195367 |
+
"loss": 0.3803,
|
| 195368 |
+
"step": 75455
|
| 195369 |
+
},
|
| 195370 |
+
{
|
| 195371 |
+
"epoch": 603.38,
|
| 195372 |
+
"learning_rate": 8.807195512820513e-06,
|
| 195373 |
+
"loss": 0.8631,
|
| 195374 |
+
"step": 75460
|
| 195375 |
+
},
|
| 195376 |
+
{
|
| 195377 |
+
"epoch": 603.42,
|
| 195378 |
+
"learning_rate": 8.807115384615386e-06,
|
| 195379 |
+
"loss": 0.9733,
|
| 195380 |
+
"step": 75465
|
| 195381 |
+
},
|
| 195382 |
+
{
|
| 195383 |
+
"epoch": 603.46,
|
| 195384 |
+
"learning_rate": 8.807035256410258e-06,
|
| 195385 |
+
"loss": 0.3768,
|
| 195386 |
+
"step": 75470
|
| 195387 |
+
},
|
| 195388 |
+
{
|
| 195389 |
+
"epoch": 603.5,
|
| 195390 |
+
"learning_rate": 8.806955128205129e-06,
|
| 195391 |
+
"loss": 0.3623,
|
| 195392 |
+
"step": 75475
|
| 195393 |
+
},
|
| 195394 |
+
{
|
| 195395 |
+
"epoch": 603.54,
|
| 195396 |
+
"learning_rate": 8.806875e-06,
|
| 195397 |
+
"loss": 0.3788,
|
| 195398 |
+
"step": 75480
|
| 195399 |
+
},
|
| 195400 |
+
{
|
| 195401 |
+
"epoch": 603.58,
|
| 195402 |
+
"learning_rate": 8.806794871794873e-06,
|
| 195403 |
+
"loss": 0.7772,
|
| 195404 |
+
"step": 75485
|
| 195405 |
+
},
|
| 195406 |
+
{
|
| 195407 |
+
"epoch": 603.62,
|
| 195408 |
+
"learning_rate": 8.806714743589743e-06,
|
| 195409 |
+
"loss": 0.9261,
|
| 195410 |
+
"step": 75490
|
| 195411 |
+
},
|
| 195412 |
+
{
|
| 195413 |
+
"epoch": 603.66,
|
| 195414 |
+
"learning_rate": 8.806634615384616e-06,
|
| 195415 |
+
"loss": 0.352,
|
| 195416 |
+
"step": 75495
|
| 195417 |
+
},
|
| 195418 |
+
{
|
| 195419 |
+
"epoch": 603.7,
|
| 195420 |
+
"learning_rate": 8.806554487179489e-06,
|
| 195421 |
+
"loss": 0.3595,
|
| 195422 |
+
"step": 75500
|
| 195423 |
+
},
|
| 195424 |
+
{
|
| 195425 |
+
"epoch": 603.74,
|
| 195426 |
+
"learning_rate": 8.806474358974359e-06,
|
| 195427 |
+
"loss": 0.3123,
|
| 195428 |
+
"step": 75505
|
| 195429 |
+
},
|
| 195430 |
+
{
|
| 195431 |
+
"epoch": 603.78,
|
| 195432 |
+
"learning_rate": 8.806394230769232e-06,
|
| 195433 |
+
"loss": 0.7048,
|
| 195434 |
+
"step": 75510
|
| 195435 |
+
},
|
| 195436 |
+
{
|
| 195437 |
+
"epoch": 603.82,
|
| 195438 |
+
"learning_rate": 8.806314102564103e-06,
|
| 195439 |
+
"loss": 0.9546,
|
| 195440 |
+
"step": 75515
|
| 195441 |
+
},
|
| 195442 |
+
{
|
| 195443 |
+
"epoch": 603.86,
|
| 195444 |
+
"learning_rate": 8.806233974358975e-06,
|
| 195445 |
+
"loss": 0.29,
|
| 195446 |
+
"step": 75520
|
| 195447 |
+
},
|
| 195448 |
+
{
|
| 195449 |
+
"epoch": 603.9,
|
| 195450 |
+
"learning_rate": 8.806153846153846e-06,
|
| 195451 |
+
"loss": 0.3054,
|
| 195452 |
+
"step": 75525
|
| 195453 |
+
},
|
| 195454 |
+
{
|
| 195455 |
+
"epoch": 603.94,
|
| 195456 |
+
"learning_rate": 8.806073717948719e-06,
|
| 195457 |
+
"loss": 0.3936,
|
| 195458 |
+
"step": 75530
|
| 195459 |
+
},
|
| 195460 |
+
{
|
| 195461 |
+
"epoch": 603.98,
|
| 195462 |
+
"learning_rate": 8.80599358974359e-06,
|
| 195463 |
+
"loss": 0.9524,
|
| 195464 |
+
"step": 75535
|
| 195465 |
+
},
|
| 195466 |
+
{
|
| 195467 |
+
"epoch": 604.0,
|
| 195468 |
+
"eval_loss": 0.4324324429035187,
|
| 195469 |
+
"eval_runtime": 41.0453,
|
| 195470 |
+
"eval_samples_per_second": 20.49,
|
| 195471 |
+
"eval_steps_per_second": 0.658,
|
| 195472 |
+
"eval_wer": 0.18972477064220183,
|
| 195473 |
+
"step": 75537
|
| 195474 |
+
},
|
| 195475 |
+
{
|
| 195476 |
+
"epoch": 604.02,
|
| 195477 |
+
"learning_rate": 8.805913461538462e-06,
|
| 195478 |
+
"loss": 0.3299,
|
| 195479 |
+
"step": 75540
|
| 195480 |
+
},
|
| 195481 |
+
{
|
| 195482 |
+
"epoch": 604.06,
|
| 195483 |
+
"learning_rate": 8.805833333333335e-06,
|
| 195484 |
+
"loss": 0.3726,
|
| 195485 |
+
"step": 75545
|
| 195486 |
+
},
|
| 195487 |
+
{
|
| 195488 |
+
"epoch": 604.1,
|
| 195489 |
+
"learning_rate": 8.805753205128206e-06,
|
| 195490 |
+
"loss": 0.3474,
|
| 195491 |
+
"step": 75550
|
| 195492 |
+
},
|
| 195493 |
+
{
|
| 195494 |
+
"epoch": 604.14,
|
| 195495 |
+
"learning_rate": 8.805673076923077e-06,
|
| 195496 |
+
"loss": 0.3935,
|
| 195497 |
+
"step": 75555
|
| 195498 |
+
},
|
| 195499 |
+
{
|
| 195500 |
+
"epoch": 604.18,
|
| 195501 |
+
"learning_rate": 8.805592948717949e-06,
|
| 195502 |
+
"loss": 0.7502,
|
| 195503 |
+
"step": 75560
|
| 195504 |
+
},
|
| 195505 |
+
{
|
| 195506 |
+
"epoch": 604.22,
|
| 195507 |
+
"learning_rate": 8.805512820512822e-06,
|
| 195508 |
+
"loss": 0.8715,
|
| 195509 |
+
"step": 75565
|
| 195510 |
+
},
|
| 195511 |
+
{
|
| 195512 |
+
"epoch": 604.26,
|
| 195513 |
+
"learning_rate": 8.805432692307693e-06,
|
| 195514 |
+
"loss": 0.4304,
|
| 195515 |
+
"step": 75570
|
| 195516 |
+
},
|
| 195517 |
+
{
|
| 195518 |
+
"epoch": 604.3,
|
| 195519 |
+
"learning_rate": 8.805352564102565e-06,
|
| 195520 |
+
"loss": 0.3175,
|
| 195521 |
+
"step": 75575
|
| 195522 |
+
},
|
| 195523 |
+
{
|
| 195524 |
+
"epoch": 604.34,
|
| 195525 |
+
"learning_rate": 8.805272435897436e-06,
|
| 195526 |
+
"loss": 0.3441,
|
| 195527 |
+
"step": 75580
|
| 195528 |
+
},
|
| 195529 |
+
{
|
| 195530 |
+
"epoch": 604.38,
|
| 195531 |
+
"learning_rate": 8.805192307692309e-06,
|
| 195532 |
+
"loss": 0.7054,
|
| 195533 |
+
"step": 75585
|
| 195534 |
+
},
|
| 195535 |
+
{
|
| 195536 |
+
"epoch": 604.42,
|
| 195537 |
+
"learning_rate": 8.80511217948718e-06,
|
| 195538 |
+
"loss": 0.8604,
|
| 195539 |
+
"step": 75590
|
| 195540 |
+
},
|
| 195541 |
+
{
|
| 195542 |
+
"epoch": 604.46,
|
| 195543 |
+
"learning_rate": 8.805032051282052e-06,
|
| 195544 |
+
"loss": 0.2919,
|
| 195545 |
+
"step": 75595
|
| 195546 |
+
},
|
| 195547 |
+
{
|
| 195548 |
+
"epoch": 604.5,
|
| 195549 |
+
"learning_rate": 8.804951923076925e-06,
|
| 195550 |
+
"loss": 0.3158,
|
| 195551 |
+
"step": 75600
|
| 195552 |
+
},
|
| 195553 |
+
{
|
| 195554 |
+
"epoch": 604.54,
|
| 195555 |
+
"learning_rate": 8.804871794871796e-06,
|
| 195556 |
+
"loss": 0.4358,
|
| 195557 |
+
"step": 75605
|
| 195558 |
+
},
|
| 195559 |
+
{
|
| 195560 |
+
"epoch": 604.58,
|
| 195561 |
+
"learning_rate": 8.804791666666667e-06,
|
| 195562 |
+
"loss": 0.8265,
|
| 195563 |
+
"step": 75610
|
| 195564 |
+
},
|
| 195565 |
+
{
|
| 195566 |
+
"epoch": 604.62,
|
| 195567 |
+
"learning_rate": 8.804711538461539e-06,
|
| 195568 |
+
"loss": 0.8654,
|
| 195569 |
+
"step": 75615
|
| 195570 |
+
},
|
| 195571 |
+
{
|
| 195572 |
+
"epoch": 604.66,
|
| 195573 |
+
"learning_rate": 8.804631410256412e-06,
|
| 195574 |
+
"loss": 0.272,
|
| 195575 |
+
"step": 75620
|
| 195576 |
+
},
|
| 195577 |
+
{
|
| 195578 |
+
"epoch": 604.7,
|
| 195579 |
+
"learning_rate": 8.804551282051283e-06,
|
| 195580 |
+
"loss": 0.2909,
|
| 195581 |
+
"step": 75625
|
| 195582 |
+
},
|
| 195583 |
+
{
|
| 195584 |
+
"epoch": 604.74,
|
| 195585 |
+
"learning_rate": 8.804471153846155e-06,
|
| 195586 |
+
"loss": 0.4165,
|
| 195587 |
+
"step": 75630
|
| 195588 |
+
},
|
| 195589 |
+
{
|
| 195590 |
+
"epoch": 604.78,
|
| 195591 |
+
"learning_rate": 8.804391025641026e-06,
|
| 195592 |
+
"loss": 0.8363,
|
| 195593 |
+
"step": 75635
|
| 195594 |
+
},
|
| 195595 |
+
{
|
| 195596 |
+
"epoch": 604.82,
|
| 195597 |
+
"learning_rate": 8.804310897435899e-06,
|
| 195598 |
+
"loss": 0.8427,
|
| 195599 |
+
"step": 75640
|
| 195600 |
+
},
|
| 195601 |
+
{
|
| 195602 |
+
"epoch": 604.86,
|
| 195603 |
+
"learning_rate": 8.80423076923077e-06,
|
| 195604 |
+
"loss": 0.3652,
|
| 195605 |
+
"step": 75645
|
| 195606 |
+
},
|
| 195607 |
+
{
|
| 195608 |
+
"epoch": 604.9,
|
| 195609 |
+
"learning_rate": 8.804150641025642e-06,
|
| 195610 |
+
"loss": 0.3643,
|
| 195611 |
+
"step": 75650
|
| 195612 |
+
},
|
| 195613 |
+
{
|
| 195614 |
+
"epoch": 604.94,
|
| 195615 |
+
"learning_rate": 8.804070512820515e-06,
|
| 195616 |
+
"loss": 0.3866,
|
| 195617 |
+
"step": 75655
|
| 195618 |
+
},
|
| 195619 |
+
{
|
| 195620 |
+
"epoch": 604.98,
|
| 195621 |
+
"learning_rate": 8.803990384615384e-06,
|
| 195622 |
+
"loss": 0.8886,
|
| 195623 |
+
"step": 75660
|
| 195624 |
+
},
|
| 195625 |
+
{
|
| 195626 |
+
"epoch": 605.0,
|
| 195627 |
+
"eval_loss": 0.3445565700531006,
|
| 195628 |
+
"eval_runtime": 40.5466,
|
| 195629 |
+
"eval_samples_per_second": 20.717,
|
| 195630 |
+
"eval_steps_per_second": 0.666,
|
| 195631 |
+
"eval_wer": 0.18799435028248587,
|
| 195632 |
+
"step": 75662
|
| 195633 |
+
},
|
| 195634 |
+
{
|
| 195635 |
+
"epoch": 605.02,
|
| 195636 |
+
"learning_rate": 8.803910256410257e-06,
|
| 195637 |
+
"loss": 0.3486,
|
| 195638 |
+
"step": 75665
|
| 195639 |
+
},
|
| 195640 |
+
{
|
| 195641 |
+
"epoch": 605.06,
|
| 195642 |
+
"learning_rate": 8.803830128205129e-06,
|
| 195643 |
+
"loss": 0.3072,
|
| 195644 |
+
"step": 75670
|
| 195645 |
+
},
|
| 195646 |
+
{
|
| 195647 |
+
"epoch": 605.1,
|
| 195648 |
+
"learning_rate": 8.80375e-06,
|
| 195649 |
+
"loss": 0.2955,
|
| 195650 |
+
"step": 75675
|
| 195651 |
+
},
|
| 195652 |
+
{
|
| 195653 |
+
"epoch": 605.14,
|
| 195654 |
+
"learning_rate": 8.803669871794872e-06,
|
| 195655 |
+
"loss": 0.3452,
|
| 195656 |
+
"step": 75680
|
| 195657 |
+
},
|
| 195658 |
+
{
|
| 195659 |
+
"epoch": 605.18,
|
| 195660 |
+
"learning_rate": 8.803589743589745e-06,
|
| 195661 |
+
"loss": 0.716,
|
| 195662 |
+
"step": 75685
|
| 195663 |
+
},
|
| 195664 |
+
{
|
| 195665 |
+
"epoch": 605.22,
|
| 195666 |
+
"learning_rate": 8.803509615384616e-06,
|
| 195667 |
+
"loss": 0.9065,
|
| 195668 |
+
"step": 75690
|
| 195669 |
+
},
|
| 195670 |
+
{
|
| 195671 |
+
"epoch": 605.26,
|
| 195672 |
+
"learning_rate": 8.803429487179487e-06,
|
| 195673 |
+
"loss": 0.2987,
|
| 195674 |
+
"step": 75695
|
| 195675 |
+
},
|
| 195676 |
+
{
|
| 195677 |
+
"epoch": 605.3,
|
| 195678 |
+
"learning_rate": 8.80334935897436e-06,
|
| 195679 |
+
"loss": 0.3722,
|
| 195680 |
+
"step": 75700
|
| 195681 |
+
},
|
| 195682 |
+
{
|
| 195683 |
+
"epoch": 605.34,
|
| 195684 |
+
"learning_rate": 8.803269230769232e-06,
|
| 195685 |
+
"loss": 0.4089,
|
| 195686 |
+
"step": 75705
|
| 195687 |
+
},
|
| 195688 |
+
{
|
| 195689 |
+
"epoch": 605.38,
|
| 195690 |
+
"learning_rate": 8.803189102564103e-06,
|
| 195691 |
+
"loss": 0.7393,
|
| 195692 |
+
"step": 75710
|
| 195693 |
+
},
|
| 195694 |
+
{
|
| 195695 |
+
"epoch": 605.42,
|
| 195696 |
+
"learning_rate": 8.803108974358974e-06,
|
| 195697 |
+
"loss": 1.0048,
|
| 195698 |
+
"step": 75715
|
| 195699 |
+
},
|
| 195700 |
+
{
|
| 195701 |
+
"epoch": 605.46,
|
| 195702 |
+
"learning_rate": 8.803028846153848e-06,
|
| 195703 |
+
"loss": 0.3025,
|
| 195704 |
+
"step": 75720
|
| 195705 |
+
},
|
| 195706 |
+
{
|
| 195707 |
+
"epoch": 605.5,
|
| 195708 |
+
"learning_rate": 8.802948717948719e-06,
|
| 195709 |
+
"loss": 0.2752,
|
| 195710 |
+
"step": 75725
|
| 195711 |
+
},
|
| 195712 |
+
{
|
| 195713 |
+
"epoch": 605.54,
|
| 195714 |
+
"learning_rate": 8.80286858974359e-06,
|
| 195715 |
+
"loss": 0.4407,
|
| 195716 |
+
"step": 75730
|
| 195717 |
+
},
|
| 195718 |
+
{
|
| 195719 |
+
"epoch": 605.58,
|
| 195720 |
+
"learning_rate": 8.802788461538462e-06,
|
| 195721 |
+
"loss": 0.7918,
|
| 195722 |
+
"step": 75735
|
| 195723 |
+
},
|
| 195724 |
+
{
|
| 195725 |
+
"epoch": 605.62,
|
| 195726 |
+
"learning_rate": 8.802708333333335e-06,
|
| 195727 |
+
"loss": 0.8038,
|
| 195728 |
+
"step": 75740
|
| 195729 |
+
},
|
| 195730 |
+
{
|
| 195731 |
+
"epoch": 605.66,
|
| 195732 |
+
"learning_rate": 8.802628205128206e-06,
|
| 195733 |
+
"loss": 0.2619,
|
| 195734 |
+
"step": 75745
|
| 195735 |
+
},
|
| 195736 |
+
{
|
| 195737 |
+
"epoch": 605.7,
|
| 195738 |
+
"learning_rate": 8.802548076923077e-06,
|
| 195739 |
+
"loss": 0.2717,
|
| 195740 |
+
"step": 75750
|
| 195741 |
+
},
|
| 195742 |
+
{
|
| 195743 |
+
"epoch": 605.74,
|
| 195744 |
+
"learning_rate": 8.80246794871795e-06,
|
| 195745 |
+
"loss": 0.3758,
|
| 195746 |
+
"step": 75755
|
| 195747 |
+
},
|
| 195748 |
+
{
|
| 195749 |
+
"epoch": 605.78,
|
| 195750 |
+
"learning_rate": 8.802387820512822e-06,
|
| 195751 |
+
"loss": 0.862,
|
| 195752 |
+
"step": 75760
|
| 195753 |
+
},
|
| 195754 |
+
{
|
| 195755 |
+
"epoch": 605.82,
|
| 195756 |
+
"learning_rate": 8.802307692307693e-06,
|
| 195757 |
+
"loss": 0.8831,
|
| 195758 |
+
"step": 75765
|
| 195759 |
+
},
|
| 195760 |
+
{
|
| 195761 |
+
"epoch": 605.86,
|
| 195762 |
+
"learning_rate": 8.802227564102564e-06,
|
| 195763 |
+
"loss": 0.2855,
|
| 195764 |
+
"step": 75770
|
| 195765 |
+
},
|
| 195766 |
+
{
|
| 195767 |
+
"epoch": 605.9,
|
| 195768 |
+
"learning_rate": 8.802147435897438e-06,
|
| 195769 |
+
"loss": 0.2964,
|
| 195770 |
+
"step": 75775
|
| 195771 |
+
},
|
| 195772 |
+
{
|
| 195773 |
+
"epoch": 605.94,
|
| 195774 |
+
"learning_rate": 8.802067307692307e-06,
|
| 195775 |
+
"loss": 0.3788,
|
| 195776 |
+
"step": 75780
|
| 195777 |
+
},
|
| 195778 |
+
{
|
| 195779 |
+
"epoch": 605.98,
|
| 195780 |
+
"learning_rate": 8.80198717948718e-06,
|
| 195781 |
+
"loss": 0.7917,
|
| 195782 |
+
"step": 75785
|
| 195783 |
+
},
|
| 195784 |
+
{
|
| 195785 |
+
"epoch": 606.0,
|
| 195786 |
+
"eval_loss": 0.40229326486587524,
|
| 195787 |
+
"eval_runtime": 40.811,
|
| 195788 |
+
"eval_samples_per_second": 20.583,
|
| 195789 |
+
"eval_steps_per_second": 0.662,
|
| 195790 |
+
"eval_wer": 0.1924281241104469,
|
| 195791 |
+
"step": 75787
|
| 195792 |
}
|
| 195793 |
],
|
| 195794 |
"max_steps": 625000,
|
| 195795 |
"num_train_epochs": 5000,
|
| 195796 |
+
"total_flos": 2.1327406074811505e+20,
|
| 195797 |
"trial_name": null,
|
| 195798 |
"trial_params": null
|
| 195799 |
}
|
model-bin/finetune/base/{checkpoint-75162 β checkpoint-75787}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1629882392.1236827/events.out.tfevents.1629882392.7e498afd5545.905.93
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:40faae71c44c577de7af7caf69225cd70fa91b8106891c9da59aeafbc4240253
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629883056.4519293/events.out.tfevents.1629883056.7e498afd5545.905.95
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:939b06a3739cc17a7e4511ffb94973599bbcb123803382e8d793b74decdad891
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629883697.070274/events.out.tfevents.1629883697.7e498afd5545.905.97
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bd4a1c43c8d0f8393356223e62609d12fe99feeffdaa48864cd0affc5cd80930
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629884340.6409886/events.out.tfevents.1629884340.7e498afd5545.905.99
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3047e66a47c520ca0d38a0885389812710a8a99fe0f4c1eee86b29e1ed75e104
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629885006.4519074/events.out.tfevents.1629885006.7e498afd5545.905.101
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:082cef36521c91b17fff53cd50db092c8004bc2f4c943a8df79734e4914e2dbf
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1629882392.7e498afd5545.905.92
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e16d61233274ee5c7a30f50978aade662d5dab45638811ad3f304ccc8670268a
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629883056.7e498afd5545.905.94
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:90b55c10715a3dd6e440f8c50d728d3f72f21cd59abdfaebc6a4eb38fc95b3af
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629883697.7e498afd5545.905.96
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a9f668ebd3dee0e05fcda258dc1e05f0bfad91436085cd5c6757bb21b1221574
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629884340.7e498afd5545.905.98
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e943e824b462a7e1503eebe32d0a77c274a8edcba1936795e22fd327cc802000
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629885006.7e498afd5545.905.100
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:45aa3d1626d97711f20adbacc8607e2484e99e05409cd57762d9b59f5f3c47cb
|
| 3 |
+
size 8622
|