"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-72051 β checkpoint-72674}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-72051 β checkpoint-72674}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-72051 β checkpoint-72674}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-72051 β checkpoint-72674}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-72051 β checkpoint-72674}/rng_state.pth +2 -2
- model-bin/finetune/base/{checkpoint-72051 β checkpoint-72674}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-72051 β checkpoint-72674}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-72051 β checkpoint-72674}/trainer_state.json +792 -3
- model-bin/finetune/base/{checkpoint-72051 β checkpoint-72674}/training_args.bin +0 -0
- model-bin/finetune/base/log/1629865712.3999202/events.out.tfevents.1629865712.7e498afd5545.905.43 +3 -0
- model-bin/finetune/base/log/1629866373.4153874/events.out.tfevents.1629866373.7e498afd5545.905.45 +3 -0
- model-bin/finetune/base/log/1629867035.5276291/events.out.tfevents.1629867035.7e498afd5545.905.47 +3 -0
- model-bin/finetune/base/log/1629867698.6983736/events.out.tfevents.1629867698.7e498afd5545.905.49 +3 -0
- model-bin/finetune/base/log/1629868357.748418/events.out.tfevents.1629868357.7e498afd5545.905.51 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629865712.7e498afd5545.905.42 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629866373.7e498afd5545.905.44 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629867035.7e498afd5545.905.46 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629867698.7e498afd5545.905.48 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629868357.7e498afd5545.905.50 +3 -0
model-bin/finetune/base/{checkpoint-72051 β checkpoint-72674}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-72051 β checkpoint-72674}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:27f7e94c0d2f13a946288d827ec9755f75add3218235b06362dd344e338875de
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-72051 β checkpoint-72674}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-72051 β checkpoint-72674}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8602c7981b5ea5c0d6a2f7a98e772f3ecff4b6e6e11024389798cfbefe92c1d7
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-72051 β checkpoint-72674}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d09a749fbdf2cfa38549a5df81720966b7cc868c88d312d1e2998e3ad599f8ec
|
| 3 |
+
size 14503
|
model-bin/finetune/base/{checkpoint-72051 β checkpoint-72674}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:06bd791a62086cb1e017bfe9d5222646f651e8165d1055ab3ef6d80ecdece72c
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-72051 β checkpoint-72674}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c835e72311c95e32d5f0b70d617e705a363bbb90fbd29eb2052c56576b97c7ab
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-72051 β checkpoint-72674}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.18412114350410416,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-69565",
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -191037,11 +191037,800 @@
|
|
| 191037 |
"eval_steps_per_second": 0.704,
|
| 191038 |
"eval_wer": 0.19118408401745432,
|
| 191039 |
"step": 72051
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 191040 |
}
|
| 191041 |
],
|
| 191042 |
"max_steps": 625000,
|
| 191043 |
"num_train_epochs": 5000,
|
| 191044 |
-
"total_flos": 2.
|
| 191045 |
"trial_name": null,
|
| 191046 |
"trial_params": null
|
| 191047 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.18412114350410416,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-69565",
|
| 4 |
+
"epoch": 581.0,
|
| 5 |
+
"global_step": 72674,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 191037 |
"eval_steps_per_second": 0.704,
|
| 191038 |
"eval_wer": 0.19118408401745432,
|
| 191039 |
"step": 72051
|
| 191040 |
+
},
|
| 191041 |
+
{
|
| 191042 |
+
"epoch": 576.03,
|
| 191043 |
+
"learning_rate": 8.86173076923077e-06,
|
| 191044 |
+
"loss": 0.339,
|
| 191045 |
+
"step": 72055
|
| 191046 |
+
},
|
| 191047 |
+
{
|
| 191048 |
+
"epoch": 576.07,
|
| 191049 |
+
"learning_rate": 8.861650641025642e-06,
|
| 191050 |
+
"loss": 0.3217,
|
| 191051 |
+
"step": 72060
|
| 191052 |
+
},
|
| 191053 |
+
{
|
| 191054 |
+
"epoch": 576.11,
|
| 191055 |
+
"learning_rate": 8.861570512820514e-06,
|
| 191056 |
+
"loss": 0.3057,
|
| 191057 |
+
"step": 72065
|
| 191058 |
+
},
|
| 191059 |
+
{
|
| 191060 |
+
"epoch": 576.15,
|
| 191061 |
+
"learning_rate": 8.861490384615385e-06,
|
| 191062 |
+
"loss": 0.4795,
|
| 191063 |
+
"step": 72070
|
| 191064 |
+
},
|
| 191065 |
+
{
|
| 191066 |
+
"epoch": 576.19,
|
| 191067 |
+
"learning_rate": 8.861410256410257e-06,
|
| 191068 |
+
"loss": 0.9112,
|
| 191069 |
+
"step": 72075
|
| 191070 |
+
},
|
| 191071 |
+
{
|
| 191072 |
+
"epoch": 576.23,
|
| 191073 |
+
"learning_rate": 8.86133012820513e-06,
|
| 191074 |
+
"loss": 0.7722,
|
| 191075 |
+
"step": 72080
|
| 191076 |
+
},
|
| 191077 |
+
{
|
| 191078 |
+
"epoch": 576.27,
|
| 191079 |
+
"learning_rate": 8.861250000000001e-06,
|
| 191080 |
+
"loss": 0.3769,
|
| 191081 |
+
"step": 72085
|
| 191082 |
+
},
|
| 191083 |
+
{
|
| 191084 |
+
"epoch": 576.31,
|
| 191085 |
+
"learning_rate": 8.861169871794872e-06,
|
| 191086 |
+
"loss": 0.3351,
|
| 191087 |
+
"step": 72090
|
| 191088 |
+
},
|
| 191089 |
+
{
|
| 191090 |
+
"epoch": 576.35,
|
| 191091 |
+
"learning_rate": 8.861089743589745e-06,
|
| 191092 |
+
"loss": 0.3966,
|
| 191093 |
+
"step": 72095
|
| 191094 |
+
},
|
| 191095 |
+
{
|
| 191096 |
+
"epoch": 576.39,
|
| 191097 |
+
"learning_rate": 8.861009615384617e-06,
|
| 191098 |
+
"loss": 1.0172,
|
| 191099 |
+
"step": 72100
|
| 191100 |
+
},
|
| 191101 |
+
{
|
| 191102 |
+
"epoch": 576.43,
|
| 191103 |
+
"learning_rate": 8.860929487179488e-06,
|
| 191104 |
+
"loss": 0.7078,
|
| 191105 |
+
"step": 72105
|
| 191106 |
+
},
|
| 191107 |
+
{
|
| 191108 |
+
"epoch": 576.47,
|
| 191109 |
+
"learning_rate": 8.86084935897436e-06,
|
| 191110 |
+
"loss": 0.2888,
|
| 191111 |
+
"step": 72110
|
| 191112 |
+
},
|
| 191113 |
+
{
|
| 191114 |
+
"epoch": 576.51,
|
| 191115 |
+
"learning_rate": 8.860769230769232e-06,
|
| 191116 |
+
"loss": 0.3262,
|
| 191117 |
+
"step": 72115
|
| 191118 |
+
},
|
| 191119 |
+
{
|
| 191120 |
+
"epoch": 576.55,
|
| 191121 |
+
"learning_rate": 8.860689102564102e-06,
|
| 191122 |
+
"loss": 0.4345,
|
| 191123 |
+
"step": 72120
|
| 191124 |
+
},
|
| 191125 |
+
{
|
| 191126 |
+
"epoch": 576.59,
|
| 191127 |
+
"learning_rate": 8.860608974358975e-06,
|
| 191128 |
+
"loss": 0.9028,
|
| 191129 |
+
"step": 72125
|
| 191130 |
+
},
|
| 191131 |
+
{
|
| 191132 |
+
"epoch": 576.63,
|
| 191133 |
+
"learning_rate": 8.860528846153847e-06,
|
| 191134 |
+
"loss": 0.6099,
|
| 191135 |
+
"step": 72130
|
| 191136 |
+
},
|
| 191137 |
+
{
|
| 191138 |
+
"epoch": 576.67,
|
| 191139 |
+
"learning_rate": 8.860448717948718e-06,
|
| 191140 |
+
"loss": 0.3631,
|
| 191141 |
+
"step": 72135
|
| 191142 |
+
},
|
| 191143 |
+
{
|
| 191144 |
+
"epoch": 576.71,
|
| 191145 |
+
"learning_rate": 8.860368589743591e-06,
|
| 191146 |
+
"loss": 0.3006,
|
| 191147 |
+
"step": 72140
|
| 191148 |
+
},
|
| 191149 |
+
{
|
| 191150 |
+
"epoch": 576.75,
|
| 191151 |
+
"learning_rate": 8.860288461538462e-06,
|
| 191152 |
+
"loss": 0.3792,
|
| 191153 |
+
"step": 72145
|
| 191154 |
+
},
|
| 191155 |
+
{
|
| 191156 |
+
"epoch": 576.79,
|
| 191157 |
+
"learning_rate": 8.860208333333334e-06,
|
| 191158 |
+
"loss": 0.8652,
|
| 191159 |
+
"step": 72150
|
| 191160 |
+
},
|
| 191161 |
+
{
|
| 191162 |
+
"epoch": 576.83,
|
| 191163 |
+
"learning_rate": 8.860128205128205e-06,
|
| 191164 |
+
"loss": 0.7947,
|
| 191165 |
+
"step": 72155
|
| 191166 |
+
},
|
| 191167 |
+
{
|
| 191168 |
+
"epoch": 576.87,
|
| 191169 |
+
"learning_rate": 8.860048076923078e-06,
|
| 191170 |
+
"loss": 0.2886,
|
| 191171 |
+
"step": 72160
|
| 191172 |
+
},
|
| 191173 |
+
{
|
| 191174 |
+
"epoch": 576.91,
|
| 191175 |
+
"learning_rate": 8.85996794871795e-06,
|
| 191176 |
+
"loss": 0.365,
|
| 191177 |
+
"step": 72165
|
| 191178 |
+
},
|
| 191179 |
+
{
|
| 191180 |
+
"epoch": 576.95,
|
| 191181 |
+
"learning_rate": 8.85988782051282e-06,
|
| 191182 |
+
"loss": 0.4469,
|
| 191183 |
+
"step": 72170
|
| 191184 |
+
},
|
| 191185 |
+
{
|
| 191186 |
+
"epoch": 576.99,
|
| 191187 |
+
"learning_rate": 8.859807692307692e-06,
|
| 191188 |
+
"loss": 0.9515,
|
| 191189 |
+
"step": 72175
|
| 191190 |
+
},
|
| 191191 |
+
{
|
| 191192 |
+
"epoch": 577.0,
|
| 191193 |
+
"eval_loss": 0.4047023355960846,
|
| 191194 |
+
"eval_runtime": 39.665,
|
| 191195 |
+
"eval_samples_per_second": 21.177,
|
| 191196 |
+
"eval_steps_per_second": 0.681,
|
| 191197 |
+
"eval_wer": 0.194370455690735,
|
| 191198 |
+
"step": 72176
|
| 191199 |
+
},
|
| 191200 |
+
{
|
| 191201 |
+
"epoch": 577.03,
|
| 191202 |
+
"learning_rate": 8.859727564102565e-06,
|
| 191203 |
+
"loss": 0.33,
|
| 191204 |
+
"step": 72180
|
| 191205 |
+
},
|
| 191206 |
+
{
|
| 191207 |
+
"epoch": 577.07,
|
| 191208 |
+
"learning_rate": 8.859647435897437e-06,
|
| 191209 |
+
"loss": 0.2739,
|
| 191210 |
+
"step": 72185
|
| 191211 |
+
},
|
| 191212 |
+
{
|
| 191213 |
+
"epoch": 577.11,
|
| 191214 |
+
"learning_rate": 8.859567307692308e-06,
|
| 191215 |
+
"loss": 0.3062,
|
| 191216 |
+
"step": 72190
|
| 191217 |
+
},
|
| 191218 |
+
{
|
| 191219 |
+
"epoch": 577.15,
|
| 191220 |
+
"learning_rate": 8.859487179487181e-06,
|
| 191221 |
+
"loss": 0.47,
|
| 191222 |
+
"step": 72195
|
| 191223 |
+
},
|
| 191224 |
+
{
|
| 191225 |
+
"epoch": 577.19,
|
| 191226 |
+
"learning_rate": 8.859407051282052e-06,
|
| 191227 |
+
"loss": 1.0719,
|
| 191228 |
+
"step": 72200
|
| 191229 |
+
},
|
| 191230 |
+
{
|
| 191231 |
+
"epoch": 577.23,
|
| 191232 |
+
"learning_rate": 8.859326923076924e-06,
|
| 191233 |
+
"loss": 0.6248,
|
| 191234 |
+
"step": 72205
|
| 191235 |
+
},
|
| 191236 |
+
{
|
| 191237 |
+
"epoch": 577.27,
|
| 191238 |
+
"learning_rate": 8.859246794871795e-06,
|
| 191239 |
+
"loss": 0.3089,
|
| 191240 |
+
"step": 72210
|
| 191241 |
+
},
|
| 191242 |
+
{
|
| 191243 |
+
"epoch": 577.31,
|
| 191244 |
+
"learning_rate": 8.859166666666668e-06,
|
| 191245 |
+
"loss": 0.3345,
|
| 191246 |
+
"step": 72215
|
| 191247 |
+
},
|
| 191248 |
+
{
|
| 191249 |
+
"epoch": 577.35,
|
| 191250 |
+
"learning_rate": 8.85908653846154e-06,
|
| 191251 |
+
"loss": 0.402,
|
| 191252 |
+
"step": 72220
|
| 191253 |
+
},
|
| 191254 |
+
{
|
| 191255 |
+
"epoch": 577.39,
|
| 191256 |
+
"learning_rate": 8.85900641025641e-06,
|
| 191257 |
+
"loss": 0.9915,
|
| 191258 |
+
"step": 72225
|
| 191259 |
+
},
|
| 191260 |
+
{
|
| 191261 |
+
"epoch": 577.43,
|
| 191262 |
+
"learning_rate": 8.858926282051282e-06,
|
| 191263 |
+
"loss": 0.6853,
|
| 191264 |
+
"step": 72230
|
| 191265 |
+
},
|
| 191266 |
+
{
|
| 191267 |
+
"epoch": 577.47,
|
| 191268 |
+
"learning_rate": 8.858846153846155e-06,
|
| 191269 |
+
"loss": 0.2974,
|
| 191270 |
+
"step": 72235
|
| 191271 |
+
},
|
| 191272 |
+
{
|
| 191273 |
+
"epoch": 577.51,
|
| 191274 |
+
"learning_rate": 8.858766025641027e-06,
|
| 191275 |
+
"loss": 0.3472,
|
| 191276 |
+
"step": 72240
|
| 191277 |
+
},
|
| 191278 |
+
{
|
| 191279 |
+
"epoch": 577.55,
|
| 191280 |
+
"learning_rate": 8.858685897435898e-06,
|
| 191281 |
+
"loss": 0.4193,
|
| 191282 |
+
"step": 72245
|
| 191283 |
+
},
|
| 191284 |
+
{
|
| 191285 |
+
"epoch": 577.59,
|
| 191286 |
+
"learning_rate": 8.858605769230771e-06,
|
| 191287 |
+
"loss": 0.9081,
|
| 191288 |
+
"step": 72250
|
| 191289 |
+
},
|
| 191290 |
+
{
|
| 191291 |
+
"epoch": 577.63,
|
| 191292 |
+
"learning_rate": 8.85852564102564e-06,
|
| 191293 |
+
"loss": 0.6596,
|
| 191294 |
+
"step": 72255
|
| 191295 |
+
},
|
| 191296 |
+
{
|
| 191297 |
+
"epoch": 577.67,
|
| 191298 |
+
"learning_rate": 8.858445512820514e-06,
|
| 191299 |
+
"loss": 0.2994,
|
| 191300 |
+
"step": 72260
|
| 191301 |
+
},
|
| 191302 |
+
{
|
| 191303 |
+
"epoch": 577.71,
|
| 191304 |
+
"learning_rate": 8.858365384615385e-06,
|
| 191305 |
+
"loss": 0.2584,
|
| 191306 |
+
"step": 72265
|
| 191307 |
+
},
|
| 191308 |
+
{
|
| 191309 |
+
"epoch": 577.75,
|
| 191310 |
+
"learning_rate": 8.858285256410256e-06,
|
| 191311 |
+
"loss": 0.4335,
|
| 191312 |
+
"step": 72270
|
| 191313 |
+
},
|
| 191314 |
+
{
|
| 191315 |
+
"epoch": 577.79,
|
| 191316 |
+
"learning_rate": 8.858205128205128e-06,
|
| 191317 |
+
"loss": 0.9398,
|
| 191318 |
+
"step": 72275
|
| 191319 |
+
},
|
| 191320 |
+
{
|
| 191321 |
+
"epoch": 577.83,
|
| 191322 |
+
"learning_rate": 8.858125e-06,
|
| 191323 |
+
"loss": 0.7375,
|
| 191324 |
+
"step": 72280
|
| 191325 |
+
},
|
| 191326 |
+
{
|
| 191327 |
+
"epoch": 577.87,
|
| 191328 |
+
"learning_rate": 8.858044871794872e-06,
|
| 191329 |
+
"loss": 0.3372,
|
| 191330 |
+
"step": 72285
|
| 191331 |
+
},
|
| 191332 |
+
{
|
| 191333 |
+
"epoch": 577.91,
|
| 191334 |
+
"learning_rate": 8.857964743589744e-06,
|
| 191335 |
+
"loss": 0.3376,
|
| 191336 |
+
"step": 72290
|
| 191337 |
+
},
|
| 191338 |
+
{
|
| 191339 |
+
"epoch": 577.95,
|
| 191340 |
+
"learning_rate": 8.857884615384617e-06,
|
| 191341 |
+
"loss": 0.4914,
|
| 191342 |
+
"step": 72295
|
| 191343 |
+
},
|
| 191344 |
+
{
|
| 191345 |
+
"epoch": 577.99,
|
| 191346 |
+
"learning_rate": 8.857804487179488e-06,
|
| 191347 |
+
"loss": 1.0205,
|
| 191348 |
+
"step": 72300
|
| 191349 |
+
},
|
| 191350 |
+
{
|
| 191351 |
+
"epoch": 578.0,
|
| 191352 |
+
"eval_loss": 0.447828084230423,
|
| 191353 |
+
"eval_runtime": 39.7038,
|
| 191354 |
+
"eval_samples_per_second": 21.157,
|
| 191355 |
+
"eval_steps_per_second": 0.68,
|
| 191356 |
+
"eval_wer": 0.1914162462528332,
|
| 191357 |
+
"step": 72301
|
| 191358 |
+
},
|
| 191359 |
+
{
|
| 191360 |
+
"epoch": 583.03,
|
| 191361 |
+
"learning_rate": 8.85772435897436e-06,
|
| 191362 |
+
"loss": 0.3688,
|
| 191363 |
+
"step": 72305
|
| 191364 |
+
},
|
| 191365 |
+
{
|
| 191366 |
+
"epoch": 583.07,
|
| 191367 |
+
"learning_rate": 8.85764423076923e-06,
|
| 191368 |
+
"loss": 0.3559,
|
| 191369 |
+
"step": 72310
|
| 191370 |
+
},
|
| 191371 |
+
{
|
| 191372 |
+
"epoch": 583.11,
|
| 191373 |
+
"learning_rate": 8.857564102564104e-06,
|
| 191374 |
+
"loss": 0.363,
|
| 191375 |
+
"step": 72315
|
| 191376 |
+
},
|
| 191377 |
+
{
|
| 191378 |
+
"epoch": 583.15,
|
| 191379 |
+
"learning_rate": 8.857483974358975e-06,
|
| 191380 |
+
"loss": 0.5013,
|
| 191381 |
+
"step": 72320
|
| 191382 |
+
},
|
| 191383 |
+
{
|
| 191384 |
+
"epoch": 583.19,
|
| 191385 |
+
"learning_rate": 8.857403846153846e-06,
|
| 191386 |
+
"loss": 1.1272,
|
| 191387 |
+
"step": 72325
|
| 191388 |
+
},
|
| 191389 |
+
{
|
| 191390 |
+
"epoch": 583.23,
|
| 191391 |
+
"learning_rate": 8.857323717948718e-06,
|
| 191392 |
+
"loss": 0.6567,
|
| 191393 |
+
"step": 72330
|
| 191394 |
+
},
|
| 191395 |
+
{
|
| 191396 |
+
"epoch": 583.27,
|
| 191397 |
+
"learning_rate": 8.857243589743591e-06,
|
| 191398 |
+
"loss": 0.3223,
|
| 191399 |
+
"step": 72335
|
| 191400 |
+
},
|
| 191401 |
+
{
|
| 191402 |
+
"epoch": 583.31,
|
| 191403 |
+
"learning_rate": 8.857163461538462e-06,
|
| 191404 |
+
"loss": 0.3139,
|
| 191405 |
+
"step": 72340
|
| 191406 |
+
},
|
| 191407 |
+
{
|
| 191408 |
+
"epoch": 583.35,
|
| 191409 |
+
"learning_rate": 8.857083333333334e-06,
|
| 191410 |
+
"loss": 0.3696,
|
| 191411 |
+
"step": 72345
|
| 191412 |
+
},
|
| 191413 |
+
{
|
| 191414 |
+
"epoch": 583.39,
|
| 191415 |
+
"learning_rate": 8.857003205128207e-06,
|
| 191416 |
+
"loss": 0.9257,
|
| 191417 |
+
"step": 72350
|
| 191418 |
+
},
|
| 191419 |
+
{
|
| 191420 |
+
"epoch": 583.43,
|
| 191421 |
+
"learning_rate": 8.856923076923078e-06,
|
| 191422 |
+
"loss": 0.6354,
|
| 191423 |
+
"step": 72355
|
| 191424 |
+
},
|
| 191425 |
+
{
|
| 191426 |
+
"epoch": 583.47,
|
| 191427 |
+
"learning_rate": 8.85684294871795e-06,
|
| 191428 |
+
"loss": 0.311,
|
| 191429 |
+
"step": 72360
|
| 191430 |
+
},
|
| 191431 |
+
{
|
| 191432 |
+
"epoch": 583.51,
|
| 191433 |
+
"learning_rate": 8.85676282051282e-06,
|
| 191434 |
+
"loss": 0.3108,
|
| 191435 |
+
"step": 72365
|
| 191436 |
+
},
|
| 191437 |
+
{
|
| 191438 |
+
"epoch": 583.55,
|
| 191439 |
+
"learning_rate": 8.856682692307694e-06,
|
| 191440 |
+
"loss": 0.3946,
|
| 191441 |
+
"step": 72370
|
| 191442 |
+
},
|
| 191443 |
+
{
|
| 191444 |
+
"epoch": 583.59,
|
| 191445 |
+
"learning_rate": 8.856602564102565e-06,
|
| 191446 |
+
"loss": 0.8928,
|
| 191447 |
+
"step": 72375
|
| 191448 |
+
},
|
| 191449 |
+
{
|
| 191450 |
+
"epoch": 583.63,
|
| 191451 |
+
"learning_rate": 8.856522435897436e-06,
|
| 191452 |
+
"loss": 0.688,
|
| 191453 |
+
"step": 72380
|
| 191454 |
+
},
|
| 191455 |
+
{
|
| 191456 |
+
"epoch": 583.67,
|
| 191457 |
+
"learning_rate": 8.85644230769231e-06,
|
| 191458 |
+
"loss": 0.3139,
|
| 191459 |
+
"step": 72385
|
| 191460 |
+
},
|
| 191461 |
+
{
|
| 191462 |
+
"epoch": 583.71,
|
| 191463 |
+
"learning_rate": 8.856362179487181e-06,
|
| 191464 |
+
"loss": 0.3488,
|
| 191465 |
+
"step": 72390
|
| 191466 |
+
},
|
| 191467 |
+
{
|
| 191468 |
+
"epoch": 583.76,
|
| 191469 |
+
"learning_rate": 8.856282051282052e-06,
|
| 191470 |
+
"loss": 0.4477,
|
| 191471 |
+
"step": 72395
|
| 191472 |
+
},
|
| 191473 |
+
{
|
| 191474 |
+
"epoch": 583.8,
|
| 191475 |
+
"learning_rate": 8.856201923076924e-06,
|
| 191476 |
+
"loss": 1.077,
|
| 191477 |
+
"step": 72400
|
| 191478 |
+
},
|
| 191479 |
+
{
|
| 191480 |
+
"epoch": 583.84,
|
| 191481 |
+
"learning_rate": 8.856121794871797e-06,
|
| 191482 |
+
"loss": 0.9005,
|
| 191483 |
+
"step": 72405
|
| 191484 |
+
},
|
| 191485 |
+
{
|
| 191486 |
+
"epoch": 583.88,
|
| 191487 |
+
"learning_rate": 8.856041666666666e-06,
|
| 191488 |
+
"loss": 0.2631,
|
| 191489 |
+
"step": 72410
|
| 191490 |
+
},
|
| 191491 |
+
{
|
| 191492 |
+
"epoch": 583.92,
|
| 191493 |
+
"learning_rate": 8.85596153846154e-06,
|
| 191494 |
+
"loss": 0.5206,
|
| 191495 |
+
"step": 72415
|
| 191496 |
+
},
|
| 191497 |
+
{
|
| 191498 |
+
"epoch": 583.96,
|
| 191499 |
+
"learning_rate": 8.85588141025641e-06,
|
| 191500 |
+
"loss": 0.4162,
|
| 191501 |
+
"step": 72420
|
| 191502 |
+
},
|
| 191503 |
+
{
|
| 191504 |
+
"epoch": 584.0,
|
| 191505 |
+
"learning_rate": 8.855801282051282e-06,
|
| 191506 |
+
"loss": 1.1406,
|
| 191507 |
+
"step": 72425
|
| 191508 |
+
},
|
| 191509 |
+
{
|
| 191510 |
+
"epoch": 584.0,
|
| 191511 |
+
"eval_loss": 0.3920002579689026,
|
| 191512 |
+
"eval_runtime": 40.716,
|
| 191513 |
+
"eval_samples_per_second": 20.631,
|
| 191514 |
+
"eval_steps_per_second": 0.663,
|
| 191515 |
+
"eval_wer": 0.18435091879075283,
|
| 191516 |
+
"step": 72425
|
| 191517 |
+
},
|
| 191518 |
+
{
|
| 191519 |
+
"epoch": 584.04,
|
| 191520 |
+
"learning_rate": 8.855721153846153e-06,
|
| 191521 |
+
"loss": 0.3607,
|
| 191522 |
+
"step": 72430
|
| 191523 |
+
},
|
| 191524 |
+
{
|
| 191525 |
+
"epoch": 584.08,
|
| 191526 |
+
"learning_rate": 8.855641025641027e-06,
|
| 191527 |
+
"loss": 0.3723,
|
| 191528 |
+
"step": 72435
|
| 191529 |
+
},
|
| 191530 |
+
{
|
| 191531 |
+
"epoch": 584.12,
|
| 191532 |
+
"learning_rate": 8.855560897435898e-06,
|
| 191533 |
+
"loss": 0.2966,
|
| 191534 |
+
"step": 72440
|
| 191535 |
+
},
|
| 191536 |
+
{
|
| 191537 |
+
"epoch": 584.16,
|
| 191538 |
+
"learning_rate": 8.85548076923077e-06,
|
| 191539 |
+
"loss": 0.513,
|
| 191540 |
+
"step": 72445
|
| 191541 |
+
},
|
| 191542 |
+
{
|
| 191543 |
+
"epoch": 584.2,
|
| 191544 |
+
"learning_rate": 8.855400641025642e-06,
|
| 191545 |
+
"loss": 1.1743,
|
| 191546 |
+
"step": 72450
|
| 191547 |
+
},
|
| 191548 |
+
{
|
| 191549 |
+
"epoch": 584.24,
|
| 191550 |
+
"learning_rate": 8.855320512820514e-06,
|
| 191551 |
+
"loss": 0.3608,
|
| 191552 |
+
"step": 72455
|
| 191553 |
+
},
|
| 191554 |
+
{
|
| 191555 |
+
"epoch": 584.28,
|
| 191556 |
+
"learning_rate": 8.855240384615385e-06,
|
| 191557 |
+
"loss": 0.2819,
|
| 191558 |
+
"step": 72460
|
| 191559 |
+
},
|
| 191560 |
+
{
|
| 191561 |
+
"epoch": 584.32,
|
| 191562 |
+
"learning_rate": 8.855160256410256e-06,
|
| 191563 |
+
"loss": 0.358,
|
| 191564 |
+
"step": 72465
|
| 191565 |
+
},
|
| 191566 |
+
{
|
| 191567 |
+
"epoch": 584.36,
|
| 191568 |
+
"learning_rate": 8.85508012820513e-06,
|
| 191569 |
+
"loss": 0.501,
|
| 191570 |
+
"step": 72470
|
| 191571 |
+
},
|
| 191572 |
+
{
|
| 191573 |
+
"epoch": 584.4,
|
| 191574 |
+
"learning_rate": 8.855e-06,
|
| 191575 |
+
"loss": 1.1884,
|
| 191576 |
+
"step": 72475
|
| 191577 |
+
},
|
| 191578 |
+
{
|
| 191579 |
+
"epoch": 584.44,
|
| 191580 |
+
"learning_rate": 8.854919871794872e-06,
|
| 191581 |
+
"loss": 0.3289,
|
| 191582 |
+
"step": 72480
|
| 191583 |
+
},
|
| 191584 |
+
{
|
| 191585 |
+
"epoch": 584.48,
|
| 191586 |
+
"learning_rate": 8.854839743589745e-06,
|
| 191587 |
+
"loss": 0.2994,
|
| 191588 |
+
"step": 72485
|
| 191589 |
+
},
|
| 191590 |
+
{
|
| 191591 |
+
"epoch": 584.52,
|
| 191592 |
+
"learning_rate": 8.854759615384617e-06,
|
| 191593 |
+
"loss": 0.3836,
|
| 191594 |
+
"step": 72490
|
| 191595 |
+
},
|
| 191596 |
+
{
|
| 191597 |
+
"epoch": 584.56,
|
| 191598 |
+
"learning_rate": 8.854679487179488e-06,
|
| 191599 |
+
"loss": 0.4344,
|
| 191600 |
+
"step": 72495
|
| 191601 |
+
},
|
| 191602 |
+
{
|
| 191603 |
+
"epoch": 584.6,
|
| 191604 |
+
"learning_rate": 8.85459935897436e-06,
|
| 191605 |
+
"loss": 1.2481,
|
| 191606 |
+
"step": 72500
|
| 191607 |
+
},
|
| 191608 |
+
{
|
| 191609 |
+
"epoch": 584.64,
|
| 191610 |
+
"learning_rate": 8.854519230769232e-06,
|
| 191611 |
+
"loss": 0.4116,
|
| 191612 |
+
"step": 72505
|
| 191613 |
+
},
|
| 191614 |
+
{
|
| 191615 |
+
"epoch": 584.68,
|
| 191616 |
+
"learning_rate": 8.854439102564104e-06,
|
| 191617 |
+
"loss": 0.2749,
|
| 191618 |
+
"step": 72510
|
| 191619 |
+
},
|
| 191620 |
+
{
|
| 191621 |
+
"epoch": 584.72,
|
| 191622 |
+
"learning_rate": 8.854358974358975e-06,
|
| 191623 |
+
"loss": 0.3268,
|
| 191624 |
+
"step": 72515
|
| 191625 |
+
},
|
| 191626 |
+
{
|
| 191627 |
+
"epoch": 584.76,
|
| 191628 |
+
"learning_rate": 8.854278846153846e-06,
|
| 191629 |
+
"loss": 0.4482,
|
| 191630 |
+
"step": 72520
|
| 191631 |
+
},
|
| 191632 |
+
{
|
| 191633 |
+
"epoch": 584.8,
|
| 191634 |
+
"learning_rate": 8.85419871794872e-06,
|
| 191635 |
+
"loss": 1.2621,
|
| 191636 |
+
"step": 72525
|
| 191637 |
+
},
|
| 191638 |
+
{
|
| 191639 |
+
"epoch": 584.84,
|
| 191640 |
+
"learning_rate": 8.85411858974359e-06,
|
| 191641 |
+
"loss": 0.3344,
|
| 191642 |
+
"step": 72530
|
| 191643 |
+
},
|
| 191644 |
+
{
|
| 191645 |
+
"epoch": 584.88,
|
| 191646 |
+
"learning_rate": 8.854038461538462e-06,
|
| 191647 |
+
"loss": 0.3034,
|
| 191648 |
+
"step": 72535
|
| 191649 |
+
},
|
| 191650 |
+
{
|
| 191651 |
+
"epoch": 584.92,
|
| 191652 |
+
"learning_rate": 8.853958333333335e-06,
|
| 191653 |
+
"loss": 0.3423,
|
| 191654 |
+
"step": 72540
|
| 191655 |
+
},
|
| 191656 |
+
{
|
| 191657 |
+
"epoch": 584.96,
|
| 191658 |
+
"learning_rate": 8.853878205128207e-06,
|
| 191659 |
+
"loss": 0.5005,
|
| 191660 |
+
"step": 72545
|
| 191661 |
+
},
|
| 191662 |
+
{
|
| 191663 |
+
"epoch": 585.0,
|
| 191664 |
+
"eval_loss": 0.39568430185317993,
|
| 191665 |
+
"eval_runtime": 39.2591,
|
| 191666 |
+
"eval_samples_per_second": 21.396,
|
| 191667 |
+
"eval_steps_per_second": 0.688,
|
| 191668 |
+
"eval_wer": 0.1896831245394252,
|
| 191669 |
+
"step": 72549
|
| 191670 |
+
},
|
| 191671 |
+
{
|
| 191672 |
+
"epoch": 580.01,
|
| 191673 |
+
"learning_rate": 8.853798076923078e-06,
|
| 191674 |
+
"loss": 0.4587,
|
| 191675 |
+
"step": 72550
|
| 191676 |
+
},
|
| 191677 |
+
{
|
| 191678 |
+
"epoch": 580.05,
|
| 191679 |
+
"learning_rate": 8.85371794871795e-06,
|
| 191680 |
+
"loss": 0.3441,
|
| 191681 |
+
"step": 72555
|
| 191682 |
+
},
|
| 191683 |
+
{
|
| 191684 |
+
"epoch": 580.09,
|
| 191685 |
+
"learning_rate": 8.853637820512822e-06,
|
| 191686 |
+
"loss": 0.3167,
|
| 191687 |
+
"step": 72560
|
| 191688 |
+
},
|
| 191689 |
+
{
|
| 191690 |
+
"epoch": 580.13,
|
| 191691 |
+
"learning_rate": 8.853557692307692e-06,
|
| 191692 |
+
"loss": 0.4195,
|
| 191693 |
+
"step": 72565
|
| 191694 |
+
},
|
| 191695 |
+
{
|
| 191696 |
+
"epoch": 580.17,
|
| 191697 |
+
"learning_rate": 8.853477564102565e-06,
|
| 191698 |
+
"loss": 0.5473,
|
| 191699 |
+
"step": 72570
|
| 191700 |
+
},
|
| 191701 |
+
{
|
| 191702 |
+
"epoch": 580.21,
|
| 191703 |
+
"learning_rate": 8.853397435897436e-06,
|
| 191704 |
+
"loss": 1.2351,
|
| 191705 |
+
"step": 72575
|
| 191706 |
+
},
|
| 191707 |
+
{
|
| 191708 |
+
"epoch": 580.25,
|
| 191709 |
+
"learning_rate": 8.853317307692308e-06,
|
| 191710 |
+
"loss": 0.4162,
|
| 191711 |
+
"step": 72580
|
| 191712 |
+
},
|
| 191713 |
+
{
|
| 191714 |
+
"epoch": 580.29,
|
| 191715 |
+
"learning_rate": 8.85323717948718e-06,
|
| 191716 |
+
"loss": 0.3726,
|
| 191717 |
+
"step": 72585
|
| 191718 |
+
},
|
| 191719 |
+
{
|
| 191720 |
+
"epoch": 580.33,
|
| 191721 |
+
"learning_rate": 8.853157051282052e-06,
|
| 191722 |
+
"loss": 0.311,
|
| 191723 |
+
"step": 72590
|
| 191724 |
+
},
|
| 191725 |
+
{
|
| 191726 |
+
"epoch": 580.37,
|
| 191727 |
+
"learning_rate": 8.853076923076924e-06,
|
| 191728 |
+
"loss": 0.6207,
|
| 191729 |
+
"step": 72595
|
| 191730 |
+
},
|
| 191731 |
+
{
|
| 191732 |
+
"epoch": 580.41,
|
| 191733 |
+
"learning_rate": 8.852996794871795e-06,
|
| 191734 |
+
"loss": 1.2596,
|
| 191735 |
+
"step": 72600
|
| 191736 |
+
},
|
| 191737 |
+
{
|
| 191738 |
+
"epoch": 580.45,
|
| 191739 |
+
"learning_rate": 8.852916666666668e-06,
|
| 191740 |
+
"loss": 0.3445,
|
| 191741 |
+
"step": 72605
|
| 191742 |
+
},
|
| 191743 |
+
{
|
| 191744 |
+
"epoch": 580.49,
|
| 191745 |
+
"learning_rate": 8.85283653846154e-06,
|
| 191746 |
+
"loss": 0.3112,
|
| 191747 |
+
"step": 72610
|
| 191748 |
+
},
|
| 191749 |
+
{
|
| 191750 |
+
"epoch": 580.53,
|
| 191751 |
+
"learning_rate": 8.85275641025641e-06,
|
| 191752 |
+
"loss": 0.3731,
|
| 191753 |
+
"step": 72615
|
| 191754 |
+
},
|
| 191755 |
+
{
|
| 191756 |
+
"epoch": 580.57,
|
| 191757 |
+
"learning_rate": 8.852676282051282e-06,
|
| 191758 |
+
"loss": 0.5238,
|
| 191759 |
+
"step": 72620
|
| 191760 |
+
},
|
| 191761 |
+
{
|
| 191762 |
+
"epoch": 580.61,
|
| 191763 |
+
"learning_rate": 8.852596153846155e-06,
|
| 191764 |
+
"loss": 1.196,
|
| 191765 |
+
"step": 72625
|
| 191766 |
+
},
|
| 191767 |
+
{
|
| 191768 |
+
"epoch": 580.65,
|
| 191769 |
+
"learning_rate": 8.852516025641026e-06,
|
| 191770 |
+
"loss": 0.2666,
|
| 191771 |
+
"step": 72630
|
| 191772 |
+
},
|
| 191773 |
+
{
|
| 191774 |
+
"epoch": 580.69,
|
| 191775 |
+
"learning_rate": 8.852435897435898e-06,
|
| 191776 |
+
"loss": 0.3653,
|
| 191777 |
+
"step": 72635
|
| 191778 |
+
},
|
| 191779 |
+
{
|
| 191780 |
+
"epoch": 580.73,
|
| 191781 |
+
"learning_rate": 8.852355769230771e-06,
|
| 191782 |
+
"loss": 0.3944,
|
| 191783 |
+
"step": 72640
|
| 191784 |
+
},
|
| 191785 |
+
{
|
| 191786 |
+
"epoch": 580.77,
|
| 191787 |
+
"learning_rate": 8.852275641025642e-06,
|
| 191788 |
+
"loss": 0.5596,
|
| 191789 |
+
"step": 72645
|
| 191790 |
+
},
|
| 191791 |
+
{
|
| 191792 |
+
"epoch": 580.81,
|
| 191793 |
+
"learning_rate": 8.852195512820514e-06,
|
| 191794 |
+
"loss": 1.1513,
|
| 191795 |
+
"step": 72650
|
| 191796 |
+
},
|
| 191797 |
+
{
|
| 191798 |
+
"epoch": 580.85,
|
| 191799 |
+
"learning_rate": 8.852115384615385e-06,
|
| 191800 |
+
"loss": 0.3255,
|
| 191801 |
+
"step": 72655
|
| 191802 |
+
},
|
| 191803 |
+
{
|
| 191804 |
+
"epoch": 580.89,
|
| 191805 |
+
"learning_rate": 8.852035256410258e-06,
|
| 191806 |
+
"loss": 0.25,
|
| 191807 |
+
"step": 72660
|
| 191808 |
+
},
|
| 191809 |
+
{
|
| 191810 |
+
"epoch": 580.93,
|
| 191811 |
+
"learning_rate": 8.85195512820513e-06,
|
| 191812 |
+
"loss": 0.4041,
|
| 191813 |
+
"step": 72665
|
| 191814 |
+
},
|
| 191815 |
+
{
|
| 191816 |
+
"epoch": 580.97,
|
| 191817 |
+
"learning_rate": 8.851875e-06,
|
| 191818 |
+
"loss": 0.6177,
|
| 191819 |
+
"step": 72670
|
| 191820 |
+
},
|
| 191821 |
+
{
|
| 191822 |
+
"epoch": 581.0,
|
| 191823 |
+
"eval_loss": 0.3960975706577301,
|
| 191824 |
+
"eval_runtime": 40.3396,
|
| 191825 |
+
"eval_samples_per_second": 20.823,
|
| 191826 |
+
"eval_steps_per_second": 0.669,
|
| 191827 |
+
"eval_wer": 0.19356979573777744,
|
| 191828 |
+
"step": 72674
|
| 191829 |
}
|
| 191830 |
],
|
| 191831 |
"max_steps": 625000,
|
| 191832 |
"num_train_epochs": 5000,
|
| 191833 |
+
"total_flos": 2.0451382170982056e+20,
|
| 191834 |
"trial_name": null,
|
| 191835 |
"trial_params": null
|
| 191836 |
}
|
model-bin/finetune/base/{checkpoint-72051 β checkpoint-72674}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1629865712.3999202/events.out.tfevents.1629865712.7e498afd5545.905.43
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8fe9358d32b91a31bd70c3b3f5de7f554b01d7c8e0997e77719220418e3a4d0b
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629866373.4153874/events.out.tfevents.1629866373.7e498afd5545.905.45
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:24d11bc9642f9b18f8b6e5803de4413f541482a4466333368fa876b6cc3709d3
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629867035.5276291/events.out.tfevents.1629867035.7e498afd5545.905.47
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:410a7f0161ac6cc4e8ebb05f5a9482de7b36e630250a0f7e7ebaa919175d729d
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629867698.6983736/events.out.tfevents.1629867698.7e498afd5545.905.49
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d3e789b2e73958b508e4dc81d29e447645345b6cabef52a495c211a76e64910e
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629868357.748418/events.out.tfevents.1629868357.7e498afd5545.905.51
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ca7151d4eeeeace9d594753ddac35893e62970b4085ae20b385b897f6982beb1
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1629865712.7e498afd5545.905.42
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:57bb040a5525dec5c7b9f9a950eb9fb3bd287a1af563c2b933e8b972d14f9326
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629866373.7e498afd5545.905.44
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9154ee54589c0587718af3f4a26a91428cc1ba2de06727df37fd75ec34586f4c
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629867035.7e498afd5545.905.46
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:858dfb7a5ae08a765b032454f21eac0366f1cd4b4bce7d2e0686d09489f449dd
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629867698.7e498afd5545.905.48
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:39db8ce2bf39668b96bbf7ee7e4378a20f42e93e9e02e90dd6395941f8451209
|
| 3 |
+
size 8462
|
model-bin/finetune/base/log/events.out.tfevents.1629868357.7e498afd5545.905.50
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5e1f96f05af17f0df8c92986868f317df4cb46095842d83a3ce3e7150543ed85
|
| 3 |
+
size 8622
|