"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-134152 β checkpoint-134775}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-134152 β checkpoint-134775}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-134152 β checkpoint-134775}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-134152 β checkpoint-134775}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-134152 β checkpoint-134775}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-134152 β checkpoint-134775}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-134152 β checkpoint-134775}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-134152 β checkpoint-134775}/trainer_state.json +799 -4
- model-bin/finetune/base/{checkpoint-134152 β checkpoint-134775}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630175438.376344/events.out.tfevents.1630175438.86bb0ddabf9b.4092.261 +3 -0
- model-bin/finetune/base/log/1630175824.2745113/events.out.tfevents.1630175824.86bb0ddabf9b.4092.263 +3 -0
- model-bin/finetune/base/log/1630176214.4682405/events.out.tfevents.1630176214.86bb0ddabf9b.4092.265 +3 -0
- model-bin/finetune/base/log/1630176600.245581/events.out.tfevents.1630176600.86bb0ddabf9b.4092.267 +3 -0
- model-bin/finetune/base/log/1630176983.8605223/events.out.tfevents.1630176983.86bb0ddabf9b.4092.269 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630175438.86bb0ddabf9b.4092.260 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630175824.86bb0ddabf9b.4092.262 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630176214.86bb0ddabf9b.4092.264 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630176600.86bb0ddabf9b.4092.266 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630176983.86bb0ddabf9b.4092.268 +3 -0
model-bin/finetune/base/{checkpoint-134152 β checkpoint-134775}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-134152 β checkpoint-134775}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:be3b7f7b82d9b49c3cb2566dff8f9b22035b90339ae9a63f872cf0a4b578f61b
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-134152 β checkpoint-134775}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-134152 β checkpoint-134775}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4f759d99f99404b3a158e7cf157370a55dacba03001073575daa717076ee79c9
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-134152 β checkpoint-134775}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:905f71fc4b7e73652c257064e364d2d8917dba2a6b0c4b1b7ff17c22dca783c5
|
| 3 |
size 14503
|
model-bin/finetune/base/{checkpoint-134152 β checkpoint-134775}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0384c1217168999b6e697cddefff8cc23fa43811673e445b4420748a109f8863
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-134152 β checkpoint-134775}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ddcb026dd2f7cbaec96bd348a12523cc69ccd259b9ad1e83e85474dac67e511b
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-134152 β checkpoint-134775}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1689111747851003,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-132910",
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -270048,11 +270048,806 @@
|
|
| 270048 |
"eval_steps_per_second": 0.741,
|
| 270049 |
"eval_wer": 0.1819801839047687,
|
| 270050 |
"step": 134152
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 270051 |
}
|
| 270052 |
],
|
| 270053 |
-
"max_steps":
|
| 270054 |
"num_train_epochs": 5000,
|
| 270055 |
-
"total_flos": 3.
|
| 270056 |
"trial_name": null,
|
| 270057 |
"trial_params": null
|
| 270058 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1689111747851003,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-132910",
|
| 4 |
+
"epoch": 1077.996015936255,
|
| 5 |
+
"global_step": 134775,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 270048 |
"eval_steps_per_second": 0.741,
|
| 270049 |
"eval_wer": 0.1819801839047687,
|
| 270050 |
"step": 134152
|
| 270051 |
+
},
|
| 270052 |
+
{
|
| 270053 |
+
"epoch": 1073.02,
|
| 270054 |
+
"learning_rate": 7.86695512820513e-06,
|
| 270055 |
+
"loss": 0.3217,
|
| 270056 |
+
"step": 134155
|
| 270057 |
+
},
|
| 270058 |
+
{
|
| 270059 |
+
"epoch": 1073.06,
|
| 270060 |
+
"learning_rate": 7.866875e-06,
|
| 270061 |
+
"loss": 0.2704,
|
| 270062 |
+
"step": 134160
|
| 270063 |
+
},
|
| 270064 |
+
{
|
| 270065 |
+
"epoch": 1073.1,
|
| 270066 |
+
"learning_rate": 7.866794871794872e-06,
|
| 270067 |
+
"loss": 0.2893,
|
| 270068 |
+
"step": 134165
|
| 270069 |
+
},
|
| 270070 |
+
{
|
| 270071 |
+
"epoch": 1073.14,
|
| 270072 |
+
"learning_rate": 7.866714743589744e-06,
|
| 270073 |
+
"loss": 0.3147,
|
| 270074 |
+
"step": 134170
|
| 270075 |
+
},
|
| 270076 |
+
{
|
| 270077 |
+
"epoch": 1073.18,
|
| 270078 |
+
"learning_rate": 7.866634615384617e-06,
|
| 270079 |
+
"loss": 0.6938,
|
| 270080 |
+
"step": 134175
|
| 270081 |
+
},
|
| 270082 |
+
{
|
| 270083 |
+
"epoch": 1073.22,
|
| 270084 |
+
"learning_rate": 7.866554487179488e-06,
|
| 270085 |
+
"loss": 0.8189,
|
| 270086 |
+
"step": 134180
|
| 270087 |
+
},
|
| 270088 |
+
{
|
| 270089 |
+
"epoch": 1073.26,
|
| 270090 |
+
"learning_rate": 7.86647435897436e-06,
|
| 270091 |
+
"loss": 0.2561,
|
| 270092 |
+
"step": 134185
|
| 270093 |
+
},
|
| 270094 |
+
{
|
| 270095 |
+
"epoch": 1073.3,
|
| 270096 |
+
"learning_rate": 7.86639423076923e-06,
|
| 270097 |
+
"loss": 0.2882,
|
| 270098 |
+
"step": 134190
|
| 270099 |
+
},
|
| 270100 |
+
{
|
| 270101 |
+
"epoch": 1073.34,
|
| 270102 |
+
"learning_rate": 7.866314102564104e-06,
|
| 270103 |
+
"loss": 0.3388,
|
| 270104 |
+
"step": 134195
|
| 270105 |
+
},
|
| 270106 |
+
{
|
| 270107 |
+
"epoch": 1073.38,
|
| 270108 |
+
"learning_rate": 7.866233974358975e-06,
|
| 270109 |
+
"loss": 0.7409,
|
| 270110 |
+
"step": 134200
|
| 270111 |
+
},
|
| 270112 |
+
{
|
| 270113 |
+
"epoch": 1073.42,
|
| 270114 |
+
"learning_rate": 7.866153846153846e-06,
|
| 270115 |
+
"loss": 0.8674,
|
| 270116 |
+
"step": 134205
|
| 270117 |
+
},
|
| 270118 |
+
{
|
| 270119 |
+
"epoch": 1073.46,
|
| 270120 |
+
"learning_rate": 7.86607371794872e-06,
|
| 270121 |
+
"loss": 0.2927,
|
| 270122 |
+
"step": 134210
|
| 270123 |
+
},
|
| 270124 |
+
{
|
| 270125 |
+
"epoch": 1073.5,
|
| 270126 |
+
"learning_rate": 7.86599358974359e-06,
|
| 270127 |
+
"loss": 0.2744,
|
| 270128 |
+
"step": 134215
|
| 270129 |
+
},
|
| 270130 |
+
{
|
| 270131 |
+
"epoch": 1073.54,
|
| 270132 |
+
"learning_rate": 7.865913461538462e-06,
|
| 270133 |
+
"loss": 0.2908,
|
| 270134 |
+
"step": 134220
|
| 270135 |
+
},
|
| 270136 |
+
{
|
| 270137 |
+
"epoch": 1073.58,
|
| 270138 |
+
"learning_rate": 7.865833333333334e-06,
|
| 270139 |
+
"loss": 0.6822,
|
| 270140 |
+
"step": 134225
|
| 270141 |
+
},
|
| 270142 |
+
{
|
| 270143 |
+
"epoch": 1073.62,
|
| 270144 |
+
"learning_rate": 7.865753205128207e-06,
|
| 270145 |
+
"loss": 0.7908,
|
| 270146 |
+
"step": 134230
|
| 270147 |
+
},
|
| 270148 |
+
{
|
| 270149 |
+
"epoch": 1073.66,
|
| 270150 |
+
"learning_rate": 7.865673076923078e-06,
|
| 270151 |
+
"loss": 0.2532,
|
| 270152 |
+
"step": 134235
|
| 270153 |
+
},
|
| 270154 |
+
{
|
| 270155 |
+
"epoch": 1073.7,
|
| 270156 |
+
"learning_rate": 7.86559294871795e-06,
|
| 270157 |
+
"loss": 0.3056,
|
| 270158 |
+
"step": 134240
|
| 270159 |
+
},
|
| 270160 |
+
{
|
| 270161 |
+
"epoch": 1073.74,
|
| 270162 |
+
"learning_rate": 7.86551282051282e-06,
|
| 270163 |
+
"loss": 0.4252,
|
| 270164 |
+
"step": 134245
|
| 270165 |
+
},
|
| 270166 |
+
{
|
| 270167 |
+
"epoch": 1073.78,
|
| 270168 |
+
"learning_rate": 7.865432692307694e-06,
|
| 270169 |
+
"loss": 0.6656,
|
| 270170 |
+
"step": 134250
|
| 270171 |
+
},
|
| 270172 |
+
{
|
| 270173 |
+
"epoch": 1073.82,
|
| 270174 |
+
"learning_rate": 7.865352564102565e-06,
|
| 270175 |
+
"loss": 0.8963,
|
| 270176 |
+
"step": 134255
|
| 270177 |
+
},
|
| 270178 |
+
{
|
| 270179 |
+
"epoch": 1073.86,
|
| 270180 |
+
"learning_rate": 7.865272435897436e-06,
|
| 270181 |
+
"loss": 0.3141,
|
| 270182 |
+
"step": 134260
|
| 270183 |
+
},
|
| 270184 |
+
{
|
| 270185 |
+
"epoch": 1073.9,
|
| 270186 |
+
"learning_rate": 7.86519230769231e-06,
|
| 270187 |
+
"loss": 0.3086,
|
| 270188 |
+
"step": 134265
|
| 270189 |
+
},
|
| 270190 |
+
{
|
| 270191 |
+
"epoch": 1073.94,
|
| 270192 |
+
"learning_rate": 7.86511217948718e-06,
|
| 270193 |
+
"loss": 0.4083,
|
| 270194 |
+
"step": 134270
|
| 270195 |
+
},
|
| 270196 |
+
{
|
| 270197 |
+
"epoch": 1073.98,
|
| 270198 |
+
"learning_rate": 7.865032051282052e-06,
|
| 270199 |
+
"loss": 0.7098,
|
| 270200 |
+
"step": 134275
|
| 270201 |
+
},
|
| 270202 |
+
{
|
| 270203 |
+
"epoch": 1074.0,
|
| 270204 |
+
"eval_loss": 0.38545867800712585,
|
| 270205 |
+
"eval_runtime": 36.5735,
|
| 270206 |
+
"eval_samples_per_second": 22.831,
|
| 270207 |
+
"eval_steps_per_second": 0.738,
|
| 270208 |
+
"eval_wer": 0.18589098230979434,
|
| 270209 |
+
"step": 134277
|
| 270210 |
+
},
|
| 270211 |
+
{
|
| 270212 |
+
"epoch": 1074.02,
|
| 270213 |
+
"learning_rate": 7.864951923076924e-06,
|
| 270214 |
+
"loss": 0.3174,
|
| 270215 |
+
"step": 134280
|
| 270216 |
+
},
|
| 270217 |
+
{
|
| 270218 |
+
"epoch": 1074.06,
|
| 270219 |
+
"learning_rate": 7.864871794871795e-06,
|
| 270220 |
+
"loss": 0.2214,
|
| 270221 |
+
"step": 134285
|
| 270222 |
+
},
|
| 270223 |
+
{
|
| 270224 |
+
"epoch": 1074.1,
|
| 270225 |
+
"learning_rate": 7.864791666666666e-06,
|
| 270226 |
+
"loss": 0.2406,
|
| 270227 |
+
"step": 134290
|
| 270228 |
+
},
|
| 270229 |
+
{
|
| 270230 |
+
"epoch": 1074.14,
|
| 270231 |
+
"learning_rate": 7.86471153846154e-06,
|
| 270232 |
+
"loss": 0.3307,
|
| 270233 |
+
"step": 134295
|
| 270234 |
+
},
|
| 270235 |
+
{
|
| 270236 |
+
"epoch": 1074.18,
|
| 270237 |
+
"learning_rate": 7.86463141025641e-06,
|
| 270238 |
+
"loss": 0.7062,
|
| 270239 |
+
"step": 134300
|
| 270240 |
+
},
|
| 270241 |
+
{
|
| 270242 |
+
"epoch": 1074.22,
|
| 270243 |
+
"learning_rate": 7.864551282051282e-06,
|
| 270244 |
+
"loss": 0.913,
|
| 270245 |
+
"step": 134305
|
| 270246 |
+
},
|
| 270247 |
+
{
|
| 270248 |
+
"epoch": 1074.26,
|
| 270249 |
+
"learning_rate": 7.864471153846155e-06,
|
| 270250 |
+
"loss": 0.2612,
|
| 270251 |
+
"step": 134310
|
| 270252 |
+
},
|
| 270253 |
+
{
|
| 270254 |
+
"epoch": 1074.3,
|
| 270255 |
+
"learning_rate": 7.864391025641026e-06,
|
| 270256 |
+
"loss": 0.3045,
|
| 270257 |
+
"step": 134315
|
| 270258 |
+
},
|
| 270259 |
+
{
|
| 270260 |
+
"epoch": 1074.34,
|
| 270261 |
+
"learning_rate": 7.864310897435898e-06,
|
| 270262 |
+
"loss": 0.3602,
|
| 270263 |
+
"step": 134320
|
| 270264 |
+
},
|
| 270265 |
+
{
|
| 270266 |
+
"epoch": 1074.38,
|
| 270267 |
+
"learning_rate": 7.86423076923077e-06,
|
| 270268 |
+
"loss": 0.7683,
|
| 270269 |
+
"step": 134325
|
| 270270 |
+
},
|
| 270271 |
+
{
|
| 270272 |
+
"epoch": 1074.42,
|
| 270273 |
+
"learning_rate": 7.864150641025642e-06,
|
| 270274 |
+
"loss": 0.8546,
|
| 270275 |
+
"step": 134330
|
| 270276 |
+
},
|
| 270277 |
+
{
|
| 270278 |
+
"epoch": 1074.46,
|
| 270279 |
+
"learning_rate": 7.864070512820514e-06,
|
| 270280 |
+
"loss": 0.3199,
|
| 270281 |
+
"step": 134335
|
| 270282 |
+
},
|
| 270283 |
+
{
|
| 270284 |
+
"epoch": 1074.5,
|
| 270285 |
+
"learning_rate": 7.863990384615385e-06,
|
| 270286 |
+
"loss": 0.3023,
|
| 270287 |
+
"step": 134340
|
| 270288 |
+
},
|
| 270289 |
+
{
|
| 270290 |
+
"epoch": 1074.54,
|
| 270291 |
+
"learning_rate": 7.863910256410256e-06,
|
| 270292 |
+
"loss": 0.3521,
|
| 270293 |
+
"step": 134345
|
| 270294 |
+
},
|
| 270295 |
+
{
|
| 270296 |
+
"epoch": 1074.58,
|
| 270297 |
+
"learning_rate": 7.86383012820513e-06,
|
| 270298 |
+
"loss": 0.919,
|
| 270299 |
+
"step": 134350
|
| 270300 |
+
},
|
| 270301 |
+
{
|
| 270302 |
+
"epoch": 1074.62,
|
| 270303 |
+
"learning_rate": 7.86375e-06,
|
| 270304 |
+
"loss": 0.8259,
|
| 270305 |
+
"step": 134355
|
| 270306 |
+
},
|
| 270307 |
+
{
|
| 270308 |
+
"epoch": 1074.66,
|
| 270309 |
+
"learning_rate": 7.863669871794872e-06,
|
| 270310 |
+
"loss": 0.292,
|
| 270311 |
+
"step": 134360
|
| 270312 |
+
},
|
| 270313 |
+
{
|
| 270314 |
+
"epoch": 1074.7,
|
| 270315 |
+
"learning_rate": 7.863589743589745e-06,
|
| 270316 |
+
"loss": 0.2807,
|
| 270317 |
+
"step": 134365
|
| 270318 |
+
},
|
| 270319 |
+
{
|
| 270320 |
+
"epoch": 1074.74,
|
| 270321 |
+
"learning_rate": 7.863509615384617e-06,
|
| 270322 |
+
"loss": 0.2971,
|
| 270323 |
+
"step": 134370
|
| 270324 |
+
},
|
| 270325 |
+
{
|
| 270326 |
+
"epoch": 1074.78,
|
| 270327 |
+
"learning_rate": 7.863429487179488e-06,
|
| 270328 |
+
"loss": 0.6238,
|
| 270329 |
+
"step": 134375
|
| 270330 |
+
},
|
| 270331 |
+
{
|
| 270332 |
+
"epoch": 1074.82,
|
| 270333 |
+
"learning_rate": 7.86334935897436e-06,
|
| 270334 |
+
"loss": 0.746,
|
| 270335 |
+
"step": 134380
|
| 270336 |
+
},
|
| 270337 |
+
{
|
| 270338 |
+
"epoch": 1074.86,
|
| 270339 |
+
"learning_rate": 7.863269230769232e-06,
|
| 270340 |
+
"loss": 0.2581,
|
| 270341 |
+
"step": 134385
|
| 270342 |
+
},
|
| 270343 |
+
{
|
| 270344 |
+
"epoch": 1074.9,
|
| 270345 |
+
"learning_rate": 7.863189102564102e-06,
|
| 270346 |
+
"loss": 0.2956,
|
| 270347 |
+
"step": 134390
|
| 270348 |
+
},
|
| 270349 |
+
{
|
| 270350 |
+
"epoch": 1074.94,
|
| 270351 |
+
"learning_rate": 7.863108974358975e-06,
|
| 270352 |
+
"loss": 0.3957,
|
| 270353 |
+
"step": 134395
|
| 270354 |
+
},
|
| 270355 |
+
{
|
| 270356 |
+
"epoch": 1074.98,
|
| 270357 |
+
"learning_rate": 7.863028846153848e-06,
|
| 270358 |
+
"loss": 0.787,
|
| 270359 |
+
"step": 134400
|
| 270360 |
+
},
|
| 270361 |
+
{
|
| 270362 |
+
"epoch": 1075.0,
|
| 270363 |
+
"eval_loss": 0.3511001765727997,
|
| 270364 |
+
"eval_runtime": 35.7959,
|
| 270365 |
+
"eval_samples_per_second": 23.327,
|
| 270366 |
+
"eval_steps_per_second": 0.754,
|
| 270367 |
+
"eval_wer": 0.17502042028662657,
|
| 270368 |
+
"step": 134402
|
| 270369 |
+
},
|
| 270370 |
+
{
|
| 270371 |
+
"epoch": 1083.02,
|
| 270372 |
+
"learning_rate": 7.862948717948718e-06,
|
| 270373 |
+
"loss": 0.3142,
|
| 270374 |
+
"step": 134405
|
| 270375 |
+
},
|
| 270376 |
+
{
|
| 270377 |
+
"epoch": 1083.06,
|
| 270378 |
+
"learning_rate": 7.86286858974359e-06,
|
| 270379 |
+
"loss": 0.2986,
|
| 270380 |
+
"step": 134410
|
| 270381 |
+
},
|
| 270382 |
+
{
|
| 270383 |
+
"epoch": 1083.1,
|
| 270384 |
+
"learning_rate": 7.862788461538462e-06,
|
| 270385 |
+
"loss": 0.3211,
|
| 270386 |
+
"step": 134415
|
| 270387 |
+
},
|
| 270388 |
+
{
|
| 270389 |
+
"epoch": 1083.14,
|
| 270390 |
+
"learning_rate": 7.862708333333333e-06,
|
| 270391 |
+
"loss": 0.33,
|
| 270392 |
+
"step": 134420
|
| 270393 |
+
},
|
| 270394 |
+
{
|
| 270395 |
+
"epoch": 1083.18,
|
| 270396 |
+
"learning_rate": 7.862628205128205e-06,
|
| 270397 |
+
"loss": 0.6979,
|
| 270398 |
+
"step": 134425
|
| 270399 |
+
},
|
| 270400 |
+
{
|
| 270401 |
+
"epoch": 1083.22,
|
| 270402 |
+
"learning_rate": 7.862548076923078e-06,
|
| 270403 |
+
"loss": 0.8636,
|
| 270404 |
+
"step": 134430
|
| 270405 |
+
},
|
| 270406 |
+
{
|
| 270407 |
+
"epoch": 1083.27,
|
| 270408 |
+
"learning_rate": 7.86246794871795e-06,
|
| 270409 |
+
"loss": 0.2713,
|
| 270410 |
+
"step": 134435
|
| 270411 |
+
},
|
| 270412 |
+
{
|
| 270413 |
+
"epoch": 1083.31,
|
| 270414 |
+
"learning_rate": 7.86238782051282e-06,
|
| 270415 |
+
"loss": 0.2765,
|
| 270416 |
+
"step": 134440
|
| 270417 |
+
},
|
| 270418 |
+
{
|
| 270419 |
+
"epoch": 1083.35,
|
| 270420 |
+
"learning_rate": 7.862307692307692e-06,
|
| 270421 |
+
"loss": 0.3689,
|
| 270422 |
+
"step": 134445
|
| 270423 |
+
},
|
| 270424 |
+
{
|
| 270425 |
+
"epoch": 1083.39,
|
| 270426 |
+
"learning_rate": 7.862227564102565e-06,
|
| 270427 |
+
"loss": 0.6944,
|
| 270428 |
+
"step": 134450
|
| 270429 |
+
},
|
| 270430 |
+
{
|
| 270431 |
+
"epoch": 1083.43,
|
| 270432 |
+
"learning_rate": 7.862147435897436e-06,
|
| 270433 |
+
"loss": 0.8387,
|
| 270434 |
+
"step": 134455
|
| 270435 |
+
},
|
| 270436 |
+
{
|
| 270437 |
+
"epoch": 1083.47,
|
| 270438 |
+
"learning_rate": 7.862067307692308e-06,
|
| 270439 |
+
"loss": 0.288,
|
| 270440 |
+
"step": 134460
|
| 270441 |
+
},
|
| 270442 |
+
{
|
| 270443 |
+
"epoch": 1083.51,
|
| 270444 |
+
"learning_rate": 7.86198717948718e-06,
|
| 270445 |
+
"loss": 0.335,
|
| 270446 |
+
"step": 134465
|
| 270447 |
+
},
|
| 270448 |
+
{
|
| 270449 |
+
"epoch": 1083.55,
|
| 270450 |
+
"learning_rate": 7.861907051282052e-06,
|
| 270451 |
+
"loss": 0.332,
|
| 270452 |
+
"step": 134470
|
| 270453 |
+
},
|
| 270454 |
+
{
|
| 270455 |
+
"epoch": 1083.59,
|
| 270456 |
+
"learning_rate": 7.861826923076924e-06,
|
| 270457 |
+
"loss": 0.6622,
|
| 270458 |
+
"step": 134475
|
| 270459 |
+
},
|
| 270460 |
+
{
|
| 270461 |
+
"epoch": 1083.63,
|
| 270462 |
+
"learning_rate": 7.861746794871795e-06,
|
| 270463 |
+
"loss": 0.7635,
|
| 270464 |
+
"step": 134480
|
| 270465 |
+
},
|
| 270466 |
+
{
|
| 270467 |
+
"epoch": 1083.67,
|
| 270468 |
+
"learning_rate": 7.861666666666668e-06,
|
| 270469 |
+
"loss": 0.2669,
|
| 270470 |
+
"step": 134485
|
| 270471 |
+
},
|
| 270472 |
+
{
|
| 270473 |
+
"epoch": 1083.71,
|
| 270474 |
+
"learning_rate": 7.86158653846154e-06,
|
| 270475 |
+
"loss": 0.3005,
|
| 270476 |
+
"step": 134490
|
| 270477 |
+
},
|
| 270478 |
+
{
|
| 270479 |
+
"epoch": 1083.75,
|
| 270480 |
+
"learning_rate": 7.86150641025641e-06,
|
| 270481 |
+
"loss": 0.303,
|
| 270482 |
+
"step": 134495
|
| 270483 |
+
},
|
| 270484 |
+
{
|
| 270485 |
+
"epoch": 1083.79,
|
| 270486 |
+
"learning_rate": 7.861426282051284e-06,
|
| 270487 |
+
"loss": 0.7309,
|
| 270488 |
+
"step": 134500
|
| 270489 |
+
},
|
| 270490 |
+
{
|
| 270491 |
+
"epoch": 1083.83,
|
| 270492 |
+
"learning_rate": 7.861346153846155e-06,
|
| 270493 |
+
"loss": 0.7374,
|
| 270494 |
+
"step": 134505
|
| 270495 |
+
},
|
| 270496 |
+
{
|
| 270497 |
+
"epoch": 1083.87,
|
| 270498 |
+
"learning_rate": 7.861266025641026e-06,
|
| 270499 |
+
"loss": 0.2363,
|
| 270500 |
+
"step": 134510
|
| 270501 |
+
},
|
| 270502 |
+
{
|
| 270503 |
+
"epoch": 1083.91,
|
| 270504 |
+
"learning_rate": 7.861185897435898e-06,
|
| 270505 |
+
"loss": 0.2767,
|
| 270506 |
+
"step": 134515
|
| 270507 |
+
},
|
| 270508 |
+
{
|
| 270509 |
+
"epoch": 1083.95,
|
| 270510 |
+
"learning_rate": 7.86110576923077e-06,
|
| 270511 |
+
"loss": 0.3834,
|
| 270512 |
+
"step": 134520
|
| 270513 |
+
},
|
| 270514 |
+
{
|
| 270515 |
+
"epoch": 1083.99,
|
| 270516 |
+
"learning_rate": 7.861025641025642e-06,
|
| 270517 |
+
"loss": 0.7469,
|
| 270518 |
+
"step": 134525
|
| 270519 |
+
},
|
| 270520 |
+
{
|
| 270521 |
+
"epoch": 1084.0,
|
| 270522 |
+
"eval_loss": 0.3184202015399933,
|
| 270523 |
+
"eval_runtime": 35.9608,
|
| 270524 |
+
"eval_samples_per_second": 23.22,
|
| 270525 |
+
"eval_steps_per_second": 0.751,
|
| 270526 |
+
"eval_wer": 0.17747193045997828,
|
| 270527 |
+
"step": 134526
|
| 270528 |
+
},
|
| 270529 |
+
{
|
| 270530 |
+
"epoch": 1084.03,
|
| 270531 |
+
"learning_rate": 7.860945512820514e-06,
|
| 270532 |
+
"loss": 0.302,
|
| 270533 |
+
"step": 134530
|
| 270534 |
+
},
|
| 270535 |
+
{
|
| 270536 |
+
"epoch": 1084.07,
|
| 270537 |
+
"learning_rate": 7.860865384615385e-06,
|
| 270538 |
+
"loss": 0.278,
|
| 270539 |
+
"step": 134535
|
| 270540 |
+
},
|
| 270541 |
+
{
|
| 270542 |
+
"epoch": 1084.11,
|
| 270543 |
+
"learning_rate": 7.860785256410258e-06,
|
| 270544 |
+
"loss": 0.3627,
|
| 270545 |
+
"step": 134540
|
| 270546 |
+
},
|
| 270547 |
+
{
|
| 270548 |
+
"epoch": 1084.15,
|
| 270549 |
+
"learning_rate": 7.860705128205128e-06,
|
| 270550 |
+
"loss": 0.407,
|
| 270551 |
+
"step": 134545
|
| 270552 |
+
},
|
| 270553 |
+
{
|
| 270554 |
+
"epoch": 1084.19,
|
| 270555 |
+
"learning_rate": 7.860625e-06,
|
| 270556 |
+
"loss": 0.8367,
|
| 270557 |
+
"step": 134550
|
| 270558 |
+
},
|
| 270559 |
+
{
|
| 270560 |
+
"epoch": 1084.23,
|
| 270561 |
+
"learning_rate": 7.860544871794874e-06,
|
| 270562 |
+
"loss": 0.7498,
|
| 270563 |
+
"step": 134555
|
| 270564 |
+
},
|
| 270565 |
+
{
|
| 270566 |
+
"epoch": 1084.27,
|
| 270567 |
+
"learning_rate": 7.860464743589743e-06,
|
| 270568 |
+
"loss": 0.3296,
|
| 270569 |
+
"step": 134560
|
| 270570 |
+
},
|
| 270571 |
+
{
|
| 270572 |
+
"epoch": 1084.31,
|
| 270573 |
+
"learning_rate": 7.860384615384616e-06,
|
| 270574 |
+
"loss": 0.2877,
|
| 270575 |
+
"step": 134565
|
| 270576 |
+
},
|
| 270577 |
+
{
|
| 270578 |
+
"epoch": 1084.35,
|
| 270579 |
+
"learning_rate": 7.860304487179488e-06,
|
| 270580 |
+
"loss": 0.4052,
|
| 270581 |
+
"step": 134570
|
| 270582 |
+
},
|
| 270583 |
+
{
|
| 270584 |
+
"epoch": 1084.39,
|
| 270585 |
+
"learning_rate": 7.860224358974359e-06,
|
| 270586 |
+
"loss": 0.9111,
|
| 270587 |
+
"step": 134575
|
| 270588 |
+
},
|
| 270589 |
+
{
|
| 270590 |
+
"epoch": 1084.43,
|
| 270591 |
+
"learning_rate": 7.86014423076923e-06,
|
| 270592 |
+
"loss": 0.7183,
|
| 270593 |
+
"step": 134580
|
| 270594 |
+
},
|
| 270595 |
+
{
|
| 270596 |
+
"epoch": 1084.47,
|
| 270597 |
+
"learning_rate": 7.860064102564104e-06,
|
| 270598 |
+
"loss": 0.2397,
|
| 270599 |
+
"step": 134585
|
| 270600 |
+
},
|
| 270601 |
+
{
|
| 270602 |
+
"epoch": 1084.51,
|
| 270603 |
+
"learning_rate": 7.859983974358975e-06,
|
| 270604 |
+
"loss": 0.2869,
|
| 270605 |
+
"step": 134590
|
| 270606 |
+
},
|
| 270607 |
+
{
|
| 270608 |
+
"epoch": 1084.55,
|
| 270609 |
+
"learning_rate": 7.859903846153846e-06,
|
| 270610 |
+
"loss": 0.3869,
|
| 270611 |
+
"step": 134595
|
| 270612 |
+
},
|
| 270613 |
+
{
|
| 270614 |
+
"epoch": 1084.59,
|
| 270615 |
+
"learning_rate": 7.85982371794872e-06,
|
| 270616 |
+
"loss": 0.8341,
|
| 270617 |
+
"step": 134600
|
| 270618 |
+
},
|
| 270619 |
+
{
|
| 270620 |
+
"epoch": 1084.63,
|
| 270621 |
+
"learning_rate": 7.85974358974359e-06,
|
| 270622 |
+
"loss": 0.5808,
|
| 270623 |
+
"step": 134605
|
| 270624 |
+
},
|
| 270625 |
+
{
|
| 270626 |
+
"epoch": 1084.67,
|
| 270627 |
+
"learning_rate": 7.859663461538462e-06,
|
| 270628 |
+
"loss": 0.3174,
|
| 270629 |
+
"step": 134610
|
| 270630 |
+
},
|
| 270631 |
+
{
|
| 270632 |
+
"epoch": 1084.71,
|
| 270633 |
+
"learning_rate": 7.859583333333333e-06,
|
| 270634 |
+
"loss": 0.2914,
|
| 270635 |
+
"step": 134615
|
| 270636 |
+
},
|
| 270637 |
+
{
|
| 270638 |
+
"epoch": 1084.76,
|
| 270639 |
+
"learning_rate": 7.859503205128206e-06,
|
| 270640 |
+
"loss": 0.3991,
|
| 270641 |
+
"step": 134620
|
| 270642 |
+
},
|
| 270643 |
+
{
|
| 270644 |
+
"epoch": 1084.8,
|
| 270645 |
+
"learning_rate": 7.859423076923078e-06,
|
| 270646 |
+
"loss": 0.8859,
|
| 270647 |
+
"step": 134625
|
| 270648 |
+
},
|
| 270649 |
+
{
|
| 270650 |
+
"epoch": 1084.84,
|
| 270651 |
+
"learning_rate": 7.85934294871795e-06,
|
| 270652 |
+
"loss": 0.6037,
|
| 270653 |
+
"step": 134630
|
| 270654 |
+
},
|
| 270655 |
+
{
|
| 270656 |
+
"epoch": 1084.88,
|
| 270657 |
+
"learning_rate": 7.85926282051282e-06,
|
| 270658 |
+
"loss": 0.2301,
|
| 270659 |
+
"step": 134635
|
| 270660 |
+
},
|
| 270661 |
+
{
|
| 270662 |
+
"epoch": 1084.92,
|
| 270663 |
+
"learning_rate": 7.859182692307694e-06,
|
| 270664 |
+
"loss": 0.302,
|
| 270665 |
+
"step": 134640
|
| 270666 |
+
},
|
| 270667 |
+
{
|
| 270668 |
+
"epoch": 1084.96,
|
| 270669 |
+
"learning_rate": 7.859102564102565e-06,
|
| 270670 |
+
"loss": 0.4093,
|
| 270671 |
+
"step": 134645
|
| 270672 |
+
},
|
| 270673 |
+
{
|
| 270674 |
+
"epoch": 1085.0,
|
| 270675 |
+
"learning_rate": 7.859022435897436e-06,
|
| 270676 |
+
"loss": 1.0476,
|
| 270677 |
+
"step": 134650
|
| 270678 |
+
},
|
| 270679 |
+
{
|
| 270680 |
+
"epoch": 1085.0,
|
| 270681 |
+
"eval_loss": 0.3662145137786865,
|
| 270682 |
+
"eval_runtime": 35.6491,
|
| 270683 |
+
"eval_samples_per_second": 23.423,
|
| 270684 |
+
"eval_steps_per_second": 0.757,
|
| 270685 |
+
"eval_wer": 0.17833308616342874,
|
| 270686 |
+
"step": 134650
|
| 270687 |
+
},
|
| 270688 |
+
{
|
| 270689 |
+
"epoch": 1077.04,
|
| 270690 |
+
"learning_rate": 7.85894230769231e-06,
|
| 270691 |
+
"loss": 0.3193,
|
| 270692 |
+
"step": 134655
|
| 270693 |
+
},
|
| 270694 |
+
{
|
| 270695 |
+
"epoch": 1077.08,
|
| 270696 |
+
"learning_rate": 7.85886217948718e-06,
|
| 270697 |
+
"loss": 0.3212,
|
| 270698 |
+
"step": 134660
|
| 270699 |
+
},
|
| 270700 |
+
{
|
| 270701 |
+
"epoch": 1077.12,
|
| 270702 |
+
"learning_rate": 7.858782051282052e-06,
|
| 270703 |
+
"loss": 0.3012,
|
| 270704 |
+
"step": 134665
|
| 270705 |
+
},
|
| 270706 |
+
{
|
| 270707 |
+
"epoch": 1077.16,
|
| 270708 |
+
"learning_rate": 7.858701923076923e-06,
|
| 270709 |
+
"loss": 0.4692,
|
| 270710 |
+
"step": 134670
|
| 270711 |
+
},
|
| 270712 |
+
{
|
| 270713 |
+
"epoch": 1077.2,
|
| 270714 |
+
"learning_rate": 7.858621794871796e-06,
|
| 270715 |
+
"loss": 1.165,
|
| 270716 |
+
"step": 134675
|
| 270717 |
+
},
|
| 270718 |
+
{
|
| 270719 |
+
"epoch": 1077.24,
|
| 270720 |
+
"learning_rate": 7.858541666666666e-06,
|
| 270721 |
+
"loss": 0.2905,
|
| 270722 |
+
"step": 134680
|
| 270723 |
+
},
|
| 270724 |
+
{
|
| 270725 |
+
"epoch": 1077.28,
|
| 270726 |
+
"learning_rate": 7.85846153846154e-06,
|
| 270727 |
+
"loss": 0.2562,
|
| 270728 |
+
"step": 134685
|
| 270729 |
+
},
|
| 270730 |
+
{
|
| 270731 |
+
"epoch": 1077.32,
|
| 270732 |
+
"learning_rate": 7.85838141025641e-06,
|
| 270733 |
+
"loss": 0.2847,
|
| 270734 |
+
"step": 134690
|
| 270735 |
+
},
|
| 270736 |
+
{
|
| 270737 |
+
"epoch": 1077.36,
|
| 270738 |
+
"learning_rate": 7.858301282051282e-06,
|
| 270739 |
+
"loss": 0.4418,
|
| 270740 |
+
"step": 134695
|
| 270741 |
+
},
|
| 270742 |
+
{
|
| 270743 |
+
"epoch": 1077.4,
|
| 270744 |
+
"learning_rate": 7.858221153846155e-06,
|
| 270745 |
+
"loss": 1.1527,
|
| 270746 |
+
"step": 134700
|
| 270747 |
+
},
|
| 270748 |
+
{
|
| 270749 |
+
"epoch": 1077.44,
|
| 270750 |
+
"learning_rate": 7.858141025641026e-06,
|
| 270751 |
+
"loss": 0.3339,
|
| 270752 |
+
"step": 134705
|
| 270753 |
+
},
|
| 270754 |
+
{
|
| 270755 |
+
"epoch": 1077.48,
|
| 270756 |
+
"learning_rate": 7.858060897435898e-06,
|
| 270757 |
+
"loss": 0.2692,
|
| 270758 |
+
"step": 134710
|
| 270759 |
+
},
|
| 270760 |
+
{
|
| 270761 |
+
"epoch": 1077.52,
|
| 270762 |
+
"learning_rate": 7.857980769230769e-06,
|
| 270763 |
+
"loss": 0.3078,
|
| 270764 |
+
"step": 134715
|
| 270765 |
+
},
|
| 270766 |
+
{
|
| 270767 |
+
"epoch": 1077.56,
|
| 270768 |
+
"learning_rate": 7.857900641025642e-06,
|
| 270769 |
+
"loss": 0.5131,
|
| 270770 |
+
"step": 134720
|
| 270771 |
+
},
|
| 270772 |
+
{
|
| 270773 |
+
"epoch": 1077.6,
|
| 270774 |
+
"learning_rate": 7.857820512820513e-06,
|
| 270775 |
+
"loss": 1.1702,
|
| 270776 |
+
"step": 134725
|
| 270777 |
+
},
|
| 270778 |
+
{
|
| 270779 |
+
"epoch": 1077.64,
|
| 270780 |
+
"learning_rate": 7.857740384615385e-06,
|
| 270781 |
+
"loss": 0.4159,
|
| 270782 |
+
"step": 134730
|
| 270783 |
+
},
|
| 270784 |
+
{
|
| 270785 |
+
"epoch": 1077.68,
|
| 270786 |
+
"learning_rate": 7.857660256410256e-06,
|
| 270787 |
+
"loss": 0.318,
|
| 270788 |
+
"step": 134735
|
| 270789 |
+
},
|
| 270790 |
+
{
|
| 270791 |
+
"epoch": 1077.72,
|
| 270792 |
+
"learning_rate": 7.85758012820513e-06,
|
| 270793 |
+
"loss": 0.2503,
|
| 270794 |
+
"step": 134740
|
| 270795 |
+
},
|
| 270796 |
+
{
|
| 270797 |
+
"epoch": 1077.76,
|
| 270798 |
+
"learning_rate": 7.8575e-06,
|
| 270799 |
+
"loss": 0.4202,
|
| 270800 |
+
"step": 134745
|
| 270801 |
+
},
|
| 270802 |
+
{
|
| 270803 |
+
"epoch": 1077.8,
|
| 270804 |
+
"learning_rate": 7.857419871794872e-06,
|
| 270805 |
+
"loss": 1.2349,
|
| 270806 |
+
"step": 134750
|
| 270807 |
+
},
|
| 270808 |
+
{
|
| 270809 |
+
"epoch": 1077.84,
|
| 270810 |
+
"learning_rate": 7.857339743589745e-06,
|
| 270811 |
+
"loss": 0.352,
|
| 270812 |
+
"step": 134755
|
| 270813 |
+
},
|
| 270814 |
+
{
|
| 270815 |
+
"epoch": 1077.88,
|
| 270816 |
+
"learning_rate": 7.857259615384616e-06,
|
| 270817 |
+
"loss": 0.2831,
|
| 270818 |
+
"step": 134760
|
| 270819 |
+
},
|
| 270820 |
+
{
|
| 270821 |
+
"epoch": 1077.92,
|
| 270822 |
+
"learning_rate": 7.857179487179488e-06,
|
| 270823 |
+
"loss": 0.5545,
|
| 270824 |
+
"step": 134765
|
| 270825 |
+
},
|
| 270826 |
+
{
|
| 270827 |
+
"epoch": 1077.96,
|
| 270828 |
+
"learning_rate": 7.857099358974359e-06,
|
| 270829 |
+
"loss": 0.5264,
|
| 270830 |
+
"step": 134770
|
| 270831 |
+
},
|
| 270832 |
+
{
|
| 270833 |
+
"epoch": 1078.0,
|
| 270834 |
+
"learning_rate": 7.857019230769232e-06,
|
| 270835 |
+
"loss": 1.1571,
|
| 270836 |
+
"step": 134775
|
| 270837 |
+
},
|
| 270838 |
+
{
|
| 270839 |
+
"epoch": 1078.0,
|
| 270840 |
+
"eval_loss": 0.35610613226890564,
|
| 270841 |
+
"eval_runtime": 35.6921,
|
| 270842 |
+
"eval_samples_per_second": 23.395,
|
| 270843 |
+
"eval_steps_per_second": 0.756,
|
| 270844 |
+
"eval_wer": 0.18336918626214674,
|
| 270845 |
+
"step": 134775
|
| 270846 |
}
|
| 270847 |
],
|
| 270848 |
+
"max_steps": 625000,
|
| 270849 |
"num_train_epochs": 5000,
|
| 270850 |
+
"total_flos": 3.7924993154991974e+20,
|
| 270851 |
"trial_name": null,
|
| 270852 |
"trial_params": null
|
| 270853 |
}
|
model-bin/finetune/base/{checkpoint-134152 β checkpoint-134775}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1630175438.376344/events.out.tfevents.1630175438.86bb0ddabf9b.4092.261
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f7c2bbc62c570a5fb31993d881a3c1d72ad1281d95bc75fdca3e0f5143d6bf27
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630175824.2745113/events.out.tfevents.1630175824.86bb0ddabf9b.4092.263
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:78a4a9268e26513b7ea8ea5bfb8467f75b4beca1459892c487c50e6dece31d74
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630176214.4682405/events.out.tfevents.1630176214.86bb0ddabf9b.4092.265
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:657bc50c29e889a2f92a5b7e12ab1b221a9cb1a949cd34b0379297fb33c433de
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630176600.245581/events.out.tfevents.1630176600.86bb0ddabf9b.4092.267
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2a37b54afadeb21e57793f24cbbaf29c87d88d9913a5af3e61dd136eec767fcb
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630176983.8605223/events.out.tfevents.1630176983.86bb0ddabf9b.4092.269
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:39875855dd65a15a96e2204a14bc9f19987c3d18ec16ebeaefe6e67c39d75b02
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630175438.86bb0ddabf9b.4092.260
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:866cac0cd9153ba2e8628440edd8339935fcf779c64c9f6ff71aac5296585973
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630175824.86bb0ddabf9b.4092.262
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ec8d417db3f817d9b2b0f22fe7c46e4300922a1e918814c76abca0935816a120
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630176214.86bb0ddabf9b.4092.264
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e7dd9822cab557ddf2e8dde6705a6639f242db1ee4f2e46e72819c77d097e766
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630176600.86bb0ddabf9b.4092.266
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a81dfc64f3c59dd5fec7d14f8453f692fc34d45bcf8a7ab980ced039810856b2
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630176983.86bb0ddabf9b.4092.268
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:589bcd5ccfd1df2d34418e73cc955a7bb83abf6b41fec6982d0aa4777f0ad99e
|
| 3 |
+
size 8622
|