"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-147718 β checkpoint-148340}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-147718 β checkpoint-148340}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-147718 β checkpoint-148340}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-147718 β checkpoint-148340}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-147718 β checkpoint-148340}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-147718 β checkpoint-148340}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-147718 β checkpoint-148340}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-147718 β checkpoint-148340}/trainer_state.json +799 -4
- model-bin/finetune/base/{checkpoint-147718 β checkpoint-148340}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630235074.4070776/events.out.tfevents.1630235074.cc93b136ebf5.1086.139 +3 -0
- model-bin/finetune/base/log/1630235509.4705393/events.out.tfevents.1630235509.cc93b136ebf5.1086.141 +3 -0
- model-bin/finetune/base/log/1630235947.303649/events.out.tfevents.1630235947.cc93b136ebf5.1086.143 +3 -0
- model-bin/finetune/base/log/1630236380.4833214/events.out.tfevents.1630236380.cc93b136ebf5.1086.145 +3 -0
- model-bin/finetune/base/log/1630236821.4505744/events.out.tfevents.1630236821.cc93b136ebf5.1086.147 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630235074.cc93b136ebf5.1086.138 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630235509.cc93b136ebf5.1086.140 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630235947.cc93b136ebf5.1086.142 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630236380.cc93b136ebf5.1086.144 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630236821.cc93b136ebf5.1086.146 +3 -0
model-bin/finetune/base/{checkpoint-147718 β checkpoint-148340}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-147718 β checkpoint-148340}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ca053a21f43a42ef9039777b97798ad4984877959c3297c934cb79b5f9280b8d
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-147718 β checkpoint-148340}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-147718 β checkpoint-148340}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7f6a1ad1b6bdf94b092b4dbaadb21109adb1abf08755cc31224793164fd555d6
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-147718 β checkpoint-148340}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bc771b6e0d0fae529a3ac50ef6ed92b0fc4e02d966c231e4ea817bf981f64442
|
| 3 |
size 14503
|
model-bin/finetune/base/{checkpoint-147718 β checkpoint-148340}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b54d1b931ef6d5d353a2168a55b89d51f96d90c49dc5e37c1e18abd323651d0d
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-147718 β checkpoint-148340}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:352fbfc64d23a3b022c4a65026575cbcb8c45fcce17227e15ee0f0fca4bb6241
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-147718 β checkpoint-148340}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1668011442822563,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -287307,11 +287307,806 @@
|
|
| 287307 |
"eval_steps_per_second": 0.643,
|
| 287308 |
"eval_wer": 0.17405040041143194,
|
| 287309 |
"step": 147718
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 287310 |
}
|
| 287311 |
],
|
| 287312 |
-
"max_steps":
|
| 287313 |
"num_train_epochs": 5000,
|
| 287314 |
-
"total_flos": 4.
|
| 287315 |
"trial_name": null,
|
| 287316 |
"trial_params": null
|
| 287317 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1668011442822563,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
| 4 |
+
"epoch": 1195.995983935743,
|
| 5 |
+
"global_step": 148340,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 287307 |
"eval_steps_per_second": 0.643,
|
| 287308 |
"eval_wer": 0.17405040041143194,
|
| 287309 |
"step": 147718
|
| 287310 |
+
},
|
| 287311 |
+
{
|
| 287312 |
+
"epoch": 1181.02,
|
| 287313 |
+
"learning_rate": 7.630662358642973e-06,
|
| 287314 |
+
"loss": 0.3152,
|
| 287315 |
+
"step": 147720
|
| 287316 |
+
},
|
| 287317 |
+
{
|
| 287318 |
+
"epoch": 1181.06,
|
| 287319 |
+
"learning_rate": 7.630581583198709e-06,
|
| 287320 |
+
"loss": 0.2892,
|
| 287321 |
+
"step": 147725
|
| 287322 |
+
},
|
| 287323 |
+
{
|
| 287324 |
+
"epoch": 1181.1,
|
| 287325 |
+
"learning_rate": 7.630500807754443e-06,
|
| 287326 |
+
"loss": 0.2558,
|
| 287327 |
+
"step": 147730
|
| 287328 |
+
},
|
| 287329 |
+
{
|
| 287330 |
+
"epoch": 1181.14,
|
| 287331 |
+
"learning_rate": 7.630420032310179e-06,
|
| 287332 |
+
"loss": 0.3198,
|
| 287333 |
+
"step": 147735
|
| 287334 |
+
},
|
| 287335 |
+
{
|
| 287336 |
+
"epoch": 1181.18,
|
| 287337 |
+
"learning_rate": 7.630339256865913e-06,
|
| 287338 |
+
"loss": 0.6625,
|
| 287339 |
+
"step": 147740
|
| 287340 |
+
},
|
| 287341 |
+
{
|
| 287342 |
+
"epoch": 1181.22,
|
| 287343 |
+
"learning_rate": 7.630258481421649e-06,
|
| 287344 |
+
"loss": 0.9346,
|
| 287345 |
+
"step": 147745
|
| 287346 |
+
},
|
| 287347 |
+
{
|
| 287348 |
+
"epoch": 1181.26,
|
| 287349 |
+
"learning_rate": 7.630177705977385e-06,
|
| 287350 |
+
"loss": 0.2707,
|
| 287351 |
+
"step": 147750
|
| 287352 |
+
},
|
| 287353 |
+
{
|
| 287354 |
+
"epoch": 1181.3,
|
| 287355 |
+
"learning_rate": 7.630096930533119e-06,
|
| 287356 |
+
"loss": 0.3105,
|
| 287357 |
+
"step": 147755
|
| 287358 |
+
},
|
| 287359 |
+
{
|
| 287360 |
+
"epoch": 1181.34,
|
| 287361 |
+
"learning_rate": 7.630016155088855e-06,
|
| 287362 |
+
"loss": 0.3038,
|
| 287363 |
+
"step": 147760
|
| 287364 |
+
},
|
| 287365 |
+
{
|
| 287366 |
+
"epoch": 1181.38,
|
| 287367 |
+
"learning_rate": 7.629935379644589e-06,
|
| 287368 |
+
"loss": 0.584,
|
| 287369 |
+
"step": 147765
|
| 287370 |
+
},
|
| 287371 |
+
{
|
| 287372 |
+
"epoch": 1181.42,
|
| 287373 |
+
"learning_rate": 7.629854604200325e-06,
|
| 287374 |
+
"loss": 1.0346,
|
| 287375 |
+
"step": 147770
|
| 287376 |
+
},
|
| 287377 |
+
{
|
| 287378 |
+
"epoch": 1181.46,
|
| 287379 |
+
"learning_rate": 7.629773828756059e-06,
|
| 287380 |
+
"loss": 0.2814,
|
| 287381 |
+
"step": 147775
|
| 287382 |
+
},
|
| 287383 |
+
{
|
| 287384 |
+
"epoch": 1181.5,
|
| 287385 |
+
"learning_rate": 7.629693053311795e-06,
|
| 287386 |
+
"loss": 0.3081,
|
| 287387 |
+
"step": 147780
|
| 287388 |
+
},
|
| 287389 |
+
{
|
| 287390 |
+
"epoch": 1181.54,
|
| 287391 |
+
"learning_rate": 7.629612277867529e-06,
|
| 287392 |
+
"loss": 0.3303,
|
| 287393 |
+
"step": 147785
|
| 287394 |
+
},
|
| 287395 |
+
{
|
| 287396 |
+
"epoch": 1181.58,
|
| 287397 |
+
"learning_rate": 7.629531502423265e-06,
|
| 287398 |
+
"loss": 0.6087,
|
| 287399 |
+
"step": 147790
|
| 287400 |
+
},
|
| 287401 |
+
{
|
| 287402 |
+
"epoch": 1181.62,
|
| 287403 |
+
"learning_rate": 7.629450726978999e-06,
|
| 287404 |
+
"loss": 1.0368,
|
| 287405 |
+
"step": 147795
|
| 287406 |
+
},
|
| 287407 |
+
{
|
| 287408 |
+
"epoch": 1181.66,
|
| 287409 |
+
"learning_rate": 7.6293699515347346e-06,
|
| 287410 |
+
"loss": 0.2708,
|
| 287411 |
+
"step": 147800
|
| 287412 |
+
},
|
| 287413 |
+
{
|
| 287414 |
+
"epoch": 1181.7,
|
| 287415 |
+
"learning_rate": 7.629289176090469e-06,
|
| 287416 |
+
"loss": 0.2804,
|
| 287417 |
+
"step": 147805
|
| 287418 |
+
},
|
| 287419 |
+
{
|
| 287420 |
+
"epoch": 1181.74,
|
| 287421 |
+
"learning_rate": 7.6292084006462045e-06,
|
| 287422 |
+
"loss": 0.3582,
|
| 287423 |
+
"step": 147810
|
| 287424 |
+
},
|
| 287425 |
+
{
|
| 287426 |
+
"epoch": 1181.78,
|
| 287427 |
+
"learning_rate": 7.6291276252019395e-06,
|
| 287428 |
+
"loss": 0.6097,
|
| 287429 |
+
"step": 147815
|
| 287430 |
+
},
|
| 287431 |
+
{
|
| 287432 |
+
"epoch": 1181.82,
|
| 287433 |
+
"learning_rate": 7.6290468497576745e-06,
|
| 287434 |
+
"loss": 0.9736,
|
| 287435 |
+
"step": 147820
|
| 287436 |
+
},
|
| 287437 |
+
{
|
| 287438 |
+
"epoch": 1181.86,
|
| 287439 |
+
"learning_rate": 7.6289660743134095e-06,
|
| 287440 |
+
"loss": 0.244,
|
| 287441 |
+
"step": 147825
|
| 287442 |
+
},
|
| 287443 |
+
{
|
| 287444 |
+
"epoch": 1181.9,
|
| 287445 |
+
"learning_rate": 7.6288852988691445e-06,
|
| 287446 |
+
"loss": 0.3411,
|
| 287447 |
+
"step": 147830
|
| 287448 |
+
},
|
| 287449 |
+
{
|
| 287450 |
+
"epoch": 1181.94,
|
| 287451 |
+
"learning_rate": 7.6288045234248795e-06,
|
| 287452 |
+
"loss": 0.3098,
|
| 287453 |
+
"step": 147835
|
| 287454 |
+
},
|
| 287455 |
+
{
|
| 287456 |
+
"epoch": 1181.98,
|
| 287457 |
+
"learning_rate": 7.6287237479806145e-06,
|
| 287458 |
+
"loss": 0.6602,
|
| 287459 |
+
"step": 147840
|
| 287460 |
+
},
|
| 287461 |
+
{
|
| 287462 |
+
"epoch": 1182.0,
|
| 287463 |
+
"eval_loss": 0.40087971091270447,
|
| 287464 |
+
"eval_runtime": 41.4361,
|
| 287465 |
+
"eval_samples_per_second": 20.272,
|
| 287466 |
+
"eval_steps_per_second": 0.652,
|
| 287467 |
+
"eval_wer": 0.17296944182332874,
|
| 287468 |
+
"step": 147843
|
| 287469 |
+
},
|
| 287470 |
+
{
|
| 287471 |
+
"epoch": 1192.02,
|
| 287472 |
+
"learning_rate": 7.6286429725363494e-06,
|
| 287473 |
+
"loss": 0.3926,
|
| 287474 |
+
"step": 147845
|
| 287475 |
+
},
|
| 287476 |
+
{
|
| 287477 |
+
"epoch": 1192.06,
|
| 287478 |
+
"learning_rate": 7.6285621970920844e-06,
|
| 287479 |
+
"loss": 0.2893,
|
| 287480 |
+
"step": 147850
|
| 287481 |
+
},
|
| 287482 |
+
{
|
| 287483 |
+
"epoch": 1192.1,
|
| 287484 |
+
"learning_rate": 7.628481421647819e-06,
|
| 287485 |
+
"loss": 0.2711,
|
| 287486 |
+
"step": 147855
|
| 287487 |
+
},
|
| 287488 |
+
{
|
| 287489 |
+
"epoch": 1192.14,
|
| 287490 |
+
"learning_rate": 7.628400646203554e-06,
|
| 287491 |
+
"loss": 0.3068,
|
| 287492 |
+
"step": 147860
|
| 287493 |
+
},
|
| 287494 |
+
{
|
| 287495 |
+
"epoch": 1192.18,
|
| 287496 |
+
"learning_rate": 7.628319870759289e-06,
|
| 287497 |
+
"loss": 0.6113,
|
| 287498 |
+
"step": 147865
|
| 287499 |
+
},
|
| 287500 |
+
{
|
| 287501 |
+
"epoch": 1192.22,
|
| 287502 |
+
"learning_rate": 7.628239095315025e-06,
|
| 287503 |
+
"loss": 0.9501,
|
| 287504 |
+
"step": 147870
|
| 287505 |
+
},
|
| 287506 |
+
{
|
| 287507 |
+
"epoch": 1192.26,
|
| 287508 |
+
"learning_rate": 7.62815831987076e-06,
|
| 287509 |
+
"loss": 0.2976,
|
| 287510 |
+
"step": 147875
|
| 287511 |
+
},
|
| 287512 |
+
{
|
| 287513 |
+
"epoch": 1192.3,
|
| 287514 |
+
"learning_rate": 7.628077544426495e-06,
|
| 287515 |
+
"loss": 0.291,
|
| 287516 |
+
"step": 147880
|
| 287517 |
+
},
|
| 287518 |
+
{
|
| 287519 |
+
"epoch": 1192.34,
|
| 287520 |
+
"learning_rate": 7.62799676898223e-06,
|
| 287521 |
+
"loss": 0.3597,
|
| 287522 |
+
"step": 147885
|
| 287523 |
+
},
|
| 287524 |
+
{
|
| 287525 |
+
"epoch": 1192.38,
|
| 287526 |
+
"learning_rate": 7.627915993537965e-06,
|
| 287527 |
+
"loss": 0.7271,
|
| 287528 |
+
"step": 147890
|
| 287529 |
+
},
|
| 287530 |
+
{
|
| 287531 |
+
"epoch": 1192.42,
|
| 287532 |
+
"learning_rate": 7.6278352180937e-06,
|
| 287533 |
+
"loss": 0.922,
|
| 287534 |
+
"step": 147895
|
| 287535 |
+
},
|
| 287536 |
+
{
|
| 287537 |
+
"epoch": 1192.46,
|
| 287538 |
+
"learning_rate": 7.627754442649435e-06,
|
| 287539 |
+
"loss": 0.3113,
|
| 287540 |
+
"step": 147900
|
| 287541 |
+
},
|
| 287542 |
+
{
|
| 287543 |
+
"epoch": 1192.5,
|
| 287544 |
+
"learning_rate": 7.62767366720517e-06,
|
| 287545 |
+
"loss": 0.2955,
|
| 287546 |
+
"step": 147905
|
| 287547 |
+
},
|
| 287548 |
+
{
|
| 287549 |
+
"epoch": 1192.54,
|
| 287550 |
+
"learning_rate": 7.627592891760905e-06,
|
| 287551 |
+
"loss": 0.302,
|
| 287552 |
+
"step": 147910
|
| 287553 |
+
},
|
| 287554 |
+
{
|
| 287555 |
+
"epoch": 1192.58,
|
| 287556 |
+
"learning_rate": 7.62751211631664e-06,
|
| 287557 |
+
"loss": 0.6507,
|
| 287558 |
+
"step": 147915
|
| 287559 |
+
},
|
| 287560 |
+
{
|
| 287561 |
+
"epoch": 1192.62,
|
| 287562 |
+
"learning_rate": 7.627431340872375e-06,
|
| 287563 |
+
"loss": 1.0984,
|
| 287564 |
+
"step": 147920
|
| 287565 |
+
},
|
| 287566 |
+
{
|
| 287567 |
+
"epoch": 1192.66,
|
| 287568 |
+
"learning_rate": 7.627350565428111e-06,
|
| 287569 |
+
"loss": 0.2543,
|
| 287570 |
+
"step": 147925
|
| 287571 |
+
},
|
| 287572 |
+
{
|
| 287573 |
+
"epoch": 1192.7,
|
| 287574 |
+
"learning_rate": 7.627269789983845e-06,
|
| 287575 |
+
"loss": 0.2675,
|
| 287576 |
+
"step": 147930
|
| 287577 |
+
},
|
| 287578 |
+
{
|
| 287579 |
+
"epoch": 1192.74,
|
| 287580 |
+
"learning_rate": 7.627189014539581e-06,
|
| 287581 |
+
"loss": 0.3055,
|
| 287582 |
+
"step": 147935
|
| 287583 |
+
},
|
| 287584 |
+
{
|
| 287585 |
+
"epoch": 1192.78,
|
| 287586 |
+
"learning_rate": 7.627108239095316e-06,
|
| 287587 |
+
"loss": 0.5833,
|
| 287588 |
+
"step": 147940
|
| 287589 |
+
},
|
| 287590 |
+
{
|
| 287591 |
+
"epoch": 1192.82,
|
| 287592 |
+
"learning_rate": 7.627027463651051e-06,
|
| 287593 |
+
"loss": 0.9175,
|
| 287594 |
+
"step": 147945
|
| 287595 |
+
},
|
| 287596 |
+
{
|
| 287597 |
+
"epoch": 1192.86,
|
| 287598 |
+
"learning_rate": 7.626946688206786e-06,
|
| 287599 |
+
"loss": 0.2454,
|
| 287600 |
+
"step": 147950
|
| 287601 |
+
},
|
| 287602 |
+
{
|
| 287603 |
+
"epoch": 1192.9,
|
| 287604 |
+
"learning_rate": 7.626865912762521e-06,
|
| 287605 |
+
"loss": 0.2769,
|
| 287606 |
+
"step": 147955
|
| 287607 |
+
},
|
| 287608 |
+
{
|
| 287609 |
+
"epoch": 1192.94,
|
| 287610 |
+
"learning_rate": 7.626785137318256e-06,
|
| 287611 |
+
"loss": 0.3101,
|
| 287612 |
+
"step": 147960
|
| 287613 |
+
},
|
| 287614 |
+
{
|
| 287615 |
+
"epoch": 1192.98,
|
| 287616 |
+
"learning_rate": 7.626704361873991e-06,
|
| 287617 |
+
"loss": 0.6418,
|
| 287618 |
+
"step": 147965
|
| 287619 |
+
},
|
| 287620 |
+
{
|
| 287621 |
+
"epoch": 1193.0,
|
| 287622 |
+
"eval_loss": 0.3407837450504303,
|
| 287623 |
+
"eval_runtime": 42.2837,
|
| 287624 |
+
"eval_samples_per_second": 19.889,
|
| 287625 |
+
"eval_steps_per_second": 0.639,
|
| 287626 |
+
"eval_wer": 0.1725676453655728,
|
| 287627 |
+
"step": 147967
|
| 287628 |
+
},
|
| 287629 |
+
{
|
| 287630 |
+
"epoch": 1193.02,
|
| 287631 |
+
"learning_rate": 7.626623586429726e-06,
|
| 287632 |
+
"loss": 0.3322,
|
| 287633 |
+
"step": 147970
|
| 287634 |
+
},
|
| 287635 |
+
{
|
| 287636 |
+
"epoch": 1193.06,
|
| 287637 |
+
"learning_rate": 7.626542810985461e-06,
|
| 287638 |
+
"loss": 0.3129,
|
| 287639 |
+
"step": 147975
|
| 287640 |
+
},
|
| 287641 |
+
{
|
| 287642 |
+
"epoch": 1193.1,
|
| 287643 |
+
"learning_rate": 7.626462035541196e-06,
|
| 287644 |
+
"loss": 0.2667,
|
| 287645 |
+
"step": 147980
|
| 287646 |
+
},
|
| 287647 |
+
{
|
| 287648 |
+
"epoch": 1193.14,
|
| 287649 |
+
"learning_rate": 7.626381260096931e-06,
|
| 287650 |
+
"loss": 0.3404,
|
| 287651 |
+
"step": 147985
|
| 287652 |
+
},
|
| 287653 |
+
{
|
| 287654 |
+
"epoch": 1193.18,
|
| 287655 |
+
"learning_rate": 7.626300484652667e-06,
|
| 287656 |
+
"loss": 0.8087,
|
| 287657 |
+
"step": 147990
|
| 287658 |
+
},
|
| 287659 |
+
{
|
| 287660 |
+
"epoch": 1193.22,
|
| 287661 |
+
"learning_rate": 7.626219709208401e-06,
|
| 287662 |
+
"loss": 0.8705,
|
| 287663 |
+
"step": 147995
|
| 287664 |
+
},
|
| 287665 |
+
{
|
| 287666 |
+
"epoch": 1193.27,
|
| 287667 |
+
"learning_rate": 7.626138933764137e-06,
|
| 287668 |
+
"loss": 0.2705,
|
| 287669 |
+
"step": 148000
|
| 287670 |
+
},
|
| 287671 |
+
{
|
| 287672 |
+
"epoch": 1193.31,
|
| 287673 |
+
"learning_rate": 7.626058158319872e-06,
|
| 287674 |
+
"loss": 0.3189,
|
| 287675 |
+
"step": 148005
|
| 287676 |
+
},
|
| 287677 |
+
{
|
| 287678 |
+
"epoch": 1193.35,
|
| 287679 |
+
"learning_rate": 7.6259773828756066e-06,
|
| 287680 |
+
"loss": 0.3232,
|
| 287681 |
+
"step": 148010
|
| 287682 |
+
},
|
| 287683 |
+
{
|
| 287684 |
+
"epoch": 1193.39,
|
| 287685 |
+
"learning_rate": 7.6258966074313416e-06,
|
| 287686 |
+
"loss": 0.7191,
|
| 287687 |
+
"step": 148015
|
| 287688 |
+
},
|
| 287689 |
+
{
|
| 287690 |
+
"epoch": 1193.43,
|
| 287691 |
+
"learning_rate": 7.6258158319870766e-06,
|
| 287692 |
+
"loss": 0.8468,
|
| 287693 |
+
"step": 148020
|
| 287694 |
+
},
|
| 287695 |
+
{
|
| 287696 |
+
"epoch": 1193.47,
|
| 287697 |
+
"learning_rate": 7.6257350565428115e-06,
|
| 287698 |
+
"loss": 0.2954,
|
| 287699 |
+
"step": 148025
|
| 287700 |
+
},
|
| 287701 |
+
{
|
| 287702 |
+
"epoch": 1193.51,
|
| 287703 |
+
"learning_rate": 7.6256542810985465e-06,
|
| 287704 |
+
"loss": 0.2542,
|
| 287705 |
+
"step": 148030
|
| 287706 |
+
},
|
| 287707 |
+
{
|
| 287708 |
+
"epoch": 1193.55,
|
| 287709 |
+
"learning_rate": 7.6255735056542815e-06,
|
| 287710 |
+
"loss": 0.3289,
|
| 287711 |
+
"step": 148035
|
| 287712 |
+
},
|
| 287713 |
+
{
|
| 287714 |
+
"epoch": 1193.59,
|
| 287715 |
+
"learning_rate": 7.6254927302100165e-06,
|
| 287716 |
+
"loss": 0.6459,
|
| 287717 |
+
"step": 148040
|
| 287718 |
+
},
|
| 287719 |
+
{
|
| 287720 |
+
"epoch": 1193.63,
|
| 287721 |
+
"learning_rate": 7.625411954765752e-06,
|
| 287722 |
+
"loss": 0.8973,
|
| 287723 |
+
"step": 148045
|
| 287724 |
+
},
|
| 287725 |
+
{
|
| 287726 |
+
"epoch": 1193.67,
|
| 287727 |
+
"learning_rate": 7.6253311793214865e-06,
|
| 287728 |
+
"loss": 0.3124,
|
| 287729 |
+
"step": 148050
|
| 287730 |
+
},
|
| 287731 |
+
{
|
| 287732 |
+
"epoch": 1193.71,
|
| 287733 |
+
"learning_rate": 7.625250403877222e-06,
|
| 287734 |
+
"loss": 0.274,
|
| 287735 |
+
"step": 148055
|
| 287736 |
+
},
|
| 287737 |
+
{
|
| 287738 |
+
"epoch": 1193.75,
|
| 287739 |
+
"learning_rate": 7.6251696284329565e-06,
|
| 287740 |
+
"loss": 0.3975,
|
| 287741 |
+
"step": 148060
|
| 287742 |
+
},
|
| 287743 |
+
{
|
| 287744 |
+
"epoch": 1193.79,
|
| 287745 |
+
"learning_rate": 7.625088852988692e-06,
|
| 287746 |
+
"loss": 0.7222,
|
| 287747 |
+
"step": 148065
|
| 287748 |
+
},
|
| 287749 |
+
{
|
| 287750 |
+
"epoch": 1193.83,
|
| 287751 |
+
"learning_rate": 7.6250080775444264e-06,
|
| 287752 |
+
"loss": 0.8491,
|
| 287753 |
+
"step": 148070
|
| 287754 |
+
},
|
| 287755 |
+
{
|
| 287756 |
+
"epoch": 1193.87,
|
| 287757 |
+
"learning_rate": 7.624927302100162e-06,
|
| 287758 |
+
"loss": 0.2277,
|
| 287759 |
+
"step": 148075
|
| 287760 |
+
},
|
| 287761 |
+
{
|
| 287762 |
+
"epoch": 1193.91,
|
| 287763 |
+
"learning_rate": 7.624846526655897e-06,
|
| 287764 |
+
"loss": 0.2811,
|
| 287765 |
+
"step": 148080
|
| 287766 |
+
},
|
| 287767 |
+
{
|
| 287768 |
+
"epoch": 1193.95,
|
| 287769 |
+
"learning_rate": 7.624765751211632e-06,
|
| 287770 |
+
"loss": 0.2995,
|
| 287771 |
+
"step": 148085
|
| 287772 |
+
},
|
| 287773 |
+
{
|
| 287774 |
+
"epoch": 1193.99,
|
| 287775 |
+
"learning_rate": 7.624684975767367e-06,
|
| 287776 |
+
"loss": 0.8542,
|
| 287777 |
+
"step": 148090
|
| 287778 |
+
},
|
| 287779 |
+
{
|
| 287780 |
+
"epoch": 1194.0,
|
| 287781 |
+
"eval_loss": 0.4103487432003021,
|
| 287782 |
+
"eval_runtime": 41.5587,
|
| 287783 |
+
"eval_samples_per_second": 20.236,
|
| 287784 |
+
"eval_steps_per_second": 0.65,
|
| 287785 |
+
"eval_wer": 0.1781133181851499,
|
| 287786 |
+
"step": 148091
|
| 287787 |
+
},
|
| 287788 |
+
{
|
| 287789 |
+
"epoch": 1184.03,
|
| 287790 |
+
"learning_rate": 7.624604200323102e-06,
|
| 287791 |
+
"loss": 0.3722,
|
| 287792 |
+
"step": 148095
|
| 287793 |
+
},
|
| 287794 |
+
{
|
| 287795 |
+
"epoch": 1184.07,
|
| 287796 |
+
"learning_rate": 7.624523424878838e-06,
|
| 287797 |
+
"loss": 0.2441,
|
| 287798 |
+
"step": 148100
|
| 287799 |
+
},
|
| 287800 |
+
{
|
| 287801 |
+
"epoch": 1184.11,
|
| 287802 |
+
"learning_rate": 7.624442649434572e-06,
|
| 287803 |
+
"loss": 0.3059,
|
| 287804 |
+
"step": 148105
|
| 287805 |
+
},
|
| 287806 |
+
{
|
| 287807 |
+
"epoch": 1184.15,
|
| 287808 |
+
"learning_rate": 7.624361873990308e-06,
|
| 287809 |
+
"loss": 0.4402,
|
| 287810 |
+
"step": 148110
|
| 287811 |
+
},
|
| 287812 |
+
{
|
| 287813 |
+
"epoch": 1184.19,
|
| 287814 |
+
"learning_rate": 7.624281098546042e-06,
|
| 287815 |
+
"loss": 0.8067,
|
| 287816 |
+
"step": 148115
|
| 287817 |
+
},
|
| 287818 |
+
{
|
| 287819 |
+
"epoch": 1184.23,
|
| 287820 |
+
"learning_rate": 7.624200323101778e-06,
|
| 287821 |
+
"loss": 0.704,
|
| 287822 |
+
"step": 148120
|
| 287823 |
+
},
|
| 287824 |
+
{
|
| 287825 |
+
"epoch": 1184.27,
|
| 287826 |
+
"learning_rate": 7.624119547657512e-06,
|
| 287827 |
+
"loss": 0.2738,
|
| 287828 |
+
"step": 148125
|
| 287829 |
+
},
|
| 287830 |
+
{
|
| 287831 |
+
"epoch": 1184.31,
|
| 287832 |
+
"learning_rate": 7.6240549273021e-06,
|
| 287833 |
+
"loss": 0.3263,
|
| 287834 |
+
"step": 148130
|
| 287835 |
+
},
|
| 287836 |
+
{
|
| 287837 |
+
"epoch": 1184.35,
|
| 287838 |
+
"learning_rate": 7.623974151857836e-06,
|
| 287839 |
+
"loss": 0.4694,
|
| 287840 |
+
"step": 148135
|
| 287841 |
+
},
|
| 287842 |
+
{
|
| 287843 |
+
"epoch": 1184.39,
|
| 287844 |
+
"learning_rate": 7.623893376413571e-06,
|
| 287845 |
+
"loss": 0.8373,
|
| 287846 |
+
"step": 148140
|
| 287847 |
+
},
|
| 287848 |
+
{
|
| 287849 |
+
"epoch": 1184.43,
|
| 287850 |
+
"learning_rate": 7.623812600969306e-06,
|
| 287851 |
+
"loss": 0.608,
|
| 287852 |
+
"step": 148145
|
| 287853 |
+
},
|
| 287854 |
+
{
|
| 287855 |
+
"epoch": 1184.47,
|
| 287856 |
+
"learning_rate": 7.623731825525041e-06,
|
| 287857 |
+
"loss": 0.2548,
|
| 287858 |
+
"step": 148150
|
| 287859 |
+
},
|
| 287860 |
+
{
|
| 287861 |
+
"epoch": 1184.51,
|
| 287862 |
+
"learning_rate": 7.623651050080776e-06,
|
| 287863 |
+
"loss": 0.2965,
|
| 287864 |
+
"step": 148155
|
| 287865 |
+
},
|
| 287866 |
+
{
|
| 287867 |
+
"epoch": 1184.55,
|
| 287868 |
+
"learning_rate": 7.623570274636511e-06,
|
| 287869 |
+
"loss": 0.3752,
|
| 287870 |
+
"step": 148160
|
| 287871 |
+
},
|
| 287872 |
+
{
|
| 287873 |
+
"epoch": 1184.59,
|
| 287874 |
+
"learning_rate": 7.623489499192246e-06,
|
| 287875 |
+
"loss": 0.8264,
|
| 287876 |
+
"step": 148165
|
| 287877 |
+
},
|
| 287878 |
+
{
|
| 287879 |
+
"epoch": 1184.63,
|
| 287880 |
+
"learning_rate": 7.623408723747981e-06,
|
| 287881 |
+
"loss": 0.6361,
|
| 287882 |
+
"step": 148170
|
| 287883 |
+
},
|
| 287884 |
+
{
|
| 287885 |
+
"epoch": 1184.67,
|
| 287886 |
+
"learning_rate": 7.623327948303716e-06,
|
| 287887 |
+
"loss": 0.2599,
|
| 287888 |
+
"step": 148175
|
| 287889 |
+
},
|
| 287890 |
+
{
|
| 287891 |
+
"epoch": 1184.71,
|
| 287892 |
+
"learning_rate": 7.623247172859452e-06,
|
| 287893 |
+
"loss": 0.3282,
|
| 287894 |
+
"step": 148180
|
| 287895 |
+
},
|
| 287896 |
+
{
|
| 287897 |
+
"epoch": 1184.75,
|
| 287898 |
+
"learning_rate": 7.623166397415186e-06,
|
| 287899 |
+
"loss": 0.3428,
|
| 287900 |
+
"step": 148185
|
| 287901 |
+
},
|
| 287902 |
+
{
|
| 287903 |
+
"epoch": 1184.79,
|
| 287904 |
+
"learning_rate": 7.623085621970922e-06,
|
| 287905 |
+
"loss": 0.7791,
|
| 287906 |
+
"step": 148190
|
| 287907 |
+
},
|
| 287908 |
+
{
|
| 287909 |
+
"epoch": 1184.83,
|
| 287910 |
+
"learning_rate": 7.623004846526657e-06,
|
| 287911 |
+
"loss": 0.7382,
|
| 287912 |
+
"step": 148195
|
| 287913 |
+
},
|
| 287914 |
+
{
|
| 287915 |
+
"epoch": 1184.87,
|
| 287916 |
+
"learning_rate": 7.622924071082392e-06,
|
| 287917 |
+
"loss": 0.232,
|
| 287918 |
+
"step": 148200
|
| 287919 |
+
},
|
| 287920 |
+
{
|
| 287921 |
+
"epoch": 1184.91,
|
| 287922 |
+
"learning_rate": 7.622843295638127e-06,
|
| 287923 |
+
"loss": 0.2919,
|
| 287924 |
+
"step": 148205
|
| 287925 |
+
},
|
| 287926 |
+
{
|
| 287927 |
+
"epoch": 1184.95,
|
| 287928 |
+
"learning_rate": 7.622762520193862e-06,
|
| 287929 |
+
"loss": 0.5011,
|
| 287930 |
+
"step": 148210
|
| 287931 |
+
},
|
| 287932 |
+
{
|
| 287933 |
+
"epoch": 1184.99,
|
| 287934 |
+
"learning_rate": 7.622681744749597e-06,
|
| 287935 |
+
"loss": 1.0441,
|
| 287936 |
+
"step": 148215
|
| 287937 |
+
},
|
| 287938 |
+
{
|
| 287939 |
+
"epoch": 1185.0,
|
| 287940 |
+
"eval_loss": 0.3193012773990631,
|
| 287941 |
+
"eval_runtime": 41.5192,
|
| 287942 |
+
"eval_samples_per_second": 20.256,
|
| 287943 |
+
"eval_steps_per_second": 0.65,
|
| 287944 |
+
"eval_wer": 0.17354712326779365,
|
| 287945 |
+
"step": 148216
|
| 287946 |
+
},
|
| 287947 |
+
{
|
| 287948 |
+
"epoch": 1195.03,
|
| 287949 |
+
"learning_rate": 7.622600969305332e-06,
|
| 287950 |
+
"loss": 0.3208,
|
| 287951 |
+
"step": 148220
|
| 287952 |
+
},
|
| 287953 |
+
{
|
| 287954 |
+
"epoch": 1195.07,
|
| 287955 |
+
"learning_rate": 7.622520193861067e-06,
|
| 287956 |
+
"loss": 0.2524,
|
| 287957 |
+
"step": 148225
|
| 287958 |
+
},
|
| 287959 |
+
{
|
| 287960 |
+
"epoch": 1195.11,
|
| 287961 |
+
"learning_rate": 7.622439418416802e-06,
|
| 287962 |
+
"loss": 0.2575,
|
| 287963 |
+
"step": 148230
|
| 287964 |
+
},
|
| 287965 |
+
{
|
| 287966 |
+
"epoch": 1195.15,
|
| 287967 |
+
"learning_rate": 7.622358642972537e-06,
|
| 287968 |
+
"loss": 0.4411,
|
| 287969 |
+
"step": 148235
|
| 287970 |
+
},
|
| 287971 |
+
{
|
| 287972 |
+
"epoch": 1195.19,
|
| 287973 |
+
"learning_rate": 7.622277867528272e-06,
|
| 287974 |
+
"loss": 0.8635,
|
| 287975 |
+
"step": 148240
|
| 287976 |
+
},
|
| 287977 |
+
{
|
| 287978 |
+
"epoch": 1195.23,
|
| 287979 |
+
"learning_rate": 7.6221970920840075e-06,
|
| 287980 |
+
"loss": 0.6527,
|
| 287981 |
+
"step": 148245
|
| 287982 |
+
},
|
| 287983 |
+
{
|
| 287984 |
+
"epoch": 1195.27,
|
| 287985 |
+
"learning_rate": 7.622116316639742e-06,
|
| 287986 |
+
"loss": 0.2968,
|
| 287987 |
+
"step": 148250
|
| 287988 |
+
},
|
| 287989 |
+
{
|
| 287990 |
+
"epoch": 1195.31,
|
| 287991 |
+
"learning_rate": 7.6220355411954775e-06,
|
| 287992 |
+
"loss": 0.2476,
|
| 287993 |
+
"step": 148255
|
| 287994 |
+
},
|
| 287995 |
+
{
|
| 287996 |
+
"epoch": 1195.35,
|
| 287997 |
+
"learning_rate": 7.6219547657512125e-06,
|
| 287998 |
+
"loss": 0.3416,
|
| 287999 |
+
"step": 148260
|
| 288000 |
+
},
|
| 288001 |
+
{
|
| 288002 |
+
"epoch": 1195.39,
|
| 288003 |
+
"learning_rate": 7.6218739903069475e-06,
|
| 288004 |
+
"loss": 0.9705,
|
| 288005 |
+
"step": 148265
|
| 288006 |
+
},
|
| 288007 |
+
{
|
| 288008 |
+
"epoch": 1195.43,
|
| 288009 |
+
"learning_rate": 7.6217932148626825e-06,
|
| 288010 |
+
"loss": 0.6557,
|
| 288011 |
+
"step": 148270
|
| 288012 |
+
},
|
| 288013 |
+
{
|
| 288014 |
+
"epoch": 1195.47,
|
| 288015 |
+
"learning_rate": 7.6217124394184175e-06,
|
| 288016 |
+
"loss": 0.2998,
|
| 288017 |
+
"step": 148275
|
| 288018 |
+
},
|
| 288019 |
+
{
|
| 288020 |
+
"epoch": 1195.51,
|
| 288021 |
+
"learning_rate": 7.6216316639741525e-06,
|
| 288022 |
+
"loss": 0.2876,
|
| 288023 |
+
"step": 148280
|
| 288024 |
+
},
|
| 288025 |
+
{
|
| 288026 |
+
"epoch": 1195.55,
|
| 288027 |
+
"learning_rate": 7.6215508885298874e-06,
|
| 288028 |
+
"loss": 0.3667,
|
| 288029 |
+
"step": 148285
|
| 288030 |
+
},
|
| 288031 |
+
{
|
| 288032 |
+
"epoch": 1195.59,
|
| 288033 |
+
"learning_rate": 7.6214701130856224e-06,
|
| 288034 |
+
"loss": 0.8482,
|
| 288035 |
+
"step": 148290
|
| 288036 |
+
},
|
| 288037 |
+
{
|
| 288038 |
+
"epoch": 1195.63,
|
| 288039 |
+
"learning_rate": 7.6213893376413574e-06,
|
| 288040 |
+
"loss": 0.6535,
|
| 288041 |
+
"step": 148295
|
| 288042 |
+
},
|
| 288043 |
+
{
|
| 288044 |
+
"epoch": 1195.67,
|
| 288045 |
+
"learning_rate": 7.621308562197092e-06,
|
| 288046 |
+
"loss": 0.2823,
|
| 288047 |
+
"step": 148300
|
| 288048 |
+
},
|
| 288049 |
+
{
|
| 288050 |
+
"epoch": 1195.71,
|
| 288051 |
+
"learning_rate": 7.621227786752827e-06,
|
| 288052 |
+
"loss": 0.2901,
|
| 288053 |
+
"step": 148305
|
| 288054 |
+
},
|
| 288055 |
+
{
|
| 288056 |
+
"epoch": 1195.76,
|
| 288057 |
+
"learning_rate": 7.621147011308563e-06,
|
| 288058 |
+
"loss": 0.3151,
|
| 288059 |
+
"step": 148310
|
| 288060 |
+
},
|
| 288061 |
+
{
|
| 288062 |
+
"epoch": 1195.8,
|
| 288063 |
+
"learning_rate": 7.621066235864298e-06,
|
| 288064 |
+
"loss": 0.939,
|
| 288065 |
+
"step": 148315
|
| 288066 |
+
},
|
| 288067 |
+
{
|
| 288068 |
+
"epoch": 1195.84,
|
| 288069 |
+
"learning_rate": 7.620985460420033e-06,
|
| 288070 |
+
"loss": 0.6048,
|
| 288071 |
+
"step": 148320
|
| 288072 |
+
},
|
| 288073 |
+
{
|
| 288074 |
+
"epoch": 1195.88,
|
| 288075 |
+
"learning_rate": 7.620904684975768e-06,
|
| 288076 |
+
"loss": 0.3069,
|
| 288077 |
+
"step": 148325
|
| 288078 |
+
},
|
| 288079 |
+
{
|
| 288080 |
+
"epoch": 1195.92,
|
| 288081 |
+
"learning_rate": 7.620823909531503e-06,
|
| 288082 |
+
"loss": 0.2925,
|
| 288083 |
+
"step": 148330
|
| 288084 |
+
},
|
| 288085 |
+
{
|
| 288086 |
+
"epoch": 1195.96,
|
| 288087 |
+
"learning_rate": 7.620743134087238e-06,
|
| 288088 |
+
"loss": 0.3316,
|
| 288089 |
+
"step": 148335
|
| 288090 |
+
},
|
| 288091 |
+
{
|
| 288092 |
+
"epoch": 1196.0,
|
| 288093 |
+
"learning_rate": 7.620662358642973e-06,
|
| 288094 |
+
"loss": 1.075,
|
| 288095 |
+
"step": 148340
|
| 288096 |
+
},
|
| 288097 |
+
{
|
| 288098 |
+
"epoch": 1196.0,
|
| 288099 |
+
"eval_loss": 0.3667512834072113,
|
| 288100 |
+
"eval_runtime": 42.6545,
|
| 288101 |
+
"eval_samples_per_second": 19.717,
|
| 288102 |
+
"eval_steps_per_second": 0.633,
|
| 288103 |
+
"eval_wer": 0.1786735277301315,
|
| 288104 |
+
"step": 148340
|
| 288105 |
}
|
| 288106 |
],
|
| 288107 |
+
"max_steps": 620000,
|
| 288108 |
"num_train_epochs": 5000,
|
| 288109 |
+
"total_flos": 4.1744157123227523e+20,
|
| 288110 |
"trial_name": null,
|
| 288111 |
"trial_params": null
|
| 288112 |
}
|
model-bin/finetune/base/{checkpoint-147718 β checkpoint-148340}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1630235074.4070776/events.out.tfevents.1630235074.cc93b136ebf5.1086.139
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:36a04c31066562f8b25d6d574c8382406aa1d11f806fec99dcdb9cea961b5e7b
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630235509.4705393/events.out.tfevents.1630235509.cc93b136ebf5.1086.141
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:229f9fc73cabb1820ed87fca0b4cffe4d1c163b0a6b9f343fd6bac07b45959ea
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630235947.303649/events.out.tfevents.1630235947.cc93b136ebf5.1086.143
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0d128741b908e3bf073f6f9cdf857941b985a1c77d660bee167bf0873d903812
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630236380.4833214/events.out.tfevents.1630236380.cc93b136ebf5.1086.145
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7bc4fcf4ae26326c2128f45a31d245e0817de203fda33b4c1cb7cd65567c3ea5
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630236821.4505744/events.out.tfevents.1630236821.cc93b136ebf5.1086.147
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3fbbf8ee946be6213439070a4ec05bc956b5bf5b43423a8cd2bf47daceb774d7
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630235074.cc93b136ebf5.1086.138
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7380b68391ab48dd6f4961e4bbfeafa0629583266670bec9db3ab25261f00492
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630235509.cc93b136ebf5.1086.140
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4d114e4dfd09ea820b1cf1bfdfc2408595f1a887c0b4f6e516ae0681f38c7b9c
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630235947.cc93b136ebf5.1086.142
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b5e7fc6301feed7f6cb72bdba0c7ee99050ef7a4889e308aef7eeeaaff2e536f
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630236380.cc93b136ebf5.1086.144
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f8281fe6e1f9fc1b9892bdadf91c54a5befc184aab952bd5aaafc1c9425c32bf
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630236821.cc93b136ebf5.1086.146
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e59e1dedf6fc9959d902922a3b13dbbaf5d9ee7aa271490ed78f48ebbb260b85
|
| 3 |
+
size 8622
|