"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/trainer_state.json +792 -3
- model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630248561.4202452/events.out.tfevents.1630248561.cc93b136ebf5.1086.199 +3 -0
- model-bin/finetune/base/log/1630248992.3105023/events.out.tfevents.1630248992.cc93b136ebf5.1086.201 +3 -0
- model-bin/finetune/base/log/1630249423.7833743/events.out.tfevents.1630249423.cc93b136ebf5.1086.203 +3 -0
- model-bin/finetune/base/log/1630249857.8968918/events.out.tfevents.1630249857.cc93b136ebf5.1086.205 +3 -0
- model-bin/finetune/base/log/1630250287.4046617/events.out.tfevents.1630250287.cc93b136ebf5.1086.207 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630248561.cc93b136ebf5.1086.198 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630248992.cc93b136ebf5.1086.200 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630249423.cc93b136ebf5.1086.202 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630249857.cc93b136ebf5.1086.204 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630250287.cc93b136ebf5.1086.206 +3 -0
model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:88106383b07c091dad16598681c368eb1f7d5dd6754f75dcdede18a251a76193
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e90c5a61e1081391cac20c2749dbcf1cb5ad34446b8a2074676f8de81385b325
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9e5c8178a08780b7e94358c9559057df556e07dd698f9fa5dc769d640159e51e
|
| 3 |
size 14503
|
model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5fe336fce1f24fb6ba716a3ec77454f1b6f380bbf0069997841f401a2fb77c2a
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d1c8fd19316f91f2cc4123c75c86425f3af495dca75999776dc0193fa6dd50cb
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1668011442822563,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -292053,11 +292053,800 @@
|
|
| 292053 |
"eval_steps_per_second": 0.639,
|
| 292054 |
"eval_wer": 0.17625899280575538,
|
| 292055 |
"step": 151448
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 292056 |
}
|
| 292057 |
],
|
| 292058 |
"max_steps": 620000,
|
| 292059 |
"num_train_epochs": 5000,
|
| 292060 |
-
"total_flos": 4.
|
| 292061 |
"trial_name": null,
|
| 292062 |
"trial_params": null
|
| 292063 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1668011442822563,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
| 4 |
+
"epoch": 1225.995983935743,
|
| 5 |
+
"global_step": 152068,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 292053 |
"eval_steps_per_second": 0.639,
|
| 292054 |
"eval_wer": 0.17625899280575538,
|
| 292055 |
"step": 151448
|
| 292056 |
+
},
|
| 292057 |
+
{
|
| 292058 |
+
"epoch": 1221.02,
|
| 292059 |
+
"learning_rate": 7.570420032310179e-06,
|
| 292060 |
+
"loss": 0.3857,
|
| 292061 |
+
"step": 151450
|
| 292062 |
+
},
|
| 292063 |
+
{
|
| 292064 |
+
"epoch": 1221.06,
|
| 292065 |
+
"learning_rate": 7.570339256865913e-06,
|
| 292066 |
+
"loss": 0.27,
|
| 292067 |
+
"step": 151455
|
| 292068 |
+
},
|
| 292069 |
+
{
|
| 292070 |
+
"epoch": 1221.1,
|
| 292071 |
+
"learning_rate": 7.5702584814216486e-06,
|
| 292072 |
+
"loss": 0.2653,
|
| 292073 |
+
"step": 151460
|
| 292074 |
+
},
|
| 292075 |
+
{
|
| 292076 |
+
"epoch": 1221.14,
|
| 292077 |
+
"learning_rate": 7.570177705977384e-06,
|
| 292078 |
+
"loss": 0.34,
|
| 292079 |
+
"step": 151465
|
| 292080 |
+
},
|
| 292081 |
+
{
|
| 292082 |
+
"epoch": 1221.18,
|
| 292083 |
+
"learning_rate": 7.5700969305331186e-06,
|
| 292084 |
+
"loss": 0.5222,
|
| 292085 |
+
"step": 151470
|
| 292086 |
+
},
|
| 292087 |
+
{
|
| 292088 |
+
"epoch": 1221.22,
|
| 292089 |
+
"learning_rate": 7.570016155088854e-06,
|
| 292090 |
+
"loss": 0.8796,
|
| 292091 |
+
"step": 151475
|
| 292092 |
+
},
|
| 292093 |
+
{
|
| 292094 |
+
"epoch": 1221.26,
|
| 292095 |
+
"learning_rate": 7.5699353796445885e-06,
|
| 292096 |
+
"loss": 0.2732,
|
| 292097 |
+
"step": 151480
|
| 292098 |
+
},
|
| 292099 |
+
{
|
| 292100 |
+
"epoch": 1221.3,
|
| 292101 |
+
"learning_rate": 7.569854604200324e-06,
|
| 292102 |
+
"loss": 0.2637,
|
| 292103 |
+
"step": 151485
|
| 292104 |
+
},
|
| 292105 |
+
{
|
| 292106 |
+
"epoch": 1221.34,
|
| 292107 |
+
"learning_rate": 7.5697738287560585e-06,
|
| 292108 |
+
"loss": 0.3104,
|
| 292109 |
+
"step": 151490
|
| 292110 |
+
},
|
| 292111 |
+
{
|
| 292112 |
+
"epoch": 1221.38,
|
| 292113 |
+
"learning_rate": 7.569693053311794e-06,
|
| 292114 |
+
"loss": 0.5735,
|
| 292115 |
+
"step": 151495
|
| 292116 |
+
},
|
| 292117 |
+
{
|
| 292118 |
+
"epoch": 1221.42,
|
| 292119 |
+
"learning_rate": 7.5696122778675285e-06,
|
| 292120 |
+
"loss": 0.9314,
|
| 292121 |
+
"step": 151500
|
| 292122 |
+
},
|
| 292123 |
+
{
|
| 292124 |
+
"epoch": 1221.46,
|
| 292125 |
+
"learning_rate": 7.569531502423264e-06,
|
| 292126 |
+
"loss": 0.2415,
|
| 292127 |
+
"step": 151505
|
| 292128 |
+
},
|
| 292129 |
+
{
|
| 292130 |
+
"epoch": 1221.5,
|
| 292131 |
+
"learning_rate": 7.5694507269789985e-06,
|
| 292132 |
+
"loss": 0.2433,
|
| 292133 |
+
"step": 151510
|
| 292134 |
+
},
|
| 292135 |
+
{
|
| 292136 |
+
"epoch": 1221.54,
|
| 292137 |
+
"learning_rate": 7.569369951534734e-06,
|
| 292138 |
+
"loss": 0.3259,
|
| 292139 |
+
"step": 151515
|
| 292140 |
+
},
|
| 292141 |
+
{
|
| 292142 |
+
"epoch": 1221.58,
|
| 292143 |
+
"learning_rate": 7.569289176090469e-06,
|
| 292144 |
+
"loss": 0.5686,
|
| 292145 |
+
"step": 151520
|
| 292146 |
+
},
|
| 292147 |
+
{
|
| 292148 |
+
"epoch": 1221.62,
|
| 292149 |
+
"learning_rate": 7.569208400646204e-06,
|
| 292150 |
+
"loss": 0.9219,
|
| 292151 |
+
"step": 151525
|
| 292152 |
+
},
|
| 292153 |
+
{
|
| 292154 |
+
"epoch": 1221.66,
|
| 292155 |
+
"learning_rate": 7.569127625201939e-06,
|
| 292156 |
+
"loss": 0.2692,
|
| 292157 |
+
"step": 151530
|
| 292158 |
+
},
|
| 292159 |
+
{
|
| 292160 |
+
"epoch": 1221.7,
|
| 292161 |
+
"learning_rate": 7.569046849757674e-06,
|
| 292162 |
+
"loss": 0.29,
|
| 292163 |
+
"step": 151535
|
| 292164 |
+
},
|
| 292165 |
+
{
|
| 292166 |
+
"epoch": 1221.74,
|
| 292167 |
+
"learning_rate": 7.56896607431341e-06,
|
| 292168 |
+
"loss": 0.2842,
|
| 292169 |
+
"step": 151540
|
| 292170 |
+
},
|
| 292171 |
+
{
|
| 292172 |
+
"epoch": 1221.78,
|
| 292173 |
+
"learning_rate": 7.568885298869144e-06,
|
| 292174 |
+
"loss": 0.5224,
|
| 292175 |
+
"step": 151545
|
| 292176 |
+
},
|
| 292177 |
+
{
|
| 292178 |
+
"epoch": 1221.82,
|
| 292179 |
+
"learning_rate": 7.56880452342488e-06,
|
| 292180 |
+
"loss": 0.9927,
|
| 292181 |
+
"step": 151550
|
| 292182 |
+
},
|
| 292183 |
+
{
|
| 292184 |
+
"epoch": 1221.86,
|
| 292185 |
+
"learning_rate": 7.568723747980614e-06,
|
| 292186 |
+
"loss": 0.2828,
|
| 292187 |
+
"step": 151555
|
| 292188 |
+
},
|
| 292189 |
+
{
|
| 292190 |
+
"epoch": 1221.9,
|
| 292191 |
+
"learning_rate": 7.56864297253635e-06,
|
| 292192 |
+
"loss": 0.2777,
|
| 292193 |
+
"step": 151560
|
| 292194 |
+
},
|
| 292195 |
+
{
|
| 292196 |
+
"epoch": 1221.94,
|
| 292197 |
+
"learning_rate": 7.568562197092084e-06,
|
| 292198 |
+
"loss": 0.3088,
|
| 292199 |
+
"step": 151565
|
| 292200 |
+
},
|
| 292201 |
+
{
|
| 292202 |
+
"epoch": 1221.98,
|
| 292203 |
+
"learning_rate": 7.56848142164782e-06,
|
| 292204 |
+
"loss": 0.6365,
|
| 292205 |
+
"step": 151570
|
| 292206 |
+
},
|
| 292207 |
+
{
|
| 292208 |
+
"epoch": 1222.0,
|
| 292209 |
+
"eval_loss": 0.5525439977645874,
|
| 292210 |
+
"eval_runtime": 39.5867,
|
| 292211 |
+
"eval_samples_per_second": 21.118,
|
| 292212 |
+
"eval_steps_per_second": 0.682,
|
| 292213 |
+
"eval_wer": 0.1825968316531494,
|
| 292214 |
+
"step": 151572
|
| 292215 |
+
},
|
| 292216 |
+
{
|
| 292217 |
+
"epoch": 1222.02,
|
| 292218 |
+
"learning_rate": 7.568400646203554e-06,
|
| 292219 |
+
"loss": 0.2871,
|
| 292220 |
+
"step": 151575
|
| 292221 |
+
},
|
| 292222 |
+
{
|
| 292223 |
+
"epoch": 1222.06,
|
| 292224 |
+
"learning_rate": 7.56831987075929e-06,
|
| 292225 |
+
"loss": 0.2949,
|
| 292226 |
+
"step": 151580
|
| 292227 |
+
},
|
| 292228 |
+
{
|
| 292229 |
+
"epoch": 1222.1,
|
| 292230 |
+
"learning_rate": 7.568239095315025e-06,
|
| 292231 |
+
"loss": 0.2781,
|
| 292232 |
+
"step": 151585
|
| 292233 |
+
},
|
| 292234 |
+
{
|
| 292235 |
+
"epoch": 1222.14,
|
| 292236 |
+
"learning_rate": 7.56815831987076e-06,
|
| 292237 |
+
"loss": 0.4072,
|
| 292238 |
+
"step": 151590
|
| 292239 |
+
},
|
| 292240 |
+
{
|
| 292241 |
+
"epoch": 1222.18,
|
| 292242 |
+
"learning_rate": 7.568077544426495e-06,
|
| 292243 |
+
"loss": 0.69,
|
| 292244 |
+
"step": 151595
|
| 292245 |
+
},
|
| 292246 |
+
{
|
| 292247 |
+
"epoch": 1222.22,
|
| 292248 |
+
"learning_rate": 7.56799676898223e-06,
|
| 292249 |
+
"loss": 0.7913,
|
| 292250 |
+
"step": 151600
|
| 292251 |
+
},
|
| 292252 |
+
{
|
| 292253 |
+
"epoch": 1222.27,
|
| 292254 |
+
"learning_rate": 7.567915993537966e-06,
|
| 292255 |
+
"loss": 0.3416,
|
| 292256 |
+
"step": 151605
|
| 292257 |
+
},
|
| 292258 |
+
{
|
| 292259 |
+
"epoch": 1222.31,
|
| 292260 |
+
"learning_rate": 7.5678352180937e-06,
|
| 292261 |
+
"loss": 0.2677,
|
| 292262 |
+
"step": 151610
|
| 292263 |
+
},
|
| 292264 |
+
{
|
| 292265 |
+
"epoch": 1222.35,
|
| 292266 |
+
"learning_rate": 7.567754442649436e-06,
|
| 292267 |
+
"loss": 0.3702,
|
| 292268 |
+
"step": 151615
|
| 292269 |
+
},
|
| 292270 |
+
{
|
| 292271 |
+
"epoch": 1222.39,
|
| 292272 |
+
"learning_rate": 7.56767366720517e-06,
|
| 292273 |
+
"loss": 0.7914,
|
| 292274 |
+
"step": 151620
|
| 292275 |
+
},
|
| 292276 |
+
{
|
| 292277 |
+
"epoch": 1222.43,
|
| 292278 |
+
"learning_rate": 7.567592891760906e-06,
|
| 292279 |
+
"loss": 0.7734,
|
| 292280 |
+
"step": 151625
|
| 292281 |
+
},
|
| 292282 |
+
{
|
| 292283 |
+
"epoch": 1222.47,
|
| 292284 |
+
"learning_rate": 7.56751211631664e-06,
|
| 292285 |
+
"loss": 0.4037,
|
| 292286 |
+
"step": 151630
|
| 292287 |
+
},
|
| 292288 |
+
{
|
| 292289 |
+
"epoch": 1222.51,
|
| 292290 |
+
"learning_rate": 7.567431340872376e-06,
|
| 292291 |
+
"loss": 0.2796,
|
| 292292 |
+
"step": 151635
|
| 292293 |
+
},
|
| 292294 |
+
{
|
| 292295 |
+
"epoch": 1222.55,
|
| 292296 |
+
"learning_rate": 7.567350565428111e-06,
|
| 292297 |
+
"loss": 0.4157,
|
| 292298 |
+
"step": 151640
|
| 292299 |
+
},
|
| 292300 |
+
{
|
| 292301 |
+
"epoch": 1222.59,
|
| 292302 |
+
"learning_rate": 7.567269789983846e-06,
|
| 292303 |
+
"loss": 0.7692,
|
| 292304 |
+
"step": 151645
|
| 292305 |
+
},
|
| 292306 |
+
{
|
| 292307 |
+
"epoch": 1222.63,
|
| 292308 |
+
"learning_rate": 7.567189014539581e-06,
|
| 292309 |
+
"loss": 0.8732,
|
| 292310 |
+
"step": 151650
|
| 292311 |
+
},
|
| 292312 |
+
{
|
| 292313 |
+
"epoch": 1222.67,
|
| 292314 |
+
"learning_rate": 7.567108239095316e-06,
|
| 292315 |
+
"loss": 0.256,
|
| 292316 |
+
"step": 151655
|
| 292317 |
+
},
|
| 292318 |
+
{
|
| 292319 |
+
"epoch": 1222.71,
|
| 292320 |
+
"learning_rate": 7.567027463651051e-06,
|
| 292321 |
+
"loss": 0.2665,
|
| 292322 |
+
"step": 151660
|
| 292323 |
+
},
|
| 292324 |
+
{
|
| 292325 |
+
"epoch": 1222.75,
|
| 292326 |
+
"learning_rate": 7.566946688206786e-06,
|
| 292327 |
+
"loss": 0.4612,
|
| 292328 |
+
"step": 151665
|
| 292329 |
+
},
|
| 292330 |
+
{
|
| 292331 |
+
"epoch": 1222.79,
|
| 292332 |
+
"learning_rate": 7.5668659127625214e-06,
|
| 292333 |
+
"loss": 0.7535,
|
| 292334 |
+
"step": 151670
|
| 292335 |
+
},
|
| 292336 |
+
{
|
| 292337 |
+
"epoch": 1222.83,
|
| 292338 |
+
"learning_rate": 7.566785137318256e-06,
|
| 292339 |
+
"loss": 0.9049,
|
| 292340 |
+
"step": 151675
|
| 292341 |
+
},
|
| 292342 |
+
{
|
| 292343 |
+
"epoch": 1222.87,
|
| 292344 |
+
"learning_rate": 7.566704361873991e-06,
|
| 292345 |
+
"loss": 0.3755,
|
| 292346 |
+
"step": 151680
|
| 292347 |
+
},
|
| 292348 |
+
{
|
| 292349 |
+
"epoch": 1222.91,
|
| 292350 |
+
"learning_rate": 7.5666235864297256e-06,
|
| 292351 |
+
"loss": 0.2588,
|
| 292352 |
+
"step": 151685
|
| 292353 |
+
},
|
| 292354 |
+
{
|
| 292355 |
+
"epoch": 1222.95,
|
| 292356 |
+
"learning_rate": 7.566542810985461e-06,
|
| 292357 |
+
"loss": 0.365,
|
| 292358 |
+
"step": 151690
|
| 292359 |
+
},
|
| 292360 |
+
{
|
| 292361 |
+
"epoch": 1222.99,
|
| 292362 |
+
"learning_rate": 7.566462035541196e-06,
|
| 292363 |
+
"loss": 0.7382,
|
| 292364 |
+
"step": 151695
|
| 292365 |
+
},
|
| 292366 |
+
{
|
| 292367 |
+
"epoch": 1223.0,
|
| 292368 |
+
"eval_loss": 0.3656209409236908,
|
| 292369 |
+
"eval_runtime": 41.975,
|
| 292370 |
+
"eval_samples_per_second": 19.917,
|
| 292371 |
+
"eval_steps_per_second": 0.643,
|
| 292372 |
+
"eval_wer": 0.16889752446747267,
|
| 292373 |
+
"step": 151696
|
| 292374 |
+
},
|
| 292375 |
+
{
|
| 292376 |
+
"epoch": 1223.03,
|
| 292377 |
+
"learning_rate": 7.566381260096931e-06,
|
| 292378 |
+
"loss": 0.3057,
|
| 292379 |
+
"step": 151700
|
| 292380 |
+
},
|
| 292381 |
+
{
|
| 292382 |
+
"epoch": 1223.07,
|
| 292383 |
+
"learning_rate": 7.566300484652666e-06,
|
| 292384 |
+
"loss": 0.3804,
|
| 292385 |
+
"step": 151705
|
| 292386 |
+
},
|
| 292387 |
+
{
|
| 292388 |
+
"epoch": 1223.11,
|
| 292389 |
+
"learning_rate": 7.566219709208401e-06,
|
| 292390 |
+
"loss": 0.3112,
|
| 292391 |
+
"step": 151710
|
| 292392 |
+
},
|
| 292393 |
+
{
|
| 292394 |
+
"epoch": 1223.15,
|
| 292395 |
+
"learning_rate": 7.566138933764136e-06,
|
| 292396 |
+
"loss": 0.3271,
|
| 292397 |
+
"step": 151715
|
| 292398 |
+
},
|
| 292399 |
+
{
|
| 292400 |
+
"epoch": 1223.19,
|
| 292401 |
+
"learning_rate": 7.566058158319871e-06,
|
| 292402 |
+
"loss": 0.8103,
|
| 292403 |
+
"step": 151720
|
| 292404 |
+
},
|
| 292405 |
+
{
|
| 292406 |
+
"epoch": 1223.23,
|
| 292407 |
+
"learning_rate": 7.565977382875606e-06,
|
| 292408 |
+
"loss": 0.6019,
|
| 292409 |
+
"step": 151725
|
| 292410 |
+
},
|
| 292411 |
+
{
|
| 292412 |
+
"epoch": 1223.27,
|
| 292413 |
+
"learning_rate": 7.565896607431341e-06,
|
| 292414 |
+
"loss": 0.2601,
|
| 292415 |
+
"step": 151730
|
| 292416 |
+
},
|
| 292417 |
+
{
|
| 292418 |
+
"epoch": 1223.31,
|
| 292419 |
+
"learning_rate": 7.565815831987076e-06,
|
| 292420 |
+
"loss": 0.2529,
|
| 292421 |
+
"step": 151735
|
| 292422 |
+
},
|
| 292423 |
+
{
|
| 292424 |
+
"epoch": 1223.35,
|
| 292425 |
+
"learning_rate": 7.565735056542811e-06,
|
| 292426 |
+
"loss": 0.3854,
|
| 292427 |
+
"step": 151740
|
| 292428 |
+
},
|
| 292429 |
+
{
|
| 292430 |
+
"epoch": 1223.4,
|
| 292431 |
+
"learning_rate": 7.565654281098547e-06,
|
| 292432 |
+
"loss": 0.8185,
|
| 292433 |
+
"step": 151745
|
| 292434 |
+
},
|
| 292435 |
+
{
|
| 292436 |
+
"epoch": 1223.44,
|
| 292437 |
+
"learning_rate": 7.565573505654281e-06,
|
| 292438 |
+
"loss": 0.5565,
|
| 292439 |
+
"step": 151750
|
| 292440 |
+
},
|
| 292441 |
+
{
|
| 292442 |
+
"epoch": 1223.48,
|
| 292443 |
+
"learning_rate": 7.565492730210017e-06,
|
| 292444 |
+
"loss": 0.2581,
|
| 292445 |
+
"step": 151755
|
| 292446 |
+
},
|
| 292447 |
+
{
|
| 292448 |
+
"epoch": 1223.52,
|
| 292449 |
+
"learning_rate": 7.565411954765752e-06,
|
| 292450 |
+
"loss": 0.2536,
|
| 292451 |
+
"step": 151760
|
| 292452 |
+
},
|
| 292453 |
+
{
|
| 292454 |
+
"epoch": 1223.56,
|
| 292455 |
+
"learning_rate": 7.565331179321487e-06,
|
| 292456 |
+
"loss": 0.3955,
|
| 292457 |
+
"step": 151765
|
| 292458 |
+
},
|
| 292459 |
+
{
|
| 292460 |
+
"epoch": 1223.6,
|
| 292461 |
+
"learning_rate": 7.565250403877222e-06,
|
| 292462 |
+
"loss": 0.8402,
|
| 292463 |
+
"step": 151770
|
| 292464 |
+
},
|
| 292465 |
+
{
|
| 292466 |
+
"epoch": 1223.64,
|
| 292467 |
+
"learning_rate": 7.565169628432957e-06,
|
| 292468 |
+
"loss": 0.6756,
|
| 292469 |
+
"step": 151775
|
| 292470 |
+
},
|
| 292471 |
+
{
|
| 292472 |
+
"epoch": 1223.68,
|
| 292473 |
+
"learning_rate": 7.565088852988692e-06,
|
| 292474 |
+
"loss": 0.2822,
|
| 292475 |
+
"step": 151780
|
| 292476 |
+
},
|
| 292477 |
+
{
|
| 292478 |
+
"epoch": 1223.72,
|
| 292479 |
+
"learning_rate": 7.565008077544427e-06,
|
| 292480 |
+
"loss": 0.3014,
|
| 292481 |
+
"step": 151785
|
| 292482 |
+
},
|
| 292483 |
+
{
|
| 292484 |
+
"epoch": 1223.76,
|
| 292485 |
+
"learning_rate": 7.564927302100162e-06,
|
| 292486 |
+
"loss": 0.4023,
|
| 292487 |
+
"step": 151790
|
| 292488 |
+
},
|
| 292489 |
+
{
|
| 292490 |
+
"epoch": 1223.8,
|
| 292491 |
+
"learning_rate": 7.564846526655897e-06,
|
| 292492 |
+
"loss": 0.8358,
|
| 292493 |
+
"step": 151795
|
| 292494 |
+
},
|
| 292495 |
+
{
|
| 292496 |
+
"epoch": 1223.84,
|
| 292497 |
+
"learning_rate": 7.564765751211632e-06,
|
| 292498 |
+
"loss": 0.7451,
|
| 292499 |
+
"step": 151800
|
| 292500 |
+
},
|
| 292501 |
+
{
|
| 292502 |
+
"epoch": 1223.88,
|
| 292503 |
+
"learning_rate": 7.564684975767367e-06,
|
| 292504 |
+
"loss": 0.2822,
|
| 292505 |
+
"step": 151805
|
| 292506 |
+
},
|
| 292507 |
+
{
|
| 292508 |
+
"epoch": 1223.92,
|
| 292509 |
+
"learning_rate": 7.564604200323103e-06,
|
| 292510 |
+
"loss": 0.2669,
|
| 292511 |
+
"step": 151810
|
| 292512 |
+
},
|
| 292513 |
+
{
|
| 292514 |
+
"epoch": 1223.96,
|
| 292515 |
+
"learning_rate": 7.564523424878838e-06,
|
| 292516 |
+
"loss": 0.4182,
|
| 292517 |
+
"step": 151815
|
| 292518 |
+
},
|
| 292519 |
+
{
|
| 292520 |
+
"epoch": 1224.0,
|
| 292521 |
+
"learning_rate": 7.564442649434573e-06,
|
| 292522 |
+
"loss": 1.1147,
|
| 292523 |
+
"step": 151820
|
| 292524 |
+
},
|
| 292525 |
+
{
|
| 292526 |
+
"epoch": 1224.0,
|
| 292527 |
+
"eval_loss": 0.34463566541671753,
|
| 292528 |
+
"eval_runtime": 39.9445,
|
| 292529 |
+
"eval_samples_per_second": 20.929,
|
| 292530 |
+
"eval_steps_per_second": 0.676,
|
| 292531 |
+
"eval_wer": 0.1835759611795446,
|
| 292532 |
+
"step": 151820
|
| 292533 |
+
},
|
| 292534 |
+
{
|
| 292535 |
+
"epoch": 1224.04,
|
| 292536 |
+
"learning_rate": 7.564361873990308e-06,
|
| 292537 |
+
"loss": 0.277,
|
| 292538 |
+
"step": 151825
|
| 292539 |
+
},
|
| 292540 |
+
{
|
| 292541 |
+
"epoch": 1224.08,
|
| 292542 |
+
"learning_rate": 7.564281098546043e-06,
|
| 292543 |
+
"loss": 0.2902,
|
| 292544 |
+
"step": 151830
|
| 292545 |
+
},
|
| 292546 |
+
{
|
| 292547 |
+
"epoch": 1224.12,
|
| 292548 |
+
"learning_rate": 7.564200323101778e-06,
|
| 292549 |
+
"loss": 0.2973,
|
| 292550 |
+
"step": 151835
|
| 292551 |
+
},
|
| 292552 |
+
{
|
| 292553 |
+
"epoch": 1224.16,
|
| 292554 |
+
"learning_rate": 7.564119547657513e-06,
|
| 292555 |
+
"loss": 0.4297,
|
| 292556 |
+
"step": 151840
|
| 292557 |
+
},
|
| 292558 |
+
{
|
| 292559 |
+
"epoch": 1224.2,
|
| 292560 |
+
"learning_rate": 7.564038772213248e-06,
|
| 292561 |
+
"loss": 1.2619,
|
| 292562 |
+
"step": 151845
|
| 292563 |
+
},
|
| 292564 |
+
{
|
| 292565 |
+
"epoch": 1224.24,
|
| 292566 |
+
"learning_rate": 7.563957996768983e-06,
|
| 292567 |
+
"loss": 0.3364,
|
| 292568 |
+
"step": 151850
|
| 292569 |
+
},
|
| 292570 |
+
{
|
| 292571 |
+
"epoch": 1224.28,
|
| 292572 |
+
"learning_rate": 7.563877221324718e-06,
|
| 292573 |
+
"loss": 0.2814,
|
| 292574 |
+
"step": 151855
|
| 292575 |
+
},
|
| 292576 |
+
{
|
| 292577 |
+
"epoch": 1224.32,
|
| 292578 |
+
"learning_rate": 7.563796445880453e-06,
|
| 292579 |
+
"loss": 0.3734,
|
| 292580 |
+
"step": 151860
|
| 292581 |
+
},
|
| 292582 |
+
{
|
| 292583 |
+
"epoch": 1224.36,
|
| 292584 |
+
"learning_rate": 7.563715670436188e-06,
|
| 292585 |
+
"loss": 0.4837,
|
| 292586 |
+
"step": 151865
|
| 292587 |
+
},
|
| 292588 |
+
{
|
| 292589 |
+
"epoch": 1224.4,
|
| 292590 |
+
"learning_rate": 7.5636348949919235e-06,
|
| 292591 |
+
"loss": 1.1229,
|
| 292592 |
+
"step": 151870
|
| 292593 |
+
},
|
| 292594 |
+
{
|
| 292595 |
+
"epoch": 1224.44,
|
| 292596 |
+
"learning_rate": 7.5635541195476585e-06,
|
| 292597 |
+
"loss": 0.3491,
|
| 292598 |
+
"step": 151875
|
| 292599 |
+
},
|
| 292600 |
+
{
|
| 292601 |
+
"epoch": 1224.48,
|
| 292602 |
+
"learning_rate": 7.5634733441033935e-06,
|
| 292603 |
+
"loss": 0.2809,
|
| 292604 |
+
"step": 151880
|
| 292605 |
+
},
|
| 292606 |
+
{
|
| 292607 |
+
"epoch": 1224.52,
|
| 292608 |
+
"learning_rate": 7.5633925686591284e-06,
|
| 292609 |
+
"loss": 0.2803,
|
| 292610 |
+
"step": 151885
|
| 292611 |
+
},
|
| 292612 |
+
{
|
| 292613 |
+
"epoch": 1224.56,
|
| 292614 |
+
"learning_rate": 7.5633117932148634e-06,
|
| 292615 |
+
"loss": 0.4462,
|
| 292616 |
+
"step": 151890
|
| 292617 |
+
},
|
| 292618 |
+
{
|
| 292619 |
+
"epoch": 1224.6,
|
| 292620 |
+
"learning_rate": 7.563231017770598e-06,
|
| 292621 |
+
"loss": 1.3136,
|
| 292622 |
+
"step": 151895
|
| 292623 |
+
},
|
| 292624 |
+
{
|
| 292625 |
+
"epoch": 1224.65,
|
| 292626 |
+
"learning_rate": 7.563150242326333e-06,
|
| 292627 |
+
"loss": 0.2907,
|
| 292628 |
+
"step": 151900
|
| 292629 |
+
},
|
| 292630 |
+
{
|
| 292631 |
+
"epoch": 1224.69,
|
| 292632 |
+
"learning_rate": 7.563069466882068e-06,
|
| 292633 |
+
"loss": 0.2524,
|
| 292634 |
+
"step": 151905
|
| 292635 |
+
},
|
| 292636 |
+
{
|
| 292637 |
+
"epoch": 1224.73,
|
| 292638 |
+
"learning_rate": 7.562988691437803e-06,
|
| 292639 |
+
"loss": 0.3055,
|
| 292640 |
+
"step": 151910
|
| 292641 |
+
},
|
| 292642 |
+
{
|
| 292643 |
+
"epoch": 1224.77,
|
| 292644 |
+
"learning_rate": 7.562907915993538e-06,
|
| 292645 |
+
"loss": 0.4168,
|
| 292646 |
+
"step": 151915
|
| 292647 |
+
},
|
| 292648 |
+
{
|
| 292649 |
+
"epoch": 1224.81,
|
| 292650 |
+
"learning_rate": 7.562827140549273e-06,
|
| 292651 |
+
"loss": 0.9734,
|
| 292652 |
+
"step": 151920
|
| 292653 |
+
},
|
| 292654 |
+
{
|
| 292655 |
+
"epoch": 1224.85,
|
| 292656 |
+
"learning_rate": 7.562746365105008e-06,
|
| 292657 |
+
"loss": 0.2457,
|
| 292658 |
+
"step": 151925
|
| 292659 |
+
},
|
| 292660 |
+
{
|
| 292661 |
+
"epoch": 1224.89,
|
| 292662 |
+
"learning_rate": 7.562665589660743e-06,
|
| 292663 |
+
"loss": 0.2464,
|
| 292664 |
+
"step": 151930
|
| 292665 |
+
},
|
| 292666 |
+
{
|
| 292667 |
+
"epoch": 1224.93,
|
| 292668 |
+
"learning_rate": 7.562584814216479e-06,
|
| 292669 |
+
"loss": 0.2614,
|
| 292670 |
+
"step": 151935
|
| 292671 |
+
},
|
| 292672 |
+
{
|
| 292673 |
+
"epoch": 1224.97,
|
| 292674 |
+
"learning_rate": 7.562504038772213e-06,
|
| 292675 |
+
"loss": 0.4991,
|
| 292676 |
+
"step": 151940
|
| 292677 |
+
},
|
| 292678 |
+
{
|
| 292679 |
+
"epoch": 1225.0,
|
| 292680 |
+
"eval_loss": 0.344937801361084,
|
| 292681 |
+
"eval_runtime": 40.5017,
|
| 292682 |
+
"eval_samples_per_second": 20.666,
|
| 292683 |
+
"eval_steps_per_second": 0.667,
|
| 292684 |
+
"eval_wer": 0.1767148683722655,
|
| 292685 |
+
"step": 151944
|
| 292686 |
+
},
|
| 292687 |
+
{
|
| 292688 |
+
"epoch": 1225.01,
|
| 292689 |
+
"learning_rate": 7.562423263327949e-06,
|
| 292690 |
+
"loss": 0.2553,
|
| 292691 |
+
"step": 151945
|
| 292692 |
+
},
|
| 292693 |
+
{
|
| 292694 |
+
"epoch": 1225.05,
|
| 292695 |
+
"learning_rate": 7.562342487883684e-06,
|
| 292696 |
+
"loss": 0.2594,
|
| 292697 |
+
"step": 151950
|
| 292698 |
+
},
|
| 292699 |
+
{
|
| 292700 |
+
"epoch": 1225.09,
|
| 292701 |
+
"learning_rate": 7.562261712439419e-06,
|
| 292702 |
+
"loss": 0.2837,
|
| 292703 |
+
"step": 151955
|
| 292704 |
+
},
|
| 292705 |
+
{
|
| 292706 |
+
"epoch": 1225.13,
|
| 292707 |
+
"learning_rate": 7.562180936995154e-06,
|
| 292708 |
+
"loss": 0.2949,
|
| 292709 |
+
"step": 151960
|
| 292710 |
+
},
|
| 292711 |
+
{
|
| 292712 |
+
"epoch": 1225.17,
|
| 292713 |
+
"learning_rate": 7.562100161550889e-06,
|
| 292714 |
+
"loss": 0.4855,
|
| 292715 |
+
"step": 151965
|
| 292716 |
+
},
|
| 292717 |
+
{
|
| 292718 |
+
"epoch": 1225.21,
|
| 292719 |
+
"learning_rate": 7.562019386106624e-06,
|
| 292720 |
+
"loss": 1.1717,
|
| 292721 |
+
"step": 151970
|
| 292722 |
+
},
|
| 292723 |
+
{
|
| 292724 |
+
"epoch": 1225.25,
|
| 292725 |
+
"learning_rate": 7.561938610662359e-06,
|
| 292726 |
+
"loss": 0.3164,
|
| 292727 |
+
"step": 151975
|
| 292728 |
+
},
|
| 292729 |
+
{
|
| 292730 |
+
"epoch": 1225.29,
|
| 292731 |
+
"learning_rate": 7.561857835218094e-06,
|
| 292732 |
+
"loss": 0.2502,
|
| 292733 |
+
"step": 151980
|
| 292734 |
+
},
|
| 292735 |
+
{
|
| 292736 |
+
"epoch": 1225.33,
|
| 292737 |
+
"learning_rate": 7.561777059773829e-06,
|
| 292738 |
+
"loss": 0.2958,
|
| 292739 |
+
"step": 151985
|
| 292740 |
+
},
|
| 292741 |
+
{
|
| 292742 |
+
"epoch": 1225.37,
|
| 292743 |
+
"learning_rate": 7.561696284329565e-06,
|
| 292744 |
+
"loss": 0.4835,
|
| 292745 |
+
"step": 151990
|
| 292746 |
+
},
|
| 292747 |
+
{
|
| 292748 |
+
"epoch": 1225.41,
|
| 292749 |
+
"learning_rate": 7.561615508885299e-06,
|
| 292750 |
+
"loss": 1.2292,
|
| 292751 |
+
"step": 151995
|
| 292752 |
+
},
|
| 292753 |
+
{
|
| 292754 |
+
"epoch": 1225.45,
|
| 292755 |
+
"learning_rate": 7.561534733441035e-06,
|
| 292756 |
+
"loss": 0.3822,
|
| 292757 |
+
"step": 152000
|
| 292758 |
+
},
|
| 292759 |
+
{
|
| 292760 |
+
"epoch": 1225.49,
|
| 292761 |
+
"learning_rate": 7.561453957996769e-06,
|
| 292762 |
+
"loss": 0.326,
|
| 292763 |
+
"step": 152005
|
| 292764 |
+
},
|
| 292765 |
+
{
|
| 292766 |
+
"epoch": 1225.53,
|
| 292767 |
+
"learning_rate": 7.561373182552505e-06,
|
| 292768 |
+
"loss": 0.3352,
|
| 292769 |
+
"step": 152010
|
| 292770 |
+
},
|
| 292771 |
+
{
|
| 292772 |
+
"epoch": 1225.57,
|
| 292773 |
+
"learning_rate": 7.56129240710824e-06,
|
| 292774 |
+
"loss": 0.5148,
|
| 292775 |
+
"step": 152015
|
| 292776 |
+
},
|
| 292777 |
+
{
|
| 292778 |
+
"epoch": 1225.61,
|
| 292779 |
+
"learning_rate": 7.561211631663975e-06,
|
| 292780 |
+
"loss": 1.0644,
|
| 292781 |
+
"step": 152020
|
| 292782 |
+
},
|
| 292783 |
+
{
|
| 292784 |
+
"epoch": 1225.65,
|
| 292785 |
+
"learning_rate": 7.56113085621971e-06,
|
| 292786 |
+
"loss": 0.2705,
|
| 292787 |
+
"step": 152025
|
| 292788 |
+
},
|
| 292789 |
+
{
|
| 292790 |
+
"epoch": 1225.69,
|
| 292791 |
+
"learning_rate": 7.561050080775445e-06,
|
| 292792 |
+
"loss": 0.2654,
|
| 292793 |
+
"step": 152030
|
| 292794 |
+
},
|
| 292795 |
+
{
|
| 292796 |
+
"epoch": 1225.73,
|
| 292797 |
+
"learning_rate": 7.56096930533118e-06,
|
| 292798 |
+
"loss": 0.2655,
|
| 292799 |
+
"step": 152035
|
| 292800 |
+
},
|
| 292801 |
+
{
|
| 292802 |
+
"epoch": 1225.77,
|
| 292803 |
+
"learning_rate": 7.560888529886915e-06,
|
| 292804 |
+
"loss": 0.5798,
|
| 292805 |
+
"step": 152040
|
| 292806 |
+
},
|
| 292807 |
+
{
|
| 292808 |
+
"epoch": 1225.81,
|
| 292809 |
+
"learning_rate": 7.560807754442651e-06,
|
| 292810 |
+
"loss": 1.0464,
|
| 292811 |
+
"step": 152045
|
| 292812 |
+
},
|
| 292813 |
+
{
|
| 292814 |
+
"epoch": 1225.85,
|
| 292815 |
+
"learning_rate": 7.560726978998385e-06,
|
| 292816 |
+
"loss": 0.264,
|
| 292817 |
+
"step": 152050
|
| 292818 |
+
},
|
| 292819 |
+
{
|
| 292820 |
+
"epoch": 1225.89,
|
| 292821 |
+
"learning_rate": 7.5606462035541206e-06,
|
| 292822 |
+
"loss": 0.2871,
|
| 292823 |
+
"step": 152055
|
| 292824 |
+
},
|
| 292825 |
+
{
|
| 292826 |
+
"epoch": 1225.93,
|
| 292827 |
+
"learning_rate": 7.560565428109855e-06,
|
| 292828 |
+
"loss": 0.3312,
|
| 292829 |
+
"step": 152060
|
| 292830 |
+
},
|
| 292831 |
+
{
|
| 292832 |
+
"epoch": 1225.97,
|
| 292833 |
+
"learning_rate": 7.5604846526655905e-06,
|
| 292834 |
+
"loss": 0.5482,
|
| 292835 |
+
"step": 152065
|
| 292836 |
+
},
|
| 292837 |
+
{
|
| 292838 |
+
"epoch": 1226.0,
|
| 292839 |
+
"eval_loss": 0.3725050091743469,
|
| 292840 |
+
"eval_runtime": 40.5085,
|
| 292841 |
+
"eval_samples_per_second": 20.662,
|
| 292842 |
+
"eval_steps_per_second": 0.667,
|
| 292843 |
+
"eval_wer": 0.1764970733903647,
|
| 292844 |
+
"step": 152068
|
| 292845 |
}
|
| 292846 |
],
|
| 292847 |
"max_steps": 620000,
|
| 292848 |
"num_train_epochs": 5000,
|
| 292849 |
+
"total_flos": 4.2793273628046655e+20,
|
| 292850 |
"trial_name": null,
|
| 292851 |
"trial_params": null
|
| 292852 |
}
|
model-bin/finetune/base/{checkpoint-151448 β checkpoint-152068}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1630248561.4202452/events.out.tfevents.1630248561.cc93b136ebf5.1086.199
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0e69c53824a99a78f62d60a3191904b4929d958e37a96aed121cea82ef6f44c6
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630248992.3105023/events.out.tfevents.1630248992.cc93b136ebf5.1086.201
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:18c11de576210df8c1bcd2e338686d0380f1a420b31373ffd06eb3aed8c1f453
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630249423.7833743/events.out.tfevents.1630249423.cc93b136ebf5.1086.203
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:568ece09adb0211ba3f15921d31806d68262b8cfa9857606486c11d4bae9edd6
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630249857.8968918/events.out.tfevents.1630249857.cc93b136ebf5.1086.205
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ca6c23b7959ed4eff92db23ba2ba96ded88c6e6bd51c3726d065275e78ced855
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630250287.4046617/events.out.tfevents.1630250287.cc93b136ebf5.1086.207
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0d4f64e2892b1172fc9ed28d3b5df11b5c1e87ce978ca5299337293b8153e478
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630248561.cc93b136ebf5.1086.198
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4839731a799003ae2c63f39a2b9d8d658d064e4e4d52528b686ea21a72782e95
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630248992.cc93b136ebf5.1086.200
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d0ef6e85b448f1e9daa36e511987fdd78a3da2fe85a0475a40bc89341de9b3cb
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630249423.cc93b136ebf5.1086.202
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7ad462f18f3ef5e7674be715c076f3a6912a2bdaa53ab1bde8f0f57357adc4f4
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630249857.cc93b136ebf5.1086.204
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8beff9185670015d1b0453cddc98932f4cf7524e092ff8056b49425d192625e8
|
| 3 |
+
size 8462
|
model-bin/finetune/base/log/events.out.tfevents.1630250287.cc93b136ebf5.1086.206
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4c75d446f14703e0a0c02a6515f68440f0868da9b0a01d30d83ccc76744caace
|
| 3 |
+
size 8622
|