"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-68444 β checkpoint-70187}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-68444 β checkpoint-70187}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-68444 β checkpoint-70187}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-68444 β checkpoint-70187}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-68444 β checkpoint-70187}/rng_state.pth +2 -2
- model-bin/finetune/base/{checkpoint-68444 β checkpoint-70187}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-68444 β checkpoint-70187}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-68444 β checkpoint-70187}/trainer_state.json +2225 -5
- model-bin/finetune/base/{checkpoint-68444 β checkpoint-70187}/training_args.bin +0 -0
- model-bin/finetune/base/log/1629852210.1896331/events.out.tfevents.1629852210.7e498afd5545.905.3 +3 -0
- model-bin/finetune/base/log/1629852859.7482474/events.out.tfevents.1629852859.7e498afd5545.905.5 +3 -0
- model-bin/finetune/base/log/1629853517.0685782/events.out.tfevents.1629853517.7e498afd5545.905.7 +3 -0
- model-bin/finetune/base/log/1629854178.1266189/events.out.tfevents.1629854178.7e498afd5545.905.9 +3 -0
- model-bin/finetune/base/log/1629854840.4334018/events.out.tfevents.1629854840.7e498afd5545.905.11 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629852210.7e498afd5545.905.2 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629852859.7e498afd5545.905.4 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629853517.7e498afd5545.905.6 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629854178.7e498afd5545.905.8 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629854840.7e498afd5545.905.10 +3 -0
model-bin/finetune/base/{checkpoint-68444 β checkpoint-70187}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-68444 β checkpoint-70187}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d42c1d3cc45fb142e63a83225b1e4e8813a7a1c71a71079fbd0436ff2fcffdeb
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-68444 β checkpoint-70187}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-68444 β checkpoint-70187}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3f360c97ee9fb26641222af40c8fbcce7706332a412dc1f8c5bc142a90772082
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-68444 β checkpoint-70187}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c855de793e0176a96ba172e2196f9871cead8ced6385a49047fad5bff62debda
|
| 3 |
+
size 14567
|
model-bin/finetune/base/{checkpoint-68444 β checkpoint-70187}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b9e8a3491d7004204355a5ab53074d965ab75a472b5554c923e87c0af173fe20
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-68444 β checkpoint-70187}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:735fb0428260ec5e2371142a8bacca8c588aafa478041f2c5de4c0d19ac647a3
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-68444 β checkpoint-70187}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
-
"best_metric": 0.
|
| 3 |
-
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -186444,11 +186444,2231 @@
|
|
| 186444 |
"eval_steps_per_second": 0.658,
|
| 186445 |
"eval_wer": 0.18525332578545145,
|
| 186446 |
"step": 68444
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 186447 |
}
|
| 186448 |
],
|
| 186449 |
"max_steps": 620000,
|
| 186450 |
"num_train_epochs": 5000,
|
| 186451 |
-
"total_flos": 1.
|
| 186452 |
"trial_name": null,
|
| 186453 |
"trial_params": null
|
| 186454 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"best_metric": 0.18412114350410416,
|
| 3 |
+
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-69565",
|
| 4 |
+
"epoch": 566.0,
|
| 5 |
+
"global_step": 70187,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 186444 |
"eval_steps_per_second": 0.658,
|
| 186445 |
"eval_wer": 0.18525332578545145,
|
| 186446 |
"step": 68444
|
| 186447 |
+
},
|
| 186448 |
+
{
|
| 186449 |
+
"epoch": 551.01,
|
| 186450 |
+
"learning_rate": 8.919567307692309e-06,
|
| 186451 |
+
"loss": 0.3373,
|
| 186452 |
+
"step": 68445
|
| 186453 |
+
},
|
| 186454 |
+
{
|
| 186455 |
+
"epoch": 551.05,
|
| 186456 |
+
"learning_rate": 8.91948717948718e-06,
|
| 186457 |
+
"loss": 0.2903,
|
| 186458 |
+
"step": 68450
|
| 186459 |
+
},
|
| 186460 |
+
{
|
| 186461 |
+
"epoch": 551.09,
|
| 186462 |
+
"learning_rate": 8.919407051282052e-06,
|
| 186463 |
+
"loss": 0.2579,
|
| 186464 |
+
"step": 68455
|
| 186465 |
+
},
|
| 186466 |
+
{
|
| 186467 |
+
"epoch": 551.13,
|
| 186468 |
+
"learning_rate": 8.919326923076925e-06,
|
| 186469 |
+
"loss": 0.3771,
|
| 186470 |
+
"step": 68460
|
| 186471 |
+
},
|
| 186472 |
+
{
|
| 186473 |
+
"epoch": 551.17,
|
| 186474 |
+
"learning_rate": 8.919246794871796e-06,
|
| 186475 |
+
"loss": 0.5803,
|
| 186476 |
+
"step": 68465
|
| 186477 |
+
},
|
| 186478 |
+
{
|
| 186479 |
+
"epoch": 551.21,
|
| 186480 |
+
"learning_rate": 8.919166666666668e-06,
|
| 186481 |
+
"loss": 1.1431,
|
| 186482 |
+
"step": 68470
|
| 186483 |
+
},
|
| 186484 |
+
{
|
| 186485 |
+
"epoch": 551.25,
|
| 186486 |
+
"learning_rate": 8.919086538461539e-06,
|
| 186487 |
+
"loss": 0.3321,
|
| 186488 |
+
"step": 68475
|
| 186489 |
+
},
|
| 186490 |
+
{
|
| 186491 |
+
"epoch": 551.29,
|
| 186492 |
+
"learning_rate": 8.919006410256412e-06,
|
| 186493 |
+
"loss": 0.3079,
|
| 186494 |
+
"step": 68480
|
| 186495 |
+
},
|
| 186496 |
+
{
|
| 186497 |
+
"epoch": 551.33,
|
| 186498 |
+
"learning_rate": 8.918926282051282e-06,
|
| 186499 |
+
"loss": 0.3175,
|
| 186500 |
+
"step": 68485
|
| 186501 |
+
},
|
| 186502 |
+
{
|
| 186503 |
+
"epoch": 551.37,
|
| 186504 |
+
"learning_rate": 8.918846153846155e-06,
|
| 186505 |
+
"loss": 0.6332,
|
| 186506 |
+
"step": 68490
|
| 186507 |
+
},
|
| 186508 |
+
{
|
| 186509 |
+
"epoch": 551.41,
|
| 186510 |
+
"learning_rate": 8.918766025641028e-06,
|
| 186511 |
+
"loss": 1.0984,
|
| 186512 |
+
"step": 68495
|
| 186513 |
+
},
|
| 186514 |
+
{
|
| 186515 |
+
"epoch": 551.45,
|
| 186516 |
+
"learning_rate": 8.918685897435897e-06,
|
| 186517 |
+
"loss": 0.3226,
|
| 186518 |
+
"step": 68500
|
| 186519 |
+
},
|
| 186520 |
+
{
|
| 186521 |
+
"epoch": 551.49,
|
| 186522 |
+
"learning_rate": 8.91860576923077e-06,
|
| 186523 |
+
"loss": 0.277,
|
| 186524 |
+
"step": 68505
|
| 186525 |
+
},
|
| 186526 |
+
{
|
| 186527 |
+
"epoch": 551.53,
|
| 186528 |
+
"learning_rate": 8.918525641025642e-06,
|
| 186529 |
+
"loss": 0.3595,
|
| 186530 |
+
"step": 68510
|
| 186531 |
+
},
|
| 186532 |
+
{
|
| 186533 |
+
"epoch": 551.57,
|
| 186534 |
+
"learning_rate": 8.918445512820513e-06,
|
| 186535 |
+
"loss": 0.6192,
|
| 186536 |
+
"step": 68515
|
| 186537 |
+
},
|
| 186538 |
+
{
|
| 186539 |
+
"epoch": 551.61,
|
| 186540 |
+
"learning_rate": 8.918365384615385e-06,
|
| 186541 |
+
"loss": 1.1009,
|
| 186542 |
+
"step": 68520
|
| 186543 |
+
},
|
| 186544 |
+
{
|
| 186545 |
+
"epoch": 551.65,
|
| 186546 |
+
"learning_rate": 8.918285256410258e-06,
|
| 186547 |
+
"loss": 0.3705,
|
| 186548 |
+
"step": 68525
|
| 186549 |
+
},
|
| 186550 |
+
{
|
| 186551 |
+
"epoch": 551.69,
|
| 186552 |
+
"learning_rate": 8.918205128205129e-06,
|
| 186553 |
+
"loss": 0.2911,
|
| 186554 |
+
"step": 68530
|
| 186555 |
+
},
|
| 186556 |
+
{
|
| 186557 |
+
"epoch": 551.73,
|
| 186558 |
+
"learning_rate": 8.918125e-06,
|
| 186559 |
+
"loss": 0.3628,
|
| 186560 |
+
"step": 68535
|
| 186561 |
+
},
|
| 186562 |
+
{
|
| 186563 |
+
"epoch": 551.77,
|
| 186564 |
+
"learning_rate": 8.918044871794872e-06,
|
| 186565 |
+
"loss": 0.533,
|
| 186566 |
+
"step": 68540
|
| 186567 |
+
},
|
| 186568 |
+
{
|
| 186569 |
+
"epoch": 551.81,
|
| 186570 |
+
"learning_rate": 8.917964743589745e-06,
|
| 186571 |
+
"loss": 1.2326,
|
| 186572 |
+
"step": 68545
|
| 186573 |
+
},
|
| 186574 |
+
{
|
| 186575 |
+
"epoch": 551.85,
|
| 186576 |
+
"learning_rate": 8.917884615384616e-06,
|
| 186577 |
+
"loss": 0.3388,
|
| 186578 |
+
"step": 68550
|
| 186579 |
+
},
|
| 186580 |
+
{
|
| 186581 |
+
"epoch": 551.89,
|
| 186582 |
+
"learning_rate": 8.917804487179487e-06,
|
| 186583 |
+
"loss": 0.3131,
|
| 186584 |
+
"step": 68555
|
| 186585 |
+
},
|
| 186586 |
+
{
|
| 186587 |
+
"epoch": 551.93,
|
| 186588 |
+
"learning_rate": 8.91772435897436e-06,
|
| 186589 |
+
"loss": 0.3515,
|
| 186590 |
+
"step": 68560
|
| 186591 |
+
},
|
| 186592 |
+
{
|
| 186593 |
+
"epoch": 551.97,
|
| 186594 |
+
"learning_rate": 8.917644230769232e-06,
|
| 186595 |
+
"loss": 0.6563,
|
| 186596 |
+
"step": 68565
|
| 186597 |
+
},
|
| 186598 |
+
{
|
| 186599 |
+
"epoch": 552.0,
|
| 186600 |
+
"eval_loss": 0.4913657009601593,
|
| 186601 |
+
"eval_runtime": 39.6316,
|
| 186602 |
+
"eval_samples_per_second": 21.17,
|
| 186603 |
+
"eval_steps_per_second": 0.681,
|
| 186604 |
+
"eval_wer": 0.19267410087881248,
|
| 186605 |
+
"step": 68568
|
| 186606 |
+
},
|
| 186607 |
+
{
|
| 186608 |
+
"epoch": 548.02,
|
| 186609 |
+
"learning_rate": 8.917564102564103e-06,
|
| 186610 |
+
"loss": 0.4739,
|
| 186611 |
+
"step": 68570
|
| 186612 |
+
},
|
| 186613 |
+
{
|
| 186614 |
+
"epoch": 548.06,
|
| 186615 |
+
"learning_rate": 8.917483974358975e-06,
|
| 186616 |
+
"loss": 0.316,
|
| 186617 |
+
"step": 68575
|
| 186618 |
+
},
|
| 186619 |
+
{
|
| 186620 |
+
"epoch": 548.1,
|
| 186621 |
+
"learning_rate": 8.917403846153848e-06,
|
| 186622 |
+
"loss": 0.3383,
|
| 186623 |
+
"step": 68580
|
| 186624 |
+
},
|
| 186625 |
+
{
|
| 186626 |
+
"epoch": 548.14,
|
| 186627 |
+
"learning_rate": 8.917323717948719e-06,
|
| 186628 |
+
"loss": 0.385,
|
| 186629 |
+
"step": 68585
|
| 186630 |
+
},
|
| 186631 |
+
{
|
| 186632 |
+
"epoch": 548.18,
|
| 186633 |
+
"learning_rate": 8.91724358974359e-06,
|
| 186634 |
+
"loss": 0.6467,
|
| 186635 |
+
"step": 68590
|
| 186636 |
+
},
|
| 186637 |
+
{
|
| 186638 |
+
"epoch": 548.22,
|
| 186639 |
+
"learning_rate": 8.917163461538463e-06,
|
| 186640 |
+
"loss": 1.0687,
|
| 186641 |
+
"step": 68595
|
| 186642 |
+
},
|
| 186643 |
+
{
|
| 186644 |
+
"epoch": 548.26,
|
| 186645 |
+
"learning_rate": 8.917083333333335e-06,
|
| 186646 |
+
"loss": 0.33,
|
| 186647 |
+
"step": 68600
|
| 186648 |
+
},
|
| 186649 |
+
{
|
| 186650 |
+
"epoch": 548.3,
|
| 186651 |
+
"learning_rate": 8.917003205128206e-06,
|
| 186652 |
+
"loss": 0.3272,
|
| 186653 |
+
"step": 68605
|
| 186654 |
+
},
|
| 186655 |
+
{
|
| 186656 |
+
"epoch": 548.34,
|
| 186657 |
+
"learning_rate": 8.916923076923077e-06,
|
| 186658 |
+
"loss": 0.3578,
|
| 186659 |
+
"step": 68610
|
| 186660 |
+
},
|
| 186661 |
+
{
|
| 186662 |
+
"epoch": 548.38,
|
| 186663 |
+
"learning_rate": 8.91684294871795e-06,
|
| 186664 |
+
"loss": 0.6878,
|
| 186665 |
+
"step": 68615
|
| 186666 |
+
},
|
| 186667 |
+
{
|
| 186668 |
+
"epoch": 548.42,
|
| 186669 |
+
"learning_rate": 8.91676282051282e-06,
|
| 186670 |
+
"loss": 0.9201,
|
| 186671 |
+
"step": 68620
|
| 186672 |
+
},
|
| 186673 |
+
{
|
| 186674 |
+
"epoch": 548.46,
|
| 186675 |
+
"learning_rate": 8.916682692307693e-06,
|
| 186676 |
+
"loss": 0.2895,
|
| 186677 |
+
"step": 68625
|
| 186678 |
+
},
|
| 186679 |
+
{
|
| 186680 |
+
"epoch": 548.5,
|
| 186681 |
+
"learning_rate": 8.916602564102565e-06,
|
| 186682 |
+
"loss": 0.3369,
|
| 186683 |
+
"step": 68630
|
| 186684 |
+
},
|
| 186685 |
+
{
|
| 186686 |
+
"epoch": 548.54,
|
| 186687 |
+
"learning_rate": 8.916522435897436e-06,
|
| 186688 |
+
"loss": 0.3617,
|
| 186689 |
+
"step": 68635
|
| 186690 |
+
},
|
| 186691 |
+
{
|
| 186692 |
+
"epoch": 548.58,
|
| 186693 |
+
"learning_rate": 8.916442307692307e-06,
|
| 186694 |
+
"loss": 0.5913,
|
| 186695 |
+
"step": 68640
|
| 186696 |
+
},
|
| 186697 |
+
{
|
| 186698 |
+
"epoch": 548.62,
|
| 186699 |
+
"learning_rate": 8.91636217948718e-06,
|
| 186700 |
+
"loss": 1.1614,
|
| 186701 |
+
"step": 68645
|
| 186702 |
+
},
|
| 186703 |
+
{
|
| 186704 |
+
"epoch": 548.66,
|
| 186705 |
+
"learning_rate": 8.916282051282052e-06,
|
| 186706 |
+
"loss": 0.2917,
|
| 186707 |
+
"step": 68650
|
| 186708 |
+
},
|
| 186709 |
+
{
|
| 186710 |
+
"epoch": 548.7,
|
| 186711 |
+
"learning_rate": 8.916201923076923e-06,
|
| 186712 |
+
"loss": 0.3465,
|
| 186713 |
+
"step": 68655
|
| 186714 |
+
},
|
| 186715 |
+
{
|
| 186716 |
+
"epoch": 548.74,
|
| 186717 |
+
"learning_rate": 8.916121794871796e-06,
|
| 186718 |
+
"loss": 0.3442,
|
| 186719 |
+
"step": 68660
|
| 186720 |
+
},
|
| 186721 |
+
{
|
| 186722 |
+
"epoch": 548.78,
|
| 186723 |
+
"learning_rate": 8.916041666666667e-06,
|
| 186724 |
+
"loss": 0.7135,
|
| 186725 |
+
"step": 68665
|
| 186726 |
+
},
|
| 186727 |
+
{
|
| 186728 |
+
"epoch": 548.82,
|
| 186729 |
+
"learning_rate": 8.915961538461539e-06,
|
| 186730 |
+
"loss": 1.1436,
|
| 186731 |
+
"step": 68670
|
| 186732 |
+
},
|
| 186733 |
+
{
|
| 186734 |
+
"epoch": 548.86,
|
| 186735 |
+
"learning_rate": 8.91588141025641e-06,
|
| 186736 |
+
"loss": 0.3227,
|
| 186737 |
+
"step": 68675
|
| 186738 |
+
},
|
| 186739 |
+
{
|
| 186740 |
+
"epoch": 548.9,
|
| 186741 |
+
"learning_rate": 8.915801282051283e-06,
|
| 186742 |
+
"loss": 0.3273,
|
| 186743 |
+
"step": 68680
|
| 186744 |
+
},
|
| 186745 |
+
{
|
| 186746 |
+
"epoch": 548.94,
|
| 186747 |
+
"learning_rate": 8.915721153846155e-06,
|
| 186748 |
+
"loss": 0.3505,
|
| 186749 |
+
"step": 68685
|
| 186750 |
+
},
|
| 186751 |
+
{
|
| 186752 |
+
"epoch": 548.98,
|
| 186753 |
+
"learning_rate": 8.915641025641026e-06,
|
| 186754 |
+
"loss": 0.6269,
|
| 186755 |
+
"step": 68690
|
| 186756 |
+
},
|
| 186757 |
+
{
|
| 186758 |
+
"epoch": 549.0,
|
| 186759 |
+
"eval_loss": 0.4549594819545746,
|
| 186760 |
+
"eval_runtime": 40.2207,
|
| 186761 |
+
"eval_samples_per_second": 20.86,
|
| 186762 |
+
"eval_steps_per_second": 0.671,
|
| 186763 |
+
"eval_wer": 0.1973022238425082,
|
| 186764 |
+
"step": 68693
|
| 186765 |
+
},
|
| 186766 |
+
{
|
| 186767 |
+
"epoch": 549.02,
|
| 186768 |
+
"learning_rate": 8.915560897435899e-06,
|
| 186769 |
+
"loss": 0.4356,
|
| 186770 |
+
"step": 68695
|
| 186771 |
+
},
|
| 186772 |
+
{
|
| 186773 |
+
"epoch": 549.06,
|
| 186774 |
+
"learning_rate": 8.91548076923077e-06,
|
| 186775 |
+
"loss": 0.3655,
|
| 186776 |
+
"step": 68700
|
| 186777 |
+
},
|
| 186778 |
+
{
|
| 186779 |
+
"epoch": 549.1,
|
| 186780 |
+
"learning_rate": 8.915400641025642e-06,
|
| 186781 |
+
"loss": 0.3215,
|
| 186782 |
+
"step": 68705
|
| 186783 |
+
},
|
| 186784 |
+
{
|
| 186785 |
+
"epoch": 549.14,
|
| 186786 |
+
"learning_rate": 8.915320512820513e-06,
|
| 186787 |
+
"loss": 0.3813,
|
| 186788 |
+
"step": 68710
|
| 186789 |
+
},
|
| 186790 |
+
{
|
| 186791 |
+
"epoch": 549.18,
|
| 186792 |
+
"learning_rate": 8.915240384615386e-06,
|
| 186793 |
+
"loss": 0.6744,
|
| 186794 |
+
"step": 68715
|
| 186795 |
+
},
|
| 186796 |
+
{
|
| 186797 |
+
"epoch": 549.22,
|
| 186798 |
+
"learning_rate": 8.915160256410257e-06,
|
| 186799 |
+
"loss": 0.9949,
|
| 186800 |
+
"step": 68720
|
| 186801 |
+
},
|
| 186802 |
+
{
|
| 186803 |
+
"epoch": 549.25,
|
| 186804 |
+
"learning_rate": 8.915080128205129e-06,
|
| 186805 |
+
"loss": 0.3279,
|
| 186806 |
+
"step": 68725
|
| 186807 |
+
},
|
| 186808 |
+
{
|
| 186809 |
+
"epoch": 549.29,
|
| 186810 |
+
"learning_rate": 8.915e-06,
|
| 186811 |
+
"loss": 0.3263,
|
| 186812 |
+
"step": 68730
|
| 186813 |
+
},
|
| 186814 |
+
{
|
| 186815 |
+
"epoch": 549.33,
|
| 186816 |
+
"learning_rate": 8.914919871794873e-06,
|
| 186817 |
+
"loss": 0.3473,
|
| 186818 |
+
"step": 68735
|
| 186819 |
+
},
|
| 186820 |
+
{
|
| 186821 |
+
"epoch": 549.37,
|
| 186822 |
+
"learning_rate": 8.914839743589745e-06,
|
| 186823 |
+
"loss": 0.628,
|
| 186824 |
+
"step": 68740
|
| 186825 |
+
},
|
| 186826 |
+
{
|
| 186827 |
+
"epoch": 549.41,
|
| 186828 |
+
"learning_rate": 8.914759615384616e-06,
|
| 186829 |
+
"loss": 0.9994,
|
| 186830 |
+
"step": 68745
|
| 186831 |
+
},
|
| 186832 |
+
{
|
| 186833 |
+
"epoch": 549.45,
|
| 186834 |
+
"learning_rate": 8.914679487179489e-06,
|
| 186835 |
+
"loss": 0.3344,
|
| 186836 |
+
"step": 68750
|
| 186837 |
+
},
|
| 186838 |
+
{
|
| 186839 |
+
"epoch": 549.49,
|
| 186840 |
+
"learning_rate": 8.91459935897436e-06,
|
| 186841 |
+
"loss": 0.2747,
|
| 186842 |
+
"step": 68755
|
| 186843 |
+
},
|
| 186844 |
+
{
|
| 186845 |
+
"epoch": 549.53,
|
| 186846 |
+
"learning_rate": 8.914519230769232e-06,
|
| 186847 |
+
"loss": 0.4123,
|
| 186848 |
+
"step": 68760
|
| 186849 |
+
},
|
| 186850 |
+
{
|
| 186851 |
+
"epoch": 549.57,
|
| 186852 |
+
"learning_rate": 8.914439102564103e-06,
|
| 186853 |
+
"loss": 0.7106,
|
| 186854 |
+
"step": 68765
|
| 186855 |
+
},
|
| 186856 |
+
{
|
| 186857 |
+
"epoch": 549.61,
|
| 186858 |
+
"learning_rate": 8.914358974358976e-06,
|
| 186859 |
+
"loss": 1.1318,
|
| 186860 |
+
"step": 68770
|
| 186861 |
+
},
|
| 186862 |
+
{
|
| 186863 |
+
"epoch": 549.65,
|
| 186864 |
+
"learning_rate": 8.914278846153846e-06,
|
| 186865 |
+
"loss": 0.4238,
|
| 186866 |
+
"step": 68775
|
| 186867 |
+
},
|
| 186868 |
+
{
|
| 186869 |
+
"epoch": 549.69,
|
| 186870 |
+
"learning_rate": 8.914198717948719e-06,
|
| 186871 |
+
"loss": 0.4404,
|
| 186872 |
+
"step": 68780
|
| 186873 |
+
},
|
| 186874 |
+
{
|
| 186875 |
+
"epoch": 549.73,
|
| 186876 |
+
"learning_rate": 8.91411858974359e-06,
|
| 186877 |
+
"loss": 0.3851,
|
| 186878 |
+
"step": 68785
|
| 186879 |
+
},
|
| 186880 |
+
{
|
| 186881 |
+
"epoch": 549.77,
|
| 186882 |
+
"learning_rate": 8.914038461538462e-06,
|
| 186883 |
+
"loss": 0.6609,
|
| 186884 |
+
"step": 68790
|
| 186885 |
+
},
|
| 186886 |
+
{
|
| 186887 |
+
"epoch": 549.81,
|
| 186888 |
+
"learning_rate": 8.913958333333335e-06,
|
| 186889 |
+
"loss": 1.2404,
|
| 186890 |
+
"step": 68795
|
| 186891 |
+
},
|
| 186892 |
+
{
|
| 186893 |
+
"epoch": 549.85,
|
| 186894 |
+
"learning_rate": 8.913878205128206e-06,
|
| 186895 |
+
"loss": 0.3031,
|
| 186896 |
+
"step": 68800
|
| 186897 |
+
},
|
| 186898 |
+
{
|
| 186899 |
+
"epoch": 549.89,
|
| 186900 |
+
"learning_rate": 8.913798076923077e-06,
|
| 186901 |
+
"loss": 0.2821,
|
| 186902 |
+
"step": 68805
|
| 186903 |
+
},
|
| 186904 |
+
{
|
| 186905 |
+
"epoch": 549.93,
|
| 186906 |
+
"learning_rate": 8.913717948717949e-06,
|
| 186907 |
+
"loss": 0.3476,
|
| 186908 |
+
"step": 68810
|
| 186909 |
+
},
|
| 186910 |
+
{
|
| 186911 |
+
"epoch": 549.97,
|
| 186912 |
+
"learning_rate": 8.913637820512822e-06,
|
| 186913 |
+
"loss": 0.6489,
|
| 186914 |
+
"step": 68815
|
| 186915 |
+
},
|
| 186916 |
+
{
|
| 186917 |
+
"epoch": 550.0,
|
| 186918 |
+
"eval_loss": 0.351525217294693,
|
| 186919 |
+
"eval_runtime": 39.2564,
|
| 186920 |
+
"eval_samples_per_second": 21.347,
|
| 186921 |
+
"eval_steps_per_second": 0.688,
|
| 186922 |
+
"eval_wer": 0.1943950177935943,
|
| 186923 |
+
"step": 68818
|
| 186924 |
+
},
|
| 186925 |
+
{
|
| 186926 |
+
"epoch": 550.02,
|
| 186927 |
+
"learning_rate": 8.913557692307693e-06,
|
| 186928 |
+
"loss": 0.4203,
|
| 186929 |
+
"step": 68820
|
| 186930 |
+
},
|
| 186931 |
+
{
|
| 186932 |
+
"epoch": 550.06,
|
| 186933 |
+
"learning_rate": 8.913477564102564e-06,
|
| 186934 |
+
"loss": 0.3611,
|
| 186935 |
+
"step": 68825
|
| 186936 |
+
},
|
| 186937 |
+
{
|
| 186938 |
+
"epoch": 550.1,
|
| 186939 |
+
"learning_rate": 8.913397435897436e-06,
|
| 186940 |
+
"loss": 0.343,
|
| 186941 |
+
"step": 68830
|
| 186942 |
+
},
|
| 186943 |
+
{
|
| 186944 |
+
"epoch": 550.14,
|
| 186945 |
+
"learning_rate": 8.913317307692309e-06,
|
| 186946 |
+
"loss": 0.3587,
|
| 186947 |
+
"step": 68835
|
| 186948 |
+
},
|
| 186949 |
+
{
|
| 186950 |
+
"epoch": 550.18,
|
| 186951 |
+
"learning_rate": 8.91323717948718e-06,
|
| 186952 |
+
"loss": 0.5963,
|
| 186953 |
+
"step": 68840
|
| 186954 |
+
},
|
| 186955 |
+
{
|
| 186956 |
+
"epoch": 550.22,
|
| 186957 |
+
"learning_rate": 8.913157051282052e-06,
|
| 186958 |
+
"loss": 1.1357,
|
| 186959 |
+
"step": 68845
|
| 186960 |
+
},
|
| 186961 |
+
{
|
| 186962 |
+
"epoch": 550.26,
|
| 186963 |
+
"learning_rate": 8.913076923076925e-06,
|
| 186964 |
+
"loss": 0.2899,
|
| 186965 |
+
"step": 68850
|
| 186966 |
+
},
|
| 186967 |
+
{
|
| 186968 |
+
"epoch": 550.3,
|
| 186969 |
+
"learning_rate": 8.912996794871796e-06,
|
| 186970 |
+
"loss": 0.3763,
|
| 186971 |
+
"step": 68855
|
| 186972 |
+
},
|
| 186973 |
+
{
|
| 186974 |
+
"epoch": 550.34,
|
| 186975 |
+
"learning_rate": 8.912916666666667e-06,
|
| 186976 |
+
"loss": 0.3944,
|
| 186977 |
+
"step": 68860
|
| 186978 |
+
},
|
| 186979 |
+
{
|
| 186980 |
+
"epoch": 550.38,
|
| 186981 |
+
"learning_rate": 8.912836538461539e-06,
|
| 186982 |
+
"loss": 0.7077,
|
| 186983 |
+
"step": 68865
|
| 186984 |
+
},
|
| 186985 |
+
{
|
| 186986 |
+
"epoch": 550.42,
|
| 186987 |
+
"learning_rate": 8.912756410256412e-06,
|
| 186988 |
+
"loss": 1.1302,
|
| 186989 |
+
"step": 68870
|
| 186990 |
+
},
|
| 186991 |
+
{
|
| 186992 |
+
"epoch": 550.46,
|
| 186993 |
+
"learning_rate": 8.912676282051283e-06,
|
| 186994 |
+
"loss": 0.2758,
|
| 186995 |
+
"step": 68875
|
| 186996 |
+
},
|
| 186997 |
+
{
|
| 186998 |
+
"epoch": 550.5,
|
| 186999 |
+
"learning_rate": 8.912596153846155e-06,
|
| 187000 |
+
"loss": 0.3103,
|
| 187001 |
+
"step": 68880
|
| 187002 |
+
},
|
| 187003 |
+
{
|
| 187004 |
+
"epoch": 550.54,
|
| 187005 |
+
"learning_rate": 8.912516025641026e-06,
|
| 187006 |
+
"loss": 0.4578,
|
| 187007 |
+
"step": 68885
|
| 187008 |
+
},
|
| 187009 |
+
{
|
| 187010 |
+
"epoch": 550.58,
|
| 187011 |
+
"learning_rate": 8.912435897435899e-06,
|
| 187012 |
+
"loss": 0.6749,
|
| 187013 |
+
"step": 68890
|
| 187014 |
+
},
|
| 187015 |
+
{
|
| 187016 |
+
"epoch": 550.62,
|
| 187017 |
+
"learning_rate": 8.91235576923077e-06,
|
| 187018 |
+
"loss": 1.0549,
|
| 187019 |
+
"step": 68895
|
| 187020 |
+
},
|
| 187021 |
+
{
|
| 187022 |
+
"epoch": 550.66,
|
| 187023 |
+
"learning_rate": 8.912275641025642e-06,
|
| 187024 |
+
"loss": 0.3558,
|
| 187025 |
+
"step": 68900
|
| 187026 |
+
},
|
| 187027 |
+
{
|
| 187028 |
+
"epoch": 550.7,
|
| 187029 |
+
"learning_rate": 8.912195512820515e-06,
|
| 187030 |
+
"loss": 0.3126,
|
| 187031 |
+
"step": 68905
|
| 187032 |
+
},
|
| 187033 |
+
{
|
| 187034 |
+
"epoch": 550.74,
|
| 187035 |
+
"learning_rate": 8.912115384615384e-06,
|
| 187036 |
+
"loss": 0.4069,
|
| 187037 |
+
"step": 68910
|
| 187038 |
+
},
|
| 187039 |
+
{
|
| 187040 |
+
"epoch": 550.78,
|
| 187041 |
+
"learning_rate": 8.912035256410257e-06,
|
| 187042 |
+
"loss": 0.6448,
|
| 187043 |
+
"step": 68915
|
| 187044 |
+
},
|
| 187045 |
+
{
|
| 187046 |
+
"epoch": 550.82,
|
| 187047 |
+
"learning_rate": 8.911955128205129e-06,
|
| 187048 |
+
"loss": 1.0045,
|
| 187049 |
+
"step": 68920
|
| 187050 |
+
},
|
| 187051 |
+
{
|
| 187052 |
+
"epoch": 550.86,
|
| 187053 |
+
"learning_rate": 8.911875e-06,
|
| 187054 |
+
"loss": 0.3323,
|
| 187055 |
+
"step": 68925
|
| 187056 |
+
},
|
| 187057 |
+
{
|
| 187058 |
+
"epoch": 550.9,
|
| 187059 |
+
"learning_rate": 8.911794871794871e-06,
|
| 187060 |
+
"loss": 0.3236,
|
| 187061 |
+
"step": 68930
|
| 187062 |
+
},
|
| 187063 |
+
{
|
| 187064 |
+
"epoch": 550.94,
|
| 187065 |
+
"learning_rate": 8.911714743589745e-06,
|
| 187066 |
+
"loss": 0.3786,
|
| 187067 |
+
"step": 68935
|
| 187068 |
+
},
|
| 187069 |
+
{
|
| 187070 |
+
"epoch": 550.98,
|
| 187071 |
+
"learning_rate": 8.911634615384616e-06,
|
| 187072 |
+
"loss": 0.6312,
|
| 187073 |
+
"step": 68940
|
| 187074 |
+
},
|
| 187075 |
+
{
|
| 187076 |
+
"epoch": 551.0,
|
| 187077 |
+
"eval_loss": 0.3938581943511963,
|
| 187078 |
+
"eval_runtime": 40.8487,
|
| 187079 |
+
"eval_samples_per_second": 20.515,
|
| 187080 |
+
"eval_steps_per_second": 0.661,
|
| 187081 |
+
"eval_wer": 0.19235865724381626,
|
| 187082 |
+
"step": 68943
|
| 187083 |
+
},
|
| 187084 |
+
{
|
| 187085 |
+
"epoch": 555.02,
|
| 187086 |
+
"learning_rate": 8.911554487179487e-06,
|
| 187087 |
+
"loss": 0.6145,
|
| 187088 |
+
"step": 68945
|
| 187089 |
+
},
|
| 187090 |
+
{
|
| 187091 |
+
"epoch": 555.06,
|
| 187092 |
+
"learning_rate": 8.91147435897436e-06,
|
| 187093 |
+
"loss": 0.3448,
|
| 187094 |
+
"step": 68950
|
| 187095 |
+
},
|
| 187096 |
+
{
|
| 187097 |
+
"epoch": 555.1,
|
| 187098 |
+
"learning_rate": 8.911394230769232e-06,
|
| 187099 |
+
"loss": 0.3279,
|
| 187100 |
+
"step": 68955
|
| 187101 |
+
},
|
| 187102 |
+
{
|
| 187103 |
+
"epoch": 555.14,
|
| 187104 |
+
"learning_rate": 8.911314102564103e-06,
|
| 187105 |
+
"loss": 0.3365,
|
| 187106 |
+
"step": 68960
|
| 187107 |
+
},
|
| 187108 |
+
{
|
| 187109 |
+
"epoch": 555.18,
|
| 187110 |
+
"learning_rate": 8.911233974358974e-06,
|
| 187111 |
+
"loss": 0.7031,
|
| 187112 |
+
"step": 68965
|
| 187113 |
+
},
|
| 187114 |
+
{
|
| 187115 |
+
"epoch": 555.22,
|
| 187116 |
+
"learning_rate": 8.911153846153847e-06,
|
| 187117 |
+
"loss": 1.028,
|
| 187118 |
+
"step": 68970
|
| 187119 |
+
},
|
| 187120 |
+
{
|
| 187121 |
+
"epoch": 555.26,
|
| 187122 |
+
"learning_rate": 8.911073717948719e-06,
|
| 187123 |
+
"loss": 0.3392,
|
| 187124 |
+
"step": 68975
|
| 187125 |
+
},
|
| 187126 |
+
{
|
| 187127 |
+
"epoch": 555.3,
|
| 187128 |
+
"learning_rate": 8.91099358974359e-06,
|
| 187129 |
+
"loss": 0.359,
|
| 187130 |
+
"step": 68980
|
| 187131 |
+
},
|
| 187132 |
+
{
|
| 187133 |
+
"epoch": 555.34,
|
| 187134 |
+
"learning_rate": 8.910913461538462e-06,
|
| 187135 |
+
"loss": 0.3914,
|
| 187136 |
+
"step": 68985
|
| 187137 |
+
},
|
| 187138 |
+
{
|
| 187139 |
+
"epoch": 555.38,
|
| 187140 |
+
"learning_rate": 8.910833333333335e-06,
|
| 187141 |
+
"loss": 0.6924,
|
| 187142 |
+
"step": 68990
|
| 187143 |
+
},
|
| 187144 |
+
{
|
| 187145 |
+
"epoch": 555.42,
|
| 187146 |
+
"learning_rate": 8.910753205128206e-06,
|
| 187147 |
+
"loss": 1.0995,
|
| 187148 |
+
"step": 68995
|
| 187149 |
+
},
|
| 187150 |
+
{
|
| 187151 |
+
"epoch": 555.46,
|
| 187152 |
+
"learning_rate": 8.910673076923077e-06,
|
| 187153 |
+
"loss": 0.3136,
|
| 187154 |
+
"step": 69000
|
| 187155 |
+
},
|
| 187156 |
+
{
|
| 187157 |
+
"epoch": 555.5,
|
| 187158 |
+
"learning_rate": 8.91059294871795e-06,
|
| 187159 |
+
"loss": 0.3497,
|
| 187160 |
+
"step": 69005
|
| 187161 |
+
},
|
| 187162 |
+
{
|
| 187163 |
+
"epoch": 555.54,
|
| 187164 |
+
"learning_rate": 8.910512820512822e-06,
|
| 187165 |
+
"loss": 0.3846,
|
| 187166 |
+
"step": 69010
|
| 187167 |
+
},
|
| 187168 |
+
{
|
| 187169 |
+
"epoch": 555.58,
|
| 187170 |
+
"learning_rate": 8.910432692307693e-06,
|
| 187171 |
+
"loss": 0.6693,
|
| 187172 |
+
"step": 69015
|
| 187173 |
+
},
|
| 187174 |
+
{
|
| 187175 |
+
"epoch": 555.62,
|
| 187176 |
+
"learning_rate": 8.910352564102564e-06,
|
| 187177 |
+
"loss": 1.1715,
|
| 187178 |
+
"step": 69020
|
| 187179 |
+
},
|
| 187180 |
+
{
|
| 187181 |
+
"epoch": 555.66,
|
| 187182 |
+
"learning_rate": 8.910272435897437e-06,
|
| 187183 |
+
"loss": 0.3312,
|
| 187184 |
+
"step": 69025
|
| 187185 |
+
},
|
| 187186 |
+
{
|
| 187187 |
+
"epoch": 555.7,
|
| 187188 |
+
"learning_rate": 8.910192307692309e-06,
|
| 187189 |
+
"loss": 0.373,
|
| 187190 |
+
"step": 69030
|
| 187191 |
+
},
|
| 187192 |
+
{
|
| 187193 |
+
"epoch": 555.74,
|
| 187194 |
+
"learning_rate": 8.91011217948718e-06,
|
| 187195 |
+
"loss": 0.3515,
|
| 187196 |
+
"step": 69035
|
| 187197 |
+
},
|
| 187198 |
+
{
|
| 187199 |
+
"epoch": 555.78,
|
| 187200 |
+
"learning_rate": 8.910032051282052e-06,
|
| 187201 |
+
"loss": 0.6376,
|
| 187202 |
+
"step": 69040
|
| 187203 |
+
},
|
| 187204 |
+
{
|
| 187205 |
+
"epoch": 555.82,
|
| 187206 |
+
"learning_rate": 8.909951923076925e-06,
|
| 187207 |
+
"loss": 1.0635,
|
| 187208 |
+
"step": 69045
|
| 187209 |
+
},
|
| 187210 |
+
{
|
| 187211 |
+
"epoch": 555.86,
|
| 187212 |
+
"learning_rate": 8.909871794871796e-06,
|
| 187213 |
+
"loss": 0.3108,
|
| 187214 |
+
"step": 69050
|
| 187215 |
+
},
|
| 187216 |
+
{
|
| 187217 |
+
"epoch": 555.9,
|
| 187218 |
+
"learning_rate": 8.909791666666667e-06,
|
| 187219 |
+
"loss": 0.3613,
|
| 187220 |
+
"step": 69055
|
| 187221 |
+
},
|
| 187222 |
+
{
|
| 187223 |
+
"epoch": 555.94,
|
| 187224 |
+
"learning_rate": 8.90971153846154e-06,
|
| 187225 |
+
"loss": 0.3515,
|
| 187226 |
+
"step": 69060
|
| 187227 |
+
},
|
| 187228 |
+
{
|
| 187229 |
+
"epoch": 555.98,
|
| 187230 |
+
"learning_rate": 8.90963141025641e-06,
|
| 187231 |
+
"loss": 0.7066,
|
| 187232 |
+
"step": 69065
|
| 187233 |
+
},
|
| 187234 |
+
{
|
| 187235 |
+
"epoch": 556.0,
|
| 187236 |
+
"eval_loss": 0.3735567331314087,
|
| 187237 |
+
"eval_runtime": 39.445,
|
| 187238 |
+
"eval_samples_per_second": 21.27,
|
| 187239 |
+
"eval_steps_per_second": 0.684,
|
| 187240 |
+
"eval_wer": 0.19350029815146094,
|
| 187241 |
+
"step": 69067
|
| 187242 |
+
},
|
| 187243 |
+
{
|
| 187244 |
+
"epoch": 556.02,
|
| 187245 |
+
"learning_rate": 8.909551282051283e-06,
|
| 187246 |
+
"loss": 0.372,
|
| 187247 |
+
"step": 69070
|
| 187248 |
+
},
|
| 187249 |
+
{
|
| 187250 |
+
"epoch": 556.06,
|
| 187251 |
+
"learning_rate": 8.909471153846154e-06,
|
| 187252 |
+
"loss": 0.3013,
|
| 187253 |
+
"step": 69075
|
| 187254 |
+
},
|
| 187255 |
+
{
|
| 187256 |
+
"epoch": 556.1,
|
| 187257 |
+
"learning_rate": 8.909391025641026e-06,
|
| 187258 |
+
"loss": 0.3405,
|
| 187259 |
+
"step": 69080
|
| 187260 |
+
},
|
| 187261 |
+
{
|
| 187262 |
+
"epoch": 556.15,
|
| 187263 |
+
"learning_rate": 8.909310897435897e-06,
|
| 187264 |
+
"loss": 0.4157,
|
| 187265 |
+
"step": 69085
|
| 187266 |
+
},
|
| 187267 |
+
{
|
| 187268 |
+
"epoch": 556.19,
|
| 187269 |
+
"learning_rate": 8.90923076923077e-06,
|
| 187270 |
+
"loss": 0.7011,
|
| 187271 |
+
"step": 69090
|
| 187272 |
+
},
|
| 187273 |
+
{
|
| 187274 |
+
"epoch": 556.23,
|
| 187275 |
+
"learning_rate": 8.909150641025642e-06,
|
| 187276 |
+
"loss": 1.0653,
|
| 187277 |
+
"step": 69095
|
| 187278 |
+
},
|
| 187279 |
+
{
|
| 187280 |
+
"epoch": 556.27,
|
| 187281 |
+
"learning_rate": 8.909070512820513e-06,
|
| 187282 |
+
"loss": 0.3359,
|
| 187283 |
+
"step": 69100
|
| 187284 |
+
},
|
| 187285 |
+
{
|
| 187286 |
+
"epoch": 556.31,
|
| 187287 |
+
"learning_rate": 8.908990384615386e-06,
|
| 187288 |
+
"loss": 0.3593,
|
| 187289 |
+
"step": 69105
|
| 187290 |
+
},
|
| 187291 |
+
{
|
| 187292 |
+
"epoch": 556.35,
|
| 187293 |
+
"learning_rate": 8.908910256410257e-06,
|
| 187294 |
+
"loss": 0.4316,
|
| 187295 |
+
"step": 69110
|
| 187296 |
+
},
|
| 187297 |
+
{
|
| 187298 |
+
"epoch": 556.39,
|
| 187299 |
+
"learning_rate": 8.908830128205129e-06,
|
| 187300 |
+
"loss": 0.748,
|
| 187301 |
+
"step": 69115
|
| 187302 |
+
},
|
| 187303 |
+
{
|
| 187304 |
+
"epoch": 556.43,
|
| 187305 |
+
"learning_rate": 8.90875e-06,
|
| 187306 |
+
"loss": 0.8545,
|
| 187307 |
+
"step": 69120
|
| 187308 |
+
},
|
| 187309 |
+
{
|
| 187310 |
+
"epoch": 556.47,
|
| 187311 |
+
"learning_rate": 8.908669871794873e-06,
|
| 187312 |
+
"loss": 0.3329,
|
| 187313 |
+
"step": 69125
|
| 187314 |
+
},
|
| 187315 |
+
{
|
| 187316 |
+
"epoch": 556.51,
|
| 187317 |
+
"learning_rate": 8.908589743589744e-06,
|
| 187318 |
+
"loss": 0.2987,
|
| 187319 |
+
"step": 69130
|
| 187320 |
+
},
|
| 187321 |
+
{
|
| 187322 |
+
"epoch": 556.55,
|
| 187323 |
+
"learning_rate": 8.908509615384616e-06,
|
| 187324 |
+
"loss": 0.3832,
|
| 187325 |
+
"step": 69135
|
| 187326 |
+
},
|
| 187327 |
+
{
|
| 187328 |
+
"epoch": 556.59,
|
| 187329 |
+
"learning_rate": 8.908429487179487e-06,
|
| 187330 |
+
"loss": 0.8224,
|
| 187331 |
+
"step": 69140
|
| 187332 |
+
},
|
| 187333 |
+
{
|
| 187334 |
+
"epoch": 556.63,
|
| 187335 |
+
"learning_rate": 8.90834935897436e-06,
|
| 187336 |
+
"loss": 0.8839,
|
| 187337 |
+
"step": 69145
|
| 187338 |
+
},
|
| 187339 |
+
{
|
| 187340 |
+
"epoch": 556.67,
|
| 187341 |
+
"learning_rate": 8.908269230769232e-06,
|
| 187342 |
+
"loss": 0.3099,
|
| 187343 |
+
"step": 69150
|
| 187344 |
+
},
|
| 187345 |
+
{
|
| 187346 |
+
"epoch": 556.71,
|
| 187347 |
+
"learning_rate": 8.908189102564103e-06,
|
| 187348 |
+
"loss": 0.2932,
|
| 187349 |
+
"step": 69155
|
| 187350 |
+
},
|
| 187351 |
+
{
|
| 187352 |
+
"epoch": 556.75,
|
| 187353 |
+
"learning_rate": 8.908108974358976e-06,
|
| 187354 |
+
"loss": 0.3511,
|
| 187355 |
+
"step": 69160
|
| 187356 |
+
},
|
| 187357 |
+
{
|
| 187358 |
+
"epoch": 556.79,
|
| 187359 |
+
"learning_rate": 8.908028846153847e-06,
|
| 187360 |
+
"loss": 0.6736,
|
| 187361 |
+
"step": 69165
|
| 187362 |
+
},
|
| 187363 |
+
{
|
| 187364 |
+
"epoch": 556.83,
|
| 187365 |
+
"learning_rate": 8.907948717948719e-06,
|
| 187366 |
+
"loss": 0.8941,
|
| 187367 |
+
"step": 69170
|
| 187368 |
+
},
|
| 187369 |
+
{
|
| 187370 |
+
"epoch": 556.87,
|
| 187371 |
+
"learning_rate": 8.90786858974359e-06,
|
| 187372 |
+
"loss": 0.3152,
|
| 187373 |
+
"step": 69175
|
| 187374 |
+
},
|
| 187375 |
+
{
|
| 187376 |
+
"epoch": 556.91,
|
| 187377 |
+
"learning_rate": 8.907788461538463e-06,
|
| 187378 |
+
"loss": 0.355,
|
| 187379 |
+
"step": 69180
|
| 187380 |
+
},
|
| 187381 |
+
{
|
| 187382 |
+
"epoch": 556.95,
|
| 187383 |
+
"learning_rate": 8.907708333333333e-06,
|
| 187384 |
+
"loss": 0.4686,
|
| 187385 |
+
"step": 69185
|
| 187386 |
+
},
|
| 187387 |
+
{
|
| 187388 |
+
"epoch": 556.99,
|
| 187389 |
+
"learning_rate": 8.907628205128206e-06,
|
| 187390 |
+
"loss": 0.8468,
|
| 187391 |
+
"step": 69190
|
| 187392 |
+
},
|
| 187393 |
+
{
|
| 187394 |
+
"epoch": 557.0,
|
| 187395 |
+
"eval_loss": 0.3936476707458496,
|
| 187396 |
+
"eval_runtime": 39.1151,
|
| 187397 |
+
"eval_samples_per_second": 21.45,
|
| 187398 |
+
"eval_steps_per_second": 0.69,
|
| 187399 |
+
"eval_wer": 0.19881948595337717,
|
| 187400 |
+
"step": 69191
|
| 187401 |
+
},
|
| 187402 |
+
{
|
| 187403 |
+
"epoch": 557.03,
|
| 187404 |
+
"learning_rate": 8.907548076923077e-06,
|
| 187405 |
+
"loss": 0.3576,
|
| 187406 |
+
"step": 69195
|
| 187407 |
+
},
|
| 187408 |
+
{
|
| 187409 |
+
"epoch": 557.07,
|
| 187410 |
+
"learning_rate": 8.907467948717949e-06,
|
| 187411 |
+
"loss": 0.2662,
|
| 187412 |
+
"step": 69200
|
| 187413 |
+
},
|
| 187414 |
+
{
|
| 187415 |
+
"epoch": 557.11,
|
| 187416 |
+
"learning_rate": 8.907387820512822e-06,
|
| 187417 |
+
"loss": 0.3339,
|
| 187418 |
+
"step": 69205
|
| 187419 |
+
},
|
| 187420 |
+
{
|
| 187421 |
+
"epoch": 557.15,
|
| 187422 |
+
"learning_rate": 8.907307692307693e-06,
|
| 187423 |
+
"loss": 0.4627,
|
| 187424 |
+
"step": 69210
|
| 187425 |
+
},
|
| 187426 |
+
{
|
| 187427 |
+
"epoch": 557.19,
|
| 187428 |
+
"learning_rate": 8.907227564102564e-06,
|
| 187429 |
+
"loss": 1.0684,
|
| 187430 |
+
"step": 69215
|
| 187431 |
+
},
|
| 187432 |
+
{
|
| 187433 |
+
"epoch": 557.23,
|
| 187434 |
+
"learning_rate": 8.907147435897436e-06,
|
| 187435 |
+
"loss": 0.7947,
|
| 187436 |
+
"step": 69220
|
| 187437 |
+
},
|
| 187438 |
+
{
|
| 187439 |
+
"epoch": 557.27,
|
| 187440 |
+
"learning_rate": 8.907067307692309e-06,
|
| 187441 |
+
"loss": 0.3128,
|
| 187442 |
+
"step": 69225
|
| 187443 |
+
},
|
| 187444 |
+
{
|
| 187445 |
+
"epoch": 557.31,
|
| 187446 |
+
"learning_rate": 8.90698717948718e-06,
|
| 187447 |
+
"loss": 0.3003,
|
| 187448 |
+
"step": 69230
|
| 187449 |
+
},
|
| 187450 |
+
{
|
| 187451 |
+
"epoch": 557.35,
|
| 187452 |
+
"learning_rate": 8.906907051282051e-06,
|
| 187453 |
+
"loss": 0.4299,
|
| 187454 |
+
"step": 69235
|
| 187455 |
+
},
|
| 187456 |
+
{
|
| 187457 |
+
"epoch": 557.39,
|
| 187458 |
+
"learning_rate": 8.906826923076923e-06,
|
| 187459 |
+
"loss": 1.0473,
|
| 187460 |
+
"step": 69240
|
| 187461 |
+
},
|
| 187462 |
+
{
|
| 187463 |
+
"epoch": 557.43,
|
| 187464 |
+
"learning_rate": 8.906746794871796e-06,
|
| 187465 |
+
"loss": 0.5692,
|
| 187466 |
+
"step": 69245
|
| 187467 |
+
},
|
| 187468 |
+
{
|
| 187469 |
+
"epoch": 557.47,
|
| 187470 |
+
"learning_rate": 8.906666666666667e-06,
|
| 187471 |
+
"loss": 0.2719,
|
| 187472 |
+
"step": 69250
|
| 187473 |
+
},
|
| 187474 |
+
{
|
| 187475 |
+
"epoch": 557.51,
|
| 187476 |
+
"learning_rate": 8.906586538461539e-06,
|
| 187477 |
+
"loss": 0.3171,
|
| 187478 |
+
"step": 69255
|
| 187479 |
+
},
|
| 187480 |
+
{
|
| 187481 |
+
"epoch": 557.55,
|
| 187482 |
+
"learning_rate": 8.906506410256412e-06,
|
| 187483 |
+
"loss": 0.4372,
|
| 187484 |
+
"step": 69260
|
| 187485 |
+
},
|
| 187486 |
+
{
|
| 187487 |
+
"epoch": 557.59,
|
| 187488 |
+
"learning_rate": 8.906426282051283e-06,
|
| 187489 |
+
"loss": 0.9317,
|
| 187490 |
+
"step": 69265
|
| 187491 |
+
},
|
| 187492 |
+
{
|
| 187493 |
+
"epoch": 557.63,
|
| 187494 |
+
"learning_rate": 8.906346153846154e-06,
|
| 187495 |
+
"loss": 0.6513,
|
| 187496 |
+
"step": 69270
|
| 187497 |
+
},
|
| 187498 |
+
{
|
| 187499 |
+
"epoch": 557.67,
|
| 187500 |
+
"learning_rate": 8.906266025641026e-06,
|
| 187501 |
+
"loss": 0.272,
|
| 187502 |
+
"step": 69275
|
| 187503 |
+
},
|
| 187504 |
+
{
|
| 187505 |
+
"epoch": 557.71,
|
| 187506 |
+
"learning_rate": 8.906185897435899e-06,
|
| 187507 |
+
"loss": 0.3266,
|
| 187508 |
+
"step": 69280
|
| 187509 |
+
},
|
| 187510 |
+
{
|
| 187511 |
+
"epoch": 557.76,
|
| 187512 |
+
"learning_rate": 8.90610576923077e-06,
|
| 187513 |
+
"loss": 0.453,
|
| 187514 |
+
"step": 69285
|
| 187515 |
+
},
|
| 187516 |
+
{
|
| 187517 |
+
"epoch": 557.8,
|
| 187518 |
+
"learning_rate": 8.906025641025641e-06,
|
| 187519 |
+
"loss": 0.8466,
|
| 187520 |
+
"step": 69290
|
| 187521 |
+
},
|
| 187522 |
+
{
|
| 187523 |
+
"epoch": 557.84,
|
| 187524 |
+
"learning_rate": 8.905945512820513e-06,
|
| 187525 |
+
"loss": 0.6527,
|
| 187526 |
+
"step": 69295
|
| 187527 |
+
},
|
| 187528 |
+
{
|
| 187529 |
+
"epoch": 557.88,
|
| 187530 |
+
"learning_rate": 8.905865384615386e-06,
|
| 187531 |
+
"loss": 0.3322,
|
| 187532 |
+
"step": 69300
|
| 187533 |
+
},
|
| 187534 |
+
{
|
| 187535 |
+
"epoch": 557.92,
|
| 187536 |
+
"learning_rate": 8.905785256410257e-06,
|
| 187537 |
+
"loss": 0.3628,
|
| 187538 |
+
"step": 69305
|
| 187539 |
+
},
|
| 187540 |
+
{
|
| 187541 |
+
"epoch": 557.96,
|
| 187542 |
+
"learning_rate": 8.905705128205129e-06,
|
| 187543 |
+
"loss": 0.4183,
|
| 187544 |
+
"step": 69310
|
| 187545 |
+
},
|
| 187546 |
+
{
|
| 187547 |
+
"epoch": 558.0,
|
| 187548 |
+
"learning_rate": 8.905625000000002e-06,
|
| 187549 |
+
"loss": 1.0645,
|
| 187550 |
+
"step": 69315
|
| 187551 |
+
},
|
| 187552 |
+
{
|
| 187553 |
+
"epoch": 558.0,
|
| 187554 |
+
"eval_loss": 0.4768179655075073,
|
| 187555 |
+
"eval_runtime": 41.3535,
|
| 187556 |
+
"eval_samples_per_second": 20.289,
|
| 187557 |
+
"eval_steps_per_second": 0.653,
|
| 187558 |
+
"eval_wer": 0.1948916076411247,
|
| 187559 |
+
"step": 69315
|
| 187560 |
+
},
|
| 187561 |
+
{
|
| 187562 |
+
"epoch": 554.04,
|
| 187563 |
+
"learning_rate": 8.905544871794873e-06,
|
| 187564 |
+
"loss": 0.3339,
|
| 187565 |
+
"step": 69320
|
| 187566 |
+
},
|
| 187567 |
+
{
|
| 187568 |
+
"epoch": 554.08,
|
| 187569 |
+
"learning_rate": 8.905464743589744e-06,
|
| 187570 |
+
"loss": 0.2938,
|
| 187571 |
+
"step": 69325
|
| 187572 |
+
},
|
| 187573 |
+
{
|
| 187574 |
+
"epoch": 554.12,
|
| 187575 |
+
"learning_rate": 8.905384615384616e-06,
|
| 187576 |
+
"loss": 0.3485,
|
| 187577 |
+
"step": 69330
|
| 187578 |
+
},
|
| 187579 |
+
{
|
| 187580 |
+
"epoch": 554.16,
|
| 187581 |
+
"learning_rate": 8.905304487179489e-06,
|
| 187582 |
+
"loss": 0.6218,
|
| 187583 |
+
"step": 69335
|
| 187584 |
+
},
|
| 187585 |
+
{
|
| 187586 |
+
"epoch": 554.2,
|
| 187587 |
+
"learning_rate": 8.905224358974358e-06,
|
| 187588 |
+
"loss": 1.2255,
|
| 187589 |
+
"step": 69340
|
| 187590 |
+
},
|
| 187591 |
+
{
|
| 187592 |
+
"epoch": 554.24,
|
| 187593 |
+
"learning_rate": 8.905144230769232e-06,
|
| 187594 |
+
"loss": 0.3385,
|
| 187595 |
+
"step": 69345
|
| 187596 |
+
},
|
| 187597 |
+
{
|
| 187598 |
+
"epoch": 554.28,
|
| 187599 |
+
"learning_rate": 8.905064102564105e-06,
|
| 187600 |
+
"loss": 0.3168,
|
| 187601 |
+
"step": 69350
|
| 187602 |
+
},
|
| 187603 |
+
{
|
| 187604 |
+
"epoch": 554.32,
|
| 187605 |
+
"learning_rate": 8.904983974358974e-06,
|
| 187606 |
+
"loss": 0.3309,
|
| 187607 |
+
"step": 69355
|
| 187608 |
+
},
|
| 187609 |
+
{
|
| 187610 |
+
"epoch": 554.36,
|
| 187611 |
+
"learning_rate": 8.904903846153847e-06,
|
| 187612 |
+
"loss": 0.5233,
|
| 187613 |
+
"step": 69360
|
| 187614 |
+
},
|
| 187615 |
+
{
|
| 187616 |
+
"epoch": 554.4,
|
| 187617 |
+
"learning_rate": 8.904823717948719e-06,
|
| 187618 |
+
"loss": 1.4065,
|
| 187619 |
+
"step": 69365
|
| 187620 |
+
},
|
| 187621 |
+
{
|
| 187622 |
+
"epoch": 554.44,
|
| 187623 |
+
"learning_rate": 8.90474358974359e-06,
|
| 187624 |
+
"loss": 0.3589,
|
| 187625 |
+
"step": 69370
|
| 187626 |
+
},
|
| 187627 |
+
{
|
| 187628 |
+
"epoch": 554.48,
|
| 187629 |
+
"learning_rate": 8.904663461538461e-06,
|
| 187630 |
+
"loss": 0.2869,
|
| 187631 |
+
"step": 69375
|
| 187632 |
+
},
|
| 187633 |
+
{
|
| 187634 |
+
"epoch": 554.52,
|
| 187635 |
+
"learning_rate": 8.904583333333334e-06,
|
| 187636 |
+
"loss": 0.3377,
|
| 187637 |
+
"step": 69380
|
| 187638 |
+
},
|
| 187639 |
+
{
|
| 187640 |
+
"epoch": 554.56,
|
| 187641 |
+
"learning_rate": 8.904503205128206e-06,
|
| 187642 |
+
"loss": 0.535,
|
| 187643 |
+
"step": 69385
|
| 187644 |
+
},
|
| 187645 |
+
{
|
| 187646 |
+
"epoch": 554.6,
|
| 187647 |
+
"learning_rate": 8.904423076923077e-06,
|
| 187648 |
+
"loss": 1.4467,
|
| 187649 |
+
"step": 69390
|
| 187650 |
+
},
|
| 187651 |
+
{
|
| 187652 |
+
"epoch": 554.64,
|
| 187653 |
+
"learning_rate": 8.904342948717948e-06,
|
| 187654 |
+
"loss": 0.5932,
|
| 187655 |
+
"step": 69395
|
| 187656 |
+
},
|
| 187657 |
+
{
|
| 187658 |
+
"epoch": 554.68,
|
| 187659 |
+
"learning_rate": 8.904262820512822e-06,
|
| 187660 |
+
"loss": 0.4281,
|
| 187661 |
+
"step": 69400
|
| 187662 |
+
},
|
| 187663 |
+
{
|
| 187664 |
+
"epoch": 554.72,
|
| 187665 |
+
"learning_rate": 8.904182692307693e-06,
|
| 187666 |
+
"loss": 0.3255,
|
| 187667 |
+
"step": 69405
|
| 187668 |
+
},
|
| 187669 |
+
{
|
| 187670 |
+
"epoch": 554.76,
|
| 187671 |
+
"learning_rate": 8.904102564102564e-06,
|
| 187672 |
+
"loss": 0.5435,
|
| 187673 |
+
"step": 69410
|
| 187674 |
+
},
|
| 187675 |
+
{
|
| 187676 |
+
"epoch": 554.8,
|
| 187677 |
+
"learning_rate": 8.904022435897437e-06,
|
| 187678 |
+
"loss": 1.2416,
|
| 187679 |
+
"step": 69415
|
| 187680 |
+
},
|
| 187681 |
+
{
|
| 187682 |
+
"epoch": 554.84,
|
| 187683 |
+
"learning_rate": 8.903942307692309e-06,
|
| 187684 |
+
"loss": 0.6299,
|
| 187685 |
+
"step": 69420
|
| 187686 |
+
},
|
| 187687 |
+
{
|
| 187688 |
+
"epoch": 554.88,
|
| 187689 |
+
"learning_rate": 8.90386217948718e-06,
|
| 187690 |
+
"loss": 0.3255,
|
| 187691 |
+
"step": 69425
|
| 187692 |
+
},
|
| 187693 |
+
{
|
| 187694 |
+
"epoch": 554.92,
|
| 187695 |
+
"learning_rate": 8.903782051282051e-06,
|
| 187696 |
+
"loss": 0.3651,
|
| 187697 |
+
"step": 69430
|
| 187698 |
+
},
|
| 187699 |
+
{
|
| 187700 |
+
"epoch": 554.96,
|
| 187701 |
+
"learning_rate": 8.903701923076924e-06,
|
| 187702 |
+
"loss": 0.4798,
|
| 187703 |
+
"step": 69435
|
| 187704 |
+
},
|
| 187705 |
+
{
|
| 187706 |
+
"epoch": 555.0,
|
| 187707 |
+
"learning_rate": 8.903621794871796e-06,
|
| 187708 |
+
"loss": 1.206,
|
| 187709 |
+
"step": 69440
|
| 187710 |
+
},
|
| 187711 |
+
{
|
| 187712 |
+
"epoch": 555.0,
|
| 187713 |
+
"eval_loss": 0.34815290570259094,
|
| 187714 |
+
"eval_runtime": 40.671,
|
| 187715 |
+
"eval_samples_per_second": 20.629,
|
| 187716 |
+
"eval_steps_per_second": 0.664,
|
| 187717 |
+
"eval_wer": 0.18589017831870933,
|
| 187718 |
+
"step": 69440
|
| 187719 |
+
},
|
| 187720 |
+
{
|
| 187721 |
+
"epoch": 555.04,
|
| 187722 |
+
"learning_rate": 8.903541666666667e-06,
|
| 187723 |
+
"loss": 0.376,
|
| 187724 |
+
"step": 69445
|
| 187725 |
+
},
|
| 187726 |
+
{
|
| 187727 |
+
"epoch": 555.08,
|
| 187728 |
+
"learning_rate": 8.90346153846154e-06,
|
| 187729 |
+
"loss": 0.2794,
|
| 187730 |
+
"step": 69450
|
| 187731 |
+
},
|
| 187732 |
+
{
|
| 187733 |
+
"epoch": 555.12,
|
| 187734 |
+
"learning_rate": 8.903381410256412e-06,
|
| 187735 |
+
"loss": 0.4327,
|
| 187736 |
+
"step": 69455
|
| 187737 |
+
},
|
| 187738 |
+
{
|
| 187739 |
+
"epoch": 555.16,
|
| 187740 |
+
"learning_rate": 8.903301282051283e-06,
|
| 187741 |
+
"loss": 0.6141,
|
| 187742 |
+
"step": 69460
|
| 187743 |
+
},
|
| 187744 |
+
{
|
| 187745 |
+
"epoch": 555.2,
|
| 187746 |
+
"learning_rate": 8.903221153846154e-06,
|
| 187747 |
+
"loss": 1.2486,
|
| 187748 |
+
"step": 69465
|
| 187749 |
+
},
|
| 187750 |
+
{
|
| 187751 |
+
"epoch": 555.24,
|
| 187752 |
+
"learning_rate": 8.903141025641027e-06,
|
| 187753 |
+
"loss": 0.5636,
|
| 187754 |
+
"step": 69470
|
| 187755 |
+
},
|
| 187756 |
+
{
|
| 187757 |
+
"epoch": 555.28,
|
| 187758 |
+
"learning_rate": 8.903060897435899e-06,
|
| 187759 |
+
"loss": 0.2681,
|
| 187760 |
+
"step": 69475
|
| 187761 |
+
},
|
| 187762 |
+
{
|
| 187763 |
+
"epoch": 555.32,
|
| 187764 |
+
"learning_rate": 8.90298076923077e-06,
|
| 187765 |
+
"loss": 0.3946,
|
| 187766 |
+
"step": 69480
|
| 187767 |
+
},
|
| 187768 |
+
{
|
| 187769 |
+
"epoch": 555.36,
|
| 187770 |
+
"learning_rate": 8.902900641025641e-06,
|
| 187771 |
+
"loss": 0.4832,
|
| 187772 |
+
"step": 69485
|
| 187773 |
+
},
|
| 187774 |
+
{
|
| 187775 |
+
"epoch": 555.4,
|
| 187776 |
+
"learning_rate": 8.902820512820514e-06,
|
| 187777 |
+
"loss": 1.2986,
|
| 187778 |
+
"step": 69490
|
| 187779 |
+
},
|
| 187780 |
+
{
|
| 187781 |
+
"epoch": 555.44,
|
| 187782 |
+
"learning_rate": 8.902740384615384e-06,
|
| 187783 |
+
"loss": 0.6412,
|
| 187784 |
+
"step": 69495
|
| 187785 |
+
},
|
| 187786 |
+
{
|
| 187787 |
+
"epoch": 555.48,
|
| 187788 |
+
"learning_rate": 8.902660256410257e-06,
|
| 187789 |
+
"loss": 0.3092,
|
| 187790 |
+
"step": 69500
|
| 187791 |
+
},
|
| 187792 |
+
{
|
| 187793 |
+
"epoch": 555.52,
|
| 187794 |
+
"learning_rate": 8.90258012820513e-06,
|
| 187795 |
+
"loss": 0.3213,
|
| 187796 |
+
"step": 69505
|
| 187797 |
+
},
|
| 187798 |
+
{
|
| 187799 |
+
"epoch": 555.56,
|
| 187800 |
+
"learning_rate": 8.9025e-06,
|
| 187801 |
+
"loss": 0.5671,
|
| 187802 |
+
"step": 69510
|
| 187803 |
+
},
|
| 187804 |
+
{
|
| 187805 |
+
"epoch": 555.6,
|
| 187806 |
+
"learning_rate": 8.902419871794873e-06,
|
| 187807 |
+
"loss": 1.5722,
|
| 187808 |
+
"step": 69515
|
| 187809 |
+
},
|
| 187810 |
+
{
|
| 187811 |
+
"epoch": 555.64,
|
| 187812 |
+
"learning_rate": 8.902339743589744e-06,
|
| 187813 |
+
"loss": 0.3612,
|
| 187814 |
+
"step": 69520
|
| 187815 |
+
},
|
| 187816 |
+
{
|
| 187817 |
+
"epoch": 555.68,
|
| 187818 |
+
"learning_rate": 8.902259615384616e-06,
|
| 187819 |
+
"loss": 0.3524,
|
| 187820 |
+
"step": 69525
|
| 187821 |
+
},
|
| 187822 |
+
{
|
| 187823 |
+
"epoch": 555.72,
|
| 187824 |
+
"learning_rate": 8.902179487179487e-06,
|
| 187825 |
+
"loss": 0.3638,
|
| 187826 |
+
"step": 69530
|
| 187827 |
+
},
|
| 187828 |
+
{
|
| 187829 |
+
"epoch": 555.76,
|
| 187830 |
+
"learning_rate": 8.90209935897436e-06,
|
| 187831 |
+
"loss": 0.5201,
|
| 187832 |
+
"step": 69535
|
| 187833 |
+
},
|
| 187834 |
+
{
|
| 187835 |
+
"epoch": 555.8,
|
| 187836 |
+
"learning_rate": 8.902019230769231e-06,
|
| 187837 |
+
"loss": 1.2726,
|
| 187838 |
+
"step": 69540
|
| 187839 |
+
},
|
| 187840 |
+
{
|
| 187841 |
+
"epoch": 555.84,
|
| 187842 |
+
"learning_rate": 8.901939102564103e-06,
|
| 187843 |
+
"loss": 0.3403,
|
| 187844 |
+
"step": 69545
|
| 187845 |
+
},
|
| 187846 |
+
{
|
| 187847 |
+
"epoch": 555.88,
|
| 187848 |
+
"learning_rate": 8.901858974358976e-06,
|
| 187849 |
+
"loss": 0.364,
|
| 187850 |
+
"step": 69550
|
| 187851 |
+
},
|
| 187852 |
+
{
|
| 187853 |
+
"epoch": 555.92,
|
| 187854 |
+
"learning_rate": 8.901778846153847e-06,
|
| 187855 |
+
"loss": 0.3376,
|
| 187856 |
+
"step": 69555
|
| 187857 |
+
},
|
| 187858 |
+
{
|
| 187859 |
+
"epoch": 555.96,
|
| 187860 |
+
"learning_rate": 8.901698717948719e-06,
|
| 187861 |
+
"loss": 0.504,
|
| 187862 |
+
"step": 69560
|
| 187863 |
+
},
|
| 187864 |
+
{
|
| 187865 |
+
"epoch": 556.0,
|
| 187866 |
+
"learning_rate": 8.90161858974359e-06,
|
| 187867 |
+
"loss": 1.2661,
|
| 187868 |
+
"step": 69565
|
| 187869 |
+
},
|
| 187870 |
+
{
|
| 187871 |
+
"epoch": 556.0,
|
| 187872 |
+
"eval_loss": 0.34267231822013855,
|
| 187873 |
+
"eval_runtime": 39.6183,
|
| 187874 |
+
"eval_samples_per_second": 21.177,
|
| 187875 |
+
"eval_steps_per_second": 0.682,
|
| 187876 |
+
"eval_wer": 0.18412114350410416,
|
| 187877 |
+
"step": 69565
|
| 187878 |
+
},
|
| 187879 |
+
{
|
| 187880 |
+
"epoch": 561.04,
|
| 187881 |
+
"learning_rate": 8.901538461538463e-06,
|
| 187882 |
+
"loss": 0.347,
|
| 187883 |
+
"step": 69570
|
| 187884 |
+
},
|
| 187885 |
+
{
|
| 187886 |
+
"epoch": 561.08,
|
| 187887 |
+
"learning_rate": 8.901458333333334e-06,
|
| 187888 |
+
"loss": 0.3164,
|
| 187889 |
+
"step": 69575
|
| 187890 |
+
},
|
| 187891 |
+
{
|
| 187892 |
+
"epoch": 561.12,
|
| 187893 |
+
"learning_rate": 8.901378205128206e-06,
|
| 187894 |
+
"loss": 0.3219,
|
| 187895 |
+
"step": 69580
|
| 187896 |
+
},
|
| 187897 |
+
{
|
| 187898 |
+
"epoch": 561.16,
|
| 187899 |
+
"learning_rate": 8.901298076923077e-06,
|
| 187900 |
+
"loss": 0.5492,
|
| 187901 |
+
"step": 69585
|
| 187902 |
+
},
|
| 187903 |
+
{
|
| 187904 |
+
"epoch": 561.2,
|
| 187905 |
+
"learning_rate": 8.90121794871795e-06,
|
| 187906 |
+
"loss": 1.1423,
|
| 187907 |
+
"step": 69590
|
| 187908 |
+
},
|
| 187909 |
+
{
|
| 187910 |
+
"epoch": 561.24,
|
| 187911 |
+
"learning_rate": 8.901137820512821e-06,
|
| 187912 |
+
"loss": 0.3753,
|
| 187913 |
+
"step": 69595
|
| 187914 |
+
},
|
| 187915 |
+
{
|
| 187916 |
+
"epoch": 561.28,
|
| 187917 |
+
"learning_rate": 8.901057692307693e-06,
|
| 187918 |
+
"loss": 0.3353,
|
| 187919 |
+
"step": 69600
|
| 187920 |
+
},
|
| 187921 |
+
{
|
| 187922 |
+
"epoch": 561.32,
|
| 187923 |
+
"learning_rate": 8.900977564102566e-06,
|
| 187924 |
+
"loss": 0.3504,
|
| 187925 |
+
"step": 69605
|
| 187926 |
+
},
|
| 187927 |
+
{
|
| 187928 |
+
"epoch": 561.36,
|
| 187929 |
+
"learning_rate": 8.900897435897437e-06,
|
| 187930 |
+
"loss": 0.582,
|
| 187931 |
+
"step": 69610
|
| 187932 |
+
},
|
| 187933 |
+
{
|
| 187934 |
+
"epoch": 561.4,
|
| 187935 |
+
"learning_rate": 8.900817307692309e-06,
|
| 187936 |
+
"loss": 1.2468,
|
| 187937 |
+
"step": 69615
|
| 187938 |
+
},
|
| 187939 |
+
{
|
| 187940 |
+
"epoch": 561.44,
|
| 187941 |
+
"learning_rate": 8.90073717948718e-06,
|
| 187942 |
+
"loss": 0.3038,
|
| 187943 |
+
"step": 69620
|
| 187944 |
+
},
|
| 187945 |
+
{
|
| 187946 |
+
"epoch": 561.48,
|
| 187947 |
+
"learning_rate": 8.900657051282053e-06,
|
| 187948 |
+
"loss": 0.3096,
|
| 187949 |
+
"step": 69625
|
| 187950 |
+
},
|
| 187951 |
+
{
|
| 187952 |
+
"epoch": 561.52,
|
| 187953 |
+
"learning_rate": 8.900576923076923e-06,
|
| 187954 |
+
"loss": 0.3303,
|
| 187955 |
+
"step": 69630
|
| 187956 |
+
},
|
| 187957 |
+
{
|
| 187958 |
+
"epoch": 561.56,
|
| 187959 |
+
"learning_rate": 8.900496794871796e-06,
|
| 187960 |
+
"loss": 0.4793,
|
| 187961 |
+
"step": 69635
|
| 187962 |
+
},
|
| 187963 |
+
{
|
| 187964 |
+
"epoch": 561.6,
|
| 187965 |
+
"learning_rate": 8.900416666666667e-06,
|
| 187966 |
+
"loss": 1.2255,
|
| 187967 |
+
"step": 69640
|
| 187968 |
+
},
|
| 187969 |
+
{
|
| 187970 |
+
"epoch": 561.64,
|
| 187971 |
+
"learning_rate": 8.900336538461538e-06,
|
| 187972 |
+
"loss": 0.3851,
|
| 187973 |
+
"step": 69645
|
| 187974 |
+
},
|
| 187975 |
+
{
|
| 187976 |
+
"epoch": 561.68,
|
| 187977 |
+
"learning_rate": 8.900256410256411e-06,
|
| 187978 |
+
"loss": 0.3737,
|
| 187979 |
+
"step": 69650
|
| 187980 |
+
},
|
| 187981 |
+
{
|
| 187982 |
+
"epoch": 561.72,
|
| 187983 |
+
"learning_rate": 8.900176282051283e-06,
|
| 187984 |
+
"loss": 0.343,
|
| 187985 |
+
"step": 69655
|
| 187986 |
+
},
|
| 187987 |
+
{
|
| 187988 |
+
"epoch": 561.76,
|
| 187989 |
+
"learning_rate": 8.900096153846154e-06,
|
| 187990 |
+
"loss": 0.4715,
|
| 187991 |
+
"step": 69660
|
| 187992 |
+
},
|
| 187993 |
+
{
|
| 187994 |
+
"epoch": 561.8,
|
| 187995 |
+
"learning_rate": 8.900016025641026e-06,
|
| 187996 |
+
"loss": 1.2698,
|
| 187997 |
+
"step": 69665
|
| 187998 |
+
},
|
| 187999 |
+
{
|
| 188000 |
+
"epoch": 561.84,
|
| 188001 |
+
"learning_rate": 8.899935897435899e-06,
|
| 188002 |
+
"loss": 0.3687,
|
| 188003 |
+
"step": 69670
|
| 188004 |
+
},
|
| 188005 |
+
{
|
| 188006 |
+
"epoch": 561.88,
|
| 188007 |
+
"learning_rate": 8.89985576923077e-06,
|
| 188008 |
+
"loss": 0.2958,
|
| 188009 |
+
"step": 69675
|
| 188010 |
+
},
|
| 188011 |
+
{
|
| 188012 |
+
"epoch": 561.92,
|
| 188013 |
+
"learning_rate": 8.899775641025641e-06,
|
| 188014 |
+
"loss": 0.365,
|
| 188015 |
+
"step": 69680
|
| 188016 |
+
},
|
| 188017 |
+
{
|
| 188018 |
+
"epoch": 561.96,
|
| 188019 |
+
"learning_rate": 8.899695512820513e-06,
|
| 188020 |
+
"loss": 0.5896,
|
| 188021 |
+
"step": 69685
|
| 188022 |
+
},
|
| 188023 |
+
{
|
| 188024 |
+
"epoch": 562.0,
|
| 188025 |
+
"eval_loss": 0.4862026274204254,
|
| 188026 |
+
"eval_runtime": 40.2027,
|
| 188027 |
+
"eval_samples_per_second": 20.869,
|
| 188028 |
+
"eval_steps_per_second": 0.672,
|
| 188029 |
+
"eval_wer": 0.19260025108928439,
|
| 188030 |
+
"step": 69689
|
| 188031 |
+
},
|
| 188032 |
+
{
|
| 188033 |
+
"epoch": 557.01,
|
| 188034 |
+
"learning_rate": 8.899615384615386e-06,
|
| 188035 |
+
"loss": 0.3987,
|
| 188036 |
+
"step": 69690
|
| 188037 |
+
},
|
| 188038 |
+
{
|
| 188039 |
+
"epoch": 557.05,
|
| 188040 |
+
"learning_rate": 8.899535256410257e-06,
|
| 188041 |
+
"loss": 0.3314,
|
| 188042 |
+
"step": 69695
|
| 188043 |
+
},
|
| 188044 |
+
{
|
| 188045 |
+
"epoch": 557.09,
|
| 188046 |
+
"learning_rate": 8.899455128205128e-06,
|
| 188047 |
+
"loss": 0.3273,
|
| 188048 |
+
"step": 69700
|
| 188049 |
+
},
|
| 188050 |
+
{
|
| 188051 |
+
"epoch": 557.13,
|
| 188052 |
+
"learning_rate": 8.899375000000002e-06,
|
| 188053 |
+
"loss": 0.3822,
|
| 188054 |
+
"step": 69705
|
| 188055 |
+
},
|
| 188056 |
+
{
|
| 188057 |
+
"epoch": 557.17,
|
| 188058 |
+
"learning_rate": 8.899294871794873e-06,
|
| 188059 |
+
"loss": 0.5403,
|
| 188060 |
+
"step": 69710
|
| 188061 |
+
},
|
| 188062 |
+
{
|
| 188063 |
+
"epoch": 557.21,
|
| 188064 |
+
"learning_rate": 8.899214743589744e-06,
|
| 188065 |
+
"loss": 1.1595,
|
| 188066 |
+
"step": 69715
|
| 188067 |
+
},
|
| 188068 |
+
{
|
| 188069 |
+
"epoch": 557.25,
|
| 188070 |
+
"learning_rate": 8.899134615384616e-06,
|
| 188071 |
+
"loss": 0.3848,
|
| 188072 |
+
"step": 69720
|
| 188073 |
+
},
|
| 188074 |
+
{
|
| 188075 |
+
"epoch": 557.29,
|
| 188076 |
+
"learning_rate": 8.899054487179489e-06,
|
| 188077 |
+
"loss": 0.3017,
|
| 188078 |
+
"step": 69725
|
| 188079 |
+
},
|
| 188080 |
+
{
|
| 188081 |
+
"epoch": 557.33,
|
| 188082 |
+
"learning_rate": 8.89897435897436e-06,
|
| 188083 |
+
"loss": 0.3478,
|
| 188084 |
+
"step": 69730
|
| 188085 |
+
},
|
| 188086 |
+
{
|
| 188087 |
+
"epoch": 557.37,
|
| 188088 |
+
"learning_rate": 8.898894230769231e-06,
|
| 188089 |
+
"loss": 0.6596,
|
| 188090 |
+
"step": 69735
|
| 188091 |
+
},
|
| 188092 |
+
{
|
| 188093 |
+
"epoch": 557.41,
|
| 188094 |
+
"learning_rate": 8.898814102564103e-06,
|
| 188095 |
+
"loss": 1.1257,
|
| 188096 |
+
"step": 69740
|
| 188097 |
+
},
|
| 188098 |
+
{
|
| 188099 |
+
"epoch": 557.45,
|
| 188100 |
+
"learning_rate": 8.898733974358976e-06,
|
| 188101 |
+
"loss": 0.3708,
|
| 188102 |
+
"step": 69745
|
| 188103 |
+
},
|
| 188104 |
+
{
|
| 188105 |
+
"epoch": 557.49,
|
| 188106 |
+
"learning_rate": 8.898653846153847e-06,
|
| 188107 |
+
"loss": 0.2876,
|
| 188108 |
+
"step": 69750
|
| 188109 |
+
},
|
| 188110 |
+
{
|
| 188111 |
+
"epoch": 557.53,
|
| 188112 |
+
"learning_rate": 8.898573717948718e-06,
|
| 188113 |
+
"loss": 0.3677,
|
| 188114 |
+
"step": 69755
|
| 188115 |
+
},
|
| 188116 |
+
{
|
| 188117 |
+
"epoch": 557.57,
|
| 188118 |
+
"learning_rate": 8.898493589743592e-06,
|
| 188119 |
+
"loss": 0.656,
|
| 188120 |
+
"step": 69760
|
| 188121 |
+
},
|
| 188122 |
+
{
|
| 188123 |
+
"epoch": 557.61,
|
| 188124 |
+
"learning_rate": 8.898413461538463e-06,
|
| 188125 |
+
"loss": 1.1714,
|
| 188126 |
+
"step": 69765
|
| 188127 |
+
},
|
| 188128 |
+
{
|
| 188129 |
+
"epoch": 557.65,
|
| 188130 |
+
"learning_rate": 8.898333333333334e-06,
|
| 188131 |
+
"loss": 0.347,
|
| 188132 |
+
"step": 69770
|
| 188133 |
+
},
|
| 188134 |
+
{
|
| 188135 |
+
"epoch": 557.69,
|
| 188136 |
+
"learning_rate": 8.898253205128206e-06,
|
| 188137 |
+
"loss": 0.3143,
|
| 188138 |
+
"step": 69775
|
| 188139 |
+
},
|
| 188140 |
+
{
|
| 188141 |
+
"epoch": 557.73,
|
| 188142 |
+
"learning_rate": 8.898173076923079e-06,
|
| 188143 |
+
"loss": 0.347,
|
| 188144 |
+
"step": 69780
|
| 188145 |
+
},
|
| 188146 |
+
{
|
| 188147 |
+
"epoch": 557.76,
|
| 188148 |
+
"learning_rate": 8.898092948717948e-06,
|
| 188149 |
+
"loss": 0.6589,
|
| 188150 |
+
"step": 69785
|
| 188151 |
+
},
|
| 188152 |
+
{
|
| 188153 |
+
"epoch": 557.8,
|
| 188154 |
+
"learning_rate": 8.898012820512821e-06,
|
| 188155 |
+
"loss": 1.2161,
|
| 188156 |
+
"step": 69790
|
| 188157 |
+
},
|
| 188158 |
+
{
|
| 188159 |
+
"epoch": 557.84,
|
| 188160 |
+
"learning_rate": 8.897932692307693e-06,
|
| 188161 |
+
"loss": 0.2778,
|
| 188162 |
+
"step": 69795
|
| 188163 |
+
},
|
| 188164 |
+
{
|
| 188165 |
+
"epoch": 557.88,
|
| 188166 |
+
"learning_rate": 8.897852564102564e-06,
|
| 188167 |
+
"loss": 0.3265,
|
| 188168 |
+
"step": 69800
|
| 188169 |
+
},
|
| 188170 |
+
{
|
| 188171 |
+
"epoch": 557.92,
|
| 188172 |
+
"learning_rate": 8.897772435897437e-06,
|
| 188173 |
+
"loss": 0.339,
|
| 188174 |
+
"step": 69805
|
| 188175 |
+
},
|
| 188176 |
+
{
|
| 188177 |
+
"epoch": 557.96,
|
| 188178 |
+
"learning_rate": 8.897692307692309e-06,
|
| 188179 |
+
"loss": 0.5141,
|
| 188180 |
+
"step": 69810
|
| 188181 |
+
},
|
| 188182 |
+
{
|
| 188183 |
+
"epoch": 558.0,
|
| 188184 |
+
"eval_loss": 0.4530166685581207,
|
| 188185 |
+
"eval_runtime": 40.5863,
|
| 188186 |
+
"eval_samples_per_second": 20.672,
|
| 188187 |
+
"eval_steps_per_second": 0.665,
|
| 188188 |
+
"eval_wer": 0.1972293109733868,
|
| 188189 |
+
"step": 69814
|
| 188190 |
+
},
|
| 188191 |
+
{
|
| 188192 |
+
"epoch": 558.01,
|
| 188193 |
+
"learning_rate": 8.89761217948718e-06,
|
| 188194 |
+
"loss": 0.3754,
|
| 188195 |
+
"step": 69815
|
| 188196 |
+
},
|
| 188197 |
+
{
|
| 188198 |
+
"epoch": 558.05,
|
| 188199 |
+
"learning_rate": 8.897532051282051e-06,
|
| 188200 |
+
"loss": 0.3432,
|
| 188201 |
+
"step": 69820
|
| 188202 |
+
},
|
| 188203 |
+
{
|
| 188204 |
+
"epoch": 558.09,
|
| 188205 |
+
"learning_rate": 8.897451923076924e-06,
|
| 188206 |
+
"loss": 0.3502,
|
| 188207 |
+
"step": 69825
|
| 188208 |
+
},
|
| 188209 |
+
{
|
| 188210 |
+
"epoch": 558.13,
|
| 188211 |
+
"learning_rate": 8.897371794871796e-06,
|
| 188212 |
+
"loss": 0.3163,
|
| 188213 |
+
"step": 69830
|
| 188214 |
+
},
|
| 188215 |
+
{
|
| 188216 |
+
"epoch": 558.17,
|
| 188217 |
+
"learning_rate": 8.897291666666667e-06,
|
| 188218 |
+
"loss": 0.5237,
|
| 188219 |
+
"step": 69835
|
| 188220 |
+
},
|
| 188221 |
+
{
|
| 188222 |
+
"epoch": 558.21,
|
| 188223 |
+
"learning_rate": 8.897211538461538e-06,
|
| 188224 |
+
"loss": 1.2681,
|
| 188225 |
+
"step": 69840
|
| 188226 |
+
},
|
| 188227 |
+
{
|
| 188228 |
+
"epoch": 558.25,
|
| 188229 |
+
"learning_rate": 8.897131410256411e-06,
|
| 188230 |
+
"loss": 0.2817,
|
| 188231 |
+
"step": 69845
|
| 188232 |
+
},
|
| 188233 |
+
{
|
| 188234 |
+
"epoch": 558.29,
|
| 188235 |
+
"learning_rate": 8.897051282051283e-06,
|
| 188236 |
+
"loss": 0.2925,
|
| 188237 |
+
"step": 69850
|
| 188238 |
+
},
|
| 188239 |
+
{
|
| 188240 |
+
"epoch": 558.33,
|
| 188241 |
+
"learning_rate": 8.896971153846154e-06,
|
| 188242 |
+
"loss": 0.3963,
|
| 188243 |
+
"step": 69855
|
| 188244 |
+
},
|
| 188245 |
+
{
|
| 188246 |
+
"epoch": 558.37,
|
| 188247 |
+
"learning_rate": 8.896891025641027e-06,
|
| 188248 |
+
"loss": 0.7644,
|
| 188249 |
+
"step": 69860
|
| 188250 |
+
},
|
| 188251 |
+
{
|
| 188252 |
+
"epoch": 558.41,
|
| 188253 |
+
"learning_rate": 8.896810897435899e-06,
|
| 188254 |
+
"loss": 1.1559,
|
| 188255 |
+
"step": 69865
|
| 188256 |
+
},
|
| 188257 |
+
{
|
| 188258 |
+
"epoch": 558.45,
|
| 188259 |
+
"learning_rate": 8.89673076923077e-06,
|
| 188260 |
+
"loss": 0.3287,
|
| 188261 |
+
"step": 69870
|
| 188262 |
+
},
|
| 188263 |
+
{
|
| 188264 |
+
"epoch": 558.49,
|
| 188265 |
+
"learning_rate": 8.896650641025641e-06,
|
| 188266 |
+
"loss": 0.2861,
|
| 188267 |
+
"step": 69875
|
| 188268 |
+
},
|
| 188269 |
+
{
|
| 188270 |
+
"epoch": 558.53,
|
| 188271 |
+
"learning_rate": 8.896570512820514e-06,
|
| 188272 |
+
"loss": 0.2943,
|
| 188273 |
+
"step": 69880
|
| 188274 |
+
},
|
| 188275 |
+
{
|
| 188276 |
+
"epoch": 558.57,
|
| 188277 |
+
"learning_rate": 8.896490384615386e-06,
|
| 188278 |
+
"loss": 0.5952,
|
| 188279 |
+
"step": 69885
|
| 188280 |
+
},
|
| 188281 |
+
{
|
| 188282 |
+
"epoch": 558.61,
|
| 188283 |
+
"learning_rate": 8.896410256410257e-06,
|
| 188284 |
+
"loss": 1.3276,
|
| 188285 |
+
"step": 69890
|
| 188286 |
+
},
|
| 188287 |
+
{
|
| 188288 |
+
"epoch": 558.65,
|
| 188289 |
+
"learning_rate": 8.896330128205128e-06,
|
| 188290 |
+
"loss": 0.3448,
|
| 188291 |
+
"step": 69895
|
| 188292 |
+
},
|
| 188293 |
+
{
|
| 188294 |
+
"epoch": 558.69,
|
| 188295 |
+
"learning_rate": 8.896250000000001e-06,
|
| 188296 |
+
"loss": 0.3679,
|
| 188297 |
+
"step": 69900
|
| 188298 |
+
},
|
| 188299 |
+
{
|
| 188300 |
+
"epoch": 558.73,
|
| 188301 |
+
"learning_rate": 8.896169871794873e-06,
|
| 188302 |
+
"loss": 0.325,
|
| 188303 |
+
"step": 69905
|
| 188304 |
+
},
|
| 188305 |
+
{
|
| 188306 |
+
"epoch": 558.77,
|
| 188307 |
+
"learning_rate": 8.896089743589744e-06,
|
| 188308 |
+
"loss": 0.575,
|
| 188309 |
+
"step": 69910
|
| 188310 |
+
},
|
| 188311 |
+
{
|
| 188312 |
+
"epoch": 558.81,
|
| 188313 |
+
"learning_rate": 8.896009615384617e-06,
|
| 188314 |
+
"loss": 1.0585,
|
| 188315 |
+
"step": 69915
|
| 188316 |
+
},
|
| 188317 |
+
{
|
| 188318 |
+
"epoch": 558.85,
|
| 188319 |
+
"learning_rate": 8.895929487179487e-06,
|
| 188320 |
+
"loss": 0.2871,
|
| 188321 |
+
"step": 69920
|
| 188322 |
+
},
|
| 188323 |
+
{
|
| 188324 |
+
"epoch": 558.89,
|
| 188325 |
+
"learning_rate": 8.89584935897436e-06,
|
| 188326 |
+
"loss": 0.3916,
|
| 188327 |
+
"step": 69925
|
| 188328 |
+
},
|
| 188329 |
+
{
|
| 188330 |
+
"epoch": 558.93,
|
| 188331 |
+
"learning_rate": 8.895769230769231e-06,
|
| 188332 |
+
"loss": 0.3471,
|
| 188333 |
+
"step": 69930
|
| 188334 |
+
},
|
| 188335 |
+
{
|
| 188336 |
+
"epoch": 558.97,
|
| 188337 |
+
"learning_rate": 8.895689102564103e-06,
|
| 188338 |
+
"loss": 0.5273,
|
| 188339 |
+
"step": 69935
|
| 188340 |
+
},
|
| 188341 |
+
{
|
| 188342 |
+
"epoch": 559.0,
|
| 188343 |
+
"eval_loss": 0.3516188859939575,
|
| 188344 |
+
"eval_runtime": 40.4439,
|
| 188345 |
+
"eval_samples_per_second": 20.72,
|
| 188346 |
+
"eval_steps_per_second": 0.668,
|
| 188347 |
+
"eval_wer": 0.1972123368920522,
|
| 188348 |
+
"step": 69939
|
| 188349 |
+
},
|
| 188350 |
+
{
|
| 188351 |
+
"epoch": 564.01,
|
| 188352 |
+
"learning_rate": 8.895608974358974e-06,
|
| 188353 |
+
"loss": 0.3957,
|
| 188354 |
+
"step": 69940
|
| 188355 |
+
},
|
| 188356 |
+
{
|
| 188357 |
+
"epoch": 564.05,
|
| 188358 |
+
"learning_rate": 8.895528846153847e-06,
|
| 188359 |
+
"loss": 0.3196,
|
| 188360 |
+
"step": 69945
|
| 188361 |
+
},
|
| 188362 |
+
{
|
| 188363 |
+
"epoch": 564.09,
|
| 188364 |
+
"learning_rate": 8.895448717948718e-06,
|
| 188365 |
+
"loss": 0.2832,
|
| 188366 |
+
"step": 69950
|
| 188367 |
+
},
|
| 188368 |
+
{
|
| 188369 |
+
"epoch": 564.13,
|
| 188370 |
+
"learning_rate": 8.89536858974359e-06,
|
| 188371 |
+
"loss": 0.3522,
|
| 188372 |
+
"step": 69955
|
| 188373 |
+
},
|
| 188374 |
+
{
|
| 188375 |
+
"epoch": 564.17,
|
| 188376 |
+
"learning_rate": 8.895288461538463e-06,
|
| 188377 |
+
"loss": 0.481,
|
| 188378 |
+
"step": 69960
|
| 188379 |
+
},
|
| 188380 |
+
{
|
| 188381 |
+
"epoch": 564.21,
|
| 188382 |
+
"learning_rate": 8.895208333333334e-06,
|
| 188383 |
+
"loss": 1.0585,
|
| 188384 |
+
"step": 69965
|
| 188385 |
+
},
|
| 188386 |
+
{
|
| 188387 |
+
"epoch": 564.25,
|
| 188388 |
+
"learning_rate": 8.895128205128206e-06,
|
| 188389 |
+
"loss": 0.3759,
|
| 188390 |
+
"step": 69970
|
| 188391 |
+
},
|
| 188392 |
+
{
|
| 188393 |
+
"epoch": 564.29,
|
| 188394 |
+
"learning_rate": 8.895048076923077e-06,
|
| 188395 |
+
"loss": 0.333,
|
| 188396 |
+
"step": 69975
|
| 188397 |
+
},
|
| 188398 |
+
{
|
| 188399 |
+
"epoch": 564.33,
|
| 188400 |
+
"learning_rate": 8.89496794871795e-06,
|
| 188401 |
+
"loss": 0.4362,
|
| 188402 |
+
"step": 69980
|
| 188403 |
+
},
|
| 188404 |
+
{
|
| 188405 |
+
"epoch": 564.37,
|
| 188406 |
+
"learning_rate": 8.894887820512821e-06,
|
| 188407 |
+
"loss": 0.5773,
|
| 188408 |
+
"step": 69985
|
| 188409 |
+
},
|
| 188410 |
+
{
|
| 188411 |
+
"epoch": 564.41,
|
| 188412 |
+
"learning_rate": 8.894807692307693e-06,
|
| 188413 |
+
"loss": 1.2406,
|
| 188414 |
+
"step": 69990
|
| 188415 |
+
},
|
| 188416 |
+
{
|
| 188417 |
+
"epoch": 564.45,
|
| 188418 |
+
"learning_rate": 8.894727564102564e-06,
|
| 188419 |
+
"loss": 0.3018,
|
| 188420 |
+
"step": 69995
|
| 188421 |
+
},
|
| 188422 |
+
{
|
| 188423 |
+
"epoch": 564.49,
|
| 188424 |
+
"learning_rate": 8.894647435897437e-06,
|
| 188425 |
+
"loss": 0.2916,
|
| 188426 |
+
"step": 70000
|
| 188427 |
+
},
|
| 188428 |
+
{
|
| 188429 |
+
"epoch": 564.53,
|
| 188430 |
+
"learning_rate": 8.894567307692308e-06,
|
| 188431 |
+
"loss": 0.312,
|
| 188432 |
+
"step": 70005
|
| 188433 |
+
},
|
| 188434 |
+
{
|
| 188435 |
+
"epoch": 564.57,
|
| 188436 |
+
"learning_rate": 8.89448717948718e-06,
|
| 188437 |
+
"loss": 0.6622,
|
| 188438 |
+
"step": 70010
|
| 188439 |
+
},
|
| 188440 |
+
{
|
| 188441 |
+
"epoch": 564.61,
|
| 188442 |
+
"learning_rate": 8.894407051282053e-06,
|
| 188443 |
+
"loss": 1.134,
|
| 188444 |
+
"step": 70015
|
| 188445 |
+
},
|
| 188446 |
+
{
|
| 188447 |
+
"epoch": 564.65,
|
| 188448 |
+
"learning_rate": 8.894326923076924e-06,
|
| 188449 |
+
"loss": 0.3032,
|
| 188450 |
+
"step": 70020
|
| 188451 |
+
},
|
| 188452 |
+
{
|
| 188453 |
+
"epoch": 564.69,
|
| 188454 |
+
"learning_rate": 8.894246794871796e-06,
|
| 188455 |
+
"loss": 0.3121,
|
| 188456 |
+
"step": 70025
|
| 188457 |
+
},
|
| 188458 |
+
{
|
| 188459 |
+
"epoch": 564.73,
|
| 188460 |
+
"learning_rate": 8.894166666666667e-06,
|
| 188461 |
+
"loss": 0.3437,
|
| 188462 |
+
"step": 70030
|
| 188463 |
+
},
|
| 188464 |
+
{
|
| 188465 |
+
"epoch": 564.77,
|
| 188466 |
+
"learning_rate": 8.89408653846154e-06,
|
| 188467 |
+
"loss": 0.5597,
|
| 188468 |
+
"step": 70035
|
| 188469 |
+
},
|
| 188470 |
+
{
|
| 188471 |
+
"epoch": 564.81,
|
| 188472 |
+
"learning_rate": 8.894006410256411e-06,
|
| 188473 |
+
"loss": 1.1312,
|
| 188474 |
+
"step": 70040
|
| 188475 |
+
},
|
| 188476 |
+
{
|
| 188477 |
+
"epoch": 564.85,
|
| 188478 |
+
"learning_rate": 8.893926282051283e-06,
|
| 188479 |
+
"loss": 0.3284,
|
| 188480 |
+
"step": 70045
|
| 188481 |
+
},
|
| 188482 |
+
{
|
| 188483 |
+
"epoch": 564.89,
|
| 188484 |
+
"learning_rate": 8.893846153846156e-06,
|
| 188485 |
+
"loss": 0.3695,
|
| 188486 |
+
"step": 70050
|
| 188487 |
+
},
|
| 188488 |
+
{
|
| 188489 |
+
"epoch": 564.93,
|
| 188490 |
+
"learning_rate": 8.893766025641027e-06,
|
| 188491 |
+
"loss": 0.3754,
|
| 188492 |
+
"step": 70055
|
| 188493 |
+
},
|
| 188494 |
+
{
|
| 188495 |
+
"epoch": 564.97,
|
| 188496 |
+
"learning_rate": 8.893685897435898e-06,
|
| 188497 |
+
"loss": 0.5765,
|
| 188498 |
+
"step": 70060
|
| 188499 |
+
},
|
| 188500 |
+
{
|
| 188501 |
+
"epoch": 565.0,
|
| 188502 |
+
"eval_loss": 0.390587717294693,
|
| 188503 |
+
"eval_runtime": 40.146,
|
| 188504 |
+
"eval_samples_per_second": 20.874,
|
| 188505 |
+
"eval_steps_per_second": 0.673,
|
| 188506 |
+
"eval_wer": 0.19272673733804477,
|
| 188507 |
+
"step": 70063
|
| 188508 |
+
},
|
| 188509 |
+
{
|
| 188510 |
+
"epoch": 565.02,
|
| 188511 |
+
"learning_rate": 8.89360576923077e-06,
|
| 188512 |
+
"loss": 0.4398,
|
| 188513 |
+
"step": 70065
|
| 188514 |
+
},
|
| 188515 |
+
{
|
| 188516 |
+
"epoch": 565.06,
|
| 188517 |
+
"learning_rate": 8.893525641025643e-06,
|
| 188518 |
+
"loss": 0.3396,
|
| 188519 |
+
"step": 70070
|
| 188520 |
+
},
|
| 188521 |
+
{
|
| 188522 |
+
"epoch": 565.1,
|
| 188523 |
+
"learning_rate": 8.893445512820513e-06,
|
| 188524 |
+
"loss": 0.3126,
|
| 188525 |
+
"step": 70075
|
| 188526 |
+
},
|
| 188527 |
+
{
|
| 188528 |
+
"epoch": 565.14,
|
| 188529 |
+
"learning_rate": 8.893365384615386e-06,
|
| 188530 |
+
"loss": 0.368,
|
| 188531 |
+
"step": 70080
|
| 188532 |
+
},
|
| 188533 |
+
{
|
| 188534 |
+
"epoch": 565.18,
|
| 188535 |
+
"learning_rate": 8.893285256410257e-06,
|
| 188536 |
+
"loss": 0.6406,
|
| 188537 |
+
"step": 70085
|
| 188538 |
+
},
|
| 188539 |
+
{
|
| 188540 |
+
"epoch": 565.22,
|
| 188541 |
+
"learning_rate": 8.893205128205128e-06,
|
| 188542 |
+
"loss": 1.042,
|
| 188543 |
+
"step": 70090
|
| 188544 |
+
},
|
| 188545 |
+
{
|
| 188546 |
+
"epoch": 565.26,
|
| 188547 |
+
"learning_rate": 8.893125e-06,
|
| 188548 |
+
"loss": 0.3395,
|
| 188549 |
+
"step": 70095
|
| 188550 |
+
},
|
| 188551 |
+
{
|
| 188552 |
+
"epoch": 565.3,
|
| 188553 |
+
"learning_rate": 8.893044871794873e-06,
|
| 188554 |
+
"loss": 0.3251,
|
| 188555 |
+
"step": 70100
|
| 188556 |
+
},
|
| 188557 |
+
{
|
| 188558 |
+
"epoch": 565.34,
|
| 188559 |
+
"learning_rate": 8.892964743589744e-06,
|
| 188560 |
+
"loss": 0.3903,
|
| 188561 |
+
"step": 70105
|
| 188562 |
+
},
|
| 188563 |
+
{
|
| 188564 |
+
"epoch": 565.38,
|
| 188565 |
+
"learning_rate": 8.892884615384615e-06,
|
| 188566 |
+
"loss": 0.6584,
|
| 188567 |
+
"step": 70110
|
| 188568 |
+
},
|
| 188569 |
+
{
|
| 188570 |
+
"epoch": 565.42,
|
| 188571 |
+
"learning_rate": 8.892804487179488e-06,
|
| 188572 |
+
"loss": 1.0984,
|
| 188573 |
+
"step": 70115
|
| 188574 |
+
},
|
| 188575 |
+
{
|
| 188576 |
+
"epoch": 565.46,
|
| 188577 |
+
"learning_rate": 8.89272435897436e-06,
|
| 188578 |
+
"loss": 0.3313,
|
| 188579 |
+
"step": 70120
|
| 188580 |
+
},
|
| 188581 |
+
{
|
| 188582 |
+
"epoch": 565.5,
|
| 188583 |
+
"learning_rate": 8.892644230769231e-06,
|
| 188584 |
+
"loss": 0.3197,
|
| 188585 |
+
"step": 70125
|
| 188586 |
+
},
|
| 188587 |
+
{
|
| 188588 |
+
"epoch": 565.54,
|
| 188589 |
+
"learning_rate": 8.892564102564103e-06,
|
| 188590 |
+
"loss": 0.4314,
|
| 188591 |
+
"step": 70130
|
| 188592 |
+
},
|
| 188593 |
+
{
|
| 188594 |
+
"epoch": 565.58,
|
| 188595 |
+
"learning_rate": 8.892483974358976e-06,
|
| 188596 |
+
"loss": 0.5728,
|
| 188597 |
+
"step": 70135
|
| 188598 |
+
},
|
| 188599 |
+
{
|
| 188600 |
+
"epoch": 565.62,
|
| 188601 |
+
"learning_rate": 8.892403846153847e-06,
|
| 188602 |
+
"loss": 1.0667,
|
| 188603 |
+
"step": 70140
|
| 188604 |
+
},
|
| 188605 |
+
{
|
| 188606 |
+
"epoch": 565.66,
|
| 188607 |
+
"learning_rate": 8.892323717948718e-06,
|
| 188608 |
+
"loss": 0.2618,
|
| 188609 |
+
"step": 70145
|
| 188610 |
+
},
|
| 188611 |
+
{
|
| 188612 |
+
"epoch": 565.7,
|
| 188613 |
+
"learning_rate": 8.89224358974359e-06,
|
| 188614 |
+
"loss": 0.2746,
|
| 188615 |
+
"step": 70150
|
| 188616 |
+
},
|
| 188617 |
+
{
|
| 188618 |
+
"epoch": 565.74,
|
| 188619 |
+
"learning_rate": 8.892163461538463e-06,
|
| 188620 |
+
"loss": 0.3862,
|
| 188621 |
+
"step": 70155
|
| 188622 |
+
},
|
| 188623 |
+
{
|
| 188624 |
+
"epoch": 565.78,
|
| 188625 |
+
"learning_rate": 8.892083333333334e-06,
|
| 188626 |
+
"loss": 0.6671,
|
| 188627 |
+
"step": 70160
|
| 188628 |
+
},
|
| 188629 |
+
{
|
| 188630 |
+
"epoch": 565.82,
|
| 188631 |
+
"learning_rate": 8.892003205128205e-06,
|
| 188632 |
+
"loss": 1.2741,
|
| 188633 |
+
"step": 70165
|
| 188634 |
+
},
|
| 188635 |
+
{
|
| 188636 |
+
"epoch": 565.86,
|
| 188637 |
+
"learning_rate": 8.891923076923079e-06,
|
| 188638 |
+
"loss": 0.2632,
|
| 188639 |
+
"step": 70170
|
| 188640 |
+
},
|
| 188641 |
+
{
|
| 188642 |
+
"epoch": 565.9,
|
| 188643 |
+
"learning_rate": 8.89184294871795e-06,
|
| 188644 |
+
"loss": 0.3287,
|
| 188645 |
+
"step": 70175
|
| 188646 |
+
},
|
| 188647 |
+
{
|
| 188648 |
+
"epoch": 565.94,
|
| 188649 |
+
"learning_rate": 8.891762820512821e-06,
|
| 188650 |
+
"loss": 0.4261,
|
| 188651 |
+
"step": 70180
|
| 188652 |
+
},
|
| 188653 |
+
{
|
| 188654 |
+
"epoch": 565.98,
|
| 188655 |
+
"learning_rate": 8.891682692307693e-06,
|
| 188656 |
+
"loss": 0.7376,
|
| 188657 |
+
"step": 70185
|
| 188658 |
+
},
|
| 188659 |
+
{
|
| 188660 |
+
"epoch": 566.0,
|
| 188661 |
+
"eval_loss": 0.37903761863708496,
|
| 188662 |
+
"eval_runtime": 40.3284,
|
| 188663 |
+
"eval_samples_per_second": 20.804,
|
| 188664 |
+
"eval_steps_per_second": 0.67,
|
| 188665 |
+
"eval_wer": 0.19290399522957663,
|
| 188666 |
+
"step": 70187
|
| 188667 |
}
|
| 188668 |
],
|
| 188669 |
"max_steps": 620000,
|
| 188670 |
"num_train_epochs": 5000,
|
| 188671 |
+
"total_flos": 1.9751188077570854e+20,
|
| 188672 |
"trial_name": null,
|
| 188673 |
"trial_params": null
|
| 188674 |
}
|
model-bin/finetune/base/{checkpoint-68444 β checkpoint-70187}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1629852210.1896331/events.out.tfevents.1629852210.7e498afd5545.905.3
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:68149130b4501f279b49c95af5739b4849c30d81b2537c4db3eef5e13e83052f
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629852859.7482474/events.out.tfevents.1629852859.7e498afd5545.905.5
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:820d9a447868e8b671a9c944cf7a173b01fccf4eb4dc0241634608ce9ee3a47f
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629853517.0685782/events.out.tfevents.1629853517.7e498afd5545.905.7
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6c896caeebc055ac1596417252dbebba435869c3cfefc38bdf1d6a7acfa309d7
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629854178.1266189/events.out.tfevents.1629854178.7e498afd5545.905.9
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4615f993eb4b009d2c1ca12dddfd5c9df6be7ced3525bf6f46b0c2b3c1908b39
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629854840.4334018/events.out.tfevents.1629854840.7e498afd5545.905.11
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5d305ac4e9a27656cdb897d260e5d0e98a803b8ff05141982ecc6da238ee5d30
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1629852210.7e498afd5545.905.2
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f451c2f60d2d81d83260193257943c10c144fe8f04fac72817f96e0cf61f9a14
|
| 3 |
+
size 8462
|
model-bin/finetune/base/log/events.out.tfevents.1629852859.7e498afd5545.905.4
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c2990e5d7ba02c7a7776171c76900dc83d7049b09f4bdf62d77bfa5e72c5e17a
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629853517.7e498afd5545.905.6
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c68c1156511fb22a45a6bbe6066f44481f65250e8ce3f7ad4e5a71436b7a9f2b
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629854178.7e498afd5545.905.8
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4b7501cb96fceed8e8999002ce1839c19aa06143709dc51b9760f0787604f761
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629854840.7e498afd5545.905.10
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3de3d39c0de58cadb428fda704ea8a11efe56a422badbc0de00120dabb99df48
|
| 3 |
+
size 8622
|