"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-149584 β checkpoint-150205}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-149584 β checkpoint-150205}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-149584 β checkpoint-150205}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-149584 β checkpoint-150205}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-149584 β checkpoint-150205}/rng_state.pth +2 -2
- model-bin/finetune/base/{checkpoint-149584 β checkpoint-150205}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-149584 β checkpoint-150205}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-149584 β checkpoint-150205}/trainer_state.json +799 -4
- model-bin/finetune/base/{checkpoint-149584 β checkpoint-150205}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630241766.1674721/events.out.tfevents.1630241766.cc93b136ebf5.1086.169 +3 -0
- model-bin/finetune/base/log/1630242210.9386408/events.out.tfevents.1630242210.cc93b136ebf5.1086.171 +3 -0
- model-bin/finetune/base/log/1630242652.9526577/events.out.tfevents.1630242652.cc93b136ebf5.1086.173 +3 -0
- model-bin/finetune/base/log/1630243093.5265884/events.out.tfevents.1630243093.cc93b136ebf5.1086.175 +3 -0
- model-bin/finetune/base/log/1630243533.7269483/events.out.tfevents.1630243533.cc93b136ebf5.1086.177 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630241766.cc93b136ebf5.1086.168 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630242210.cc93b136ebf5.1086.170 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630242652.cc93b136ebf5.1086.172 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630243093.cc93b136ebf5.1086.174 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630243533.cc93b136ebf5.1086.176 +3 -0
model-bin/finetune/base/{checkpoint-149584 β checkpoint-150205}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-149584 β checkpoint-150205}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5341661b50f4e596b6fe02c28dfc7ba32504ab70fca5178ea2952aac4a6cbdf1
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-149584 β checkpoint-150205}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-149584 β checkpoint-150205}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2831ef8aa83687ecb9c8fa7338b9766386e38377b04dd35c04a8c8c06f8148f2
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-149584 β checkpoint-150205}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0626ef614583f5ab57c8ac6ae463c7a0f18875e587996d7e4c49f829145e7e88
|
| 3 |
+
size 14503
|
model-bin/finetune/base/{checkpoint-149584 β checkpoint-150205}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3540207233bcabcb86e1d4ad4f38a1c87a6c2cdf020124389fd687b8b31fe0dd
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-149584 β checkpoint-150205}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:63f3f86c60dbd72222ca372243fb135bc1a49b12c2bc13b3077d16a8db5290be
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-149584 β checkpoint-150205}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1668011442822563,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -289680,11 +289680,806 @@
|
|
| 289680 |
"eval_steps_per_second": 0.654,
|
| 289681 |
"eval_wer": 0.17671063178605262,
|
| 289682 |
"step": 149584
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 289683 |
}
|
| 289684 |
],
|
| 289685 |
-
"max_steps":
|
| 289686 |
"num_train_epochs": 5000,
|
| 289687 |
-
"total_flos": 4.
|
| 289688 |
"trial_name": null,
|
| 289689 |
"trial_params": null
|
| 289690 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1668011442822563,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
| 4 |
+
"epoch": 1211.0,
|
| 5 |
+
"global_step": 150205,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 289680 |
"eval_steps_per_second": 0.654,
|
| 289681 |
"eval_wer": 0.17671063178605262,
|
| 289682 |
"step": 149584
|
| 289683 |
+
},
|
| 289684 |
+
{
|
| 289685 |
+
"epoch": 1206.01,
|
| 289686 |
+
"learning_rate": 7.600549273021002e-06,
|
| 289687 |
+
"loss": 0.4851,
|
| 289688 |
+
"step": 149585
|
| 289689 |
+
},
|
| 289690 |
+
{
|
| 289691 |
+
"epoch": 1206.05,
|
| 289692 |
+
"learning_rate": 7.600468497576737e-06,
|
| 289693 |
+
"loss": 0.279,
|
| 289694 |
+
"step": 149590
|
| 289695 |
+
},
|
| 289696 |
+
{
|
| 289697 |
+
"epoch": 1206.09,
|
| 289698 |
+
"learning_rate": 7.600387722132473e-06,
|
| 289699 |
+
"loss": 0.3121,
|
| 289700 |
+
"step": 149595
|
| 289701 |
+
},
|
| 289702 |
+
{
|
| 289703 |
+
"epoch": 1206.13,
|
| 289704 |
+
"learning_rate": 7.600306946688207e-06,
|
| 289705 |
+
"loss": 0.3105,
|
| 289706 |
+
"step": 149600
|
| 289707 |
+
},
|
| 289708 |
+
{
|
| 289709 |
+
"epoch": 1206.17,
|
| 289710 |
+
"learning_rate": 7.600226171243943e-06,
|
| 289711 |
+
"loss": 0.5346,
|
| 289712 |
+
"step": 149605
|
| 289713 |
+
},
|
| 289714 |
+
{
|
| 289715 |
+
"epoch": 1206.21,
|
| 289716 |
+
"learning_rate": 7.600145395799678e-06,
|
| 289717 |
+
"loss": 1.0911,
|
| 289718 |
+
"step": 149610
|
| 289719 |
+
},
|
| 289720 |
+
{
|
| 289721 |
+
"epoch": 1206.25,
|
| 289722 |
+
"learning_rate": 7.600064620355413e-06,
|
| 289723 |
+
"loss": 0.2746,
|
| 289724 |
+
"step": 149615
|
| 289725 |
+
},
|
| 289726 |
+
{
|
| 289727 |
+
"epoch": 1206.29,
|
| 289728 |
+
"learning_rate": 7.599983844911148e-06,
|
| 289729 |
+
"loss": 0.3218,
|
| 289730 |
+
"step": 149620
|
| 289731 |
+
},
|
| 289732 |
+
{
|
| 289733 |
+
"epoch": 1206.33,
|
| 289734 |
+
"learning_rate": 7.599903069466883e-06,
|
| 289735 |
+
"loss": 0.3109,
|
| 289736 |
+
"step": 149625
|
| 289737 |
+
},
|
| 289738 |
+
{
|
| 289739 |
+
"epoch": 1206.37,
|
| 289740 |
+
"learning_rate": 7.599822294022618e-06,
|
| 289741 |
+
"loss": 0.4662,
|
| 289742 |
+
"step": 149630
|
| 289743 |
+
},
|
| 289744 |
+
{
|
| 289745 |
+
"epoch": 1206.41,
|
| 289746 |
+
"learning_rate": 7.599741518578353e-06,
|
| 289747 |
+
"loss": 1.0338,
|
| 289748 |
+
"step": 149635
|
| 289749 |
+
},
|
| 289750 |
+
{
|
| 289751 |
+
"epoch": 1206.45,
|
| 289752 |
+
"learning_rate": 7.599660743134088e-06,
|
| 289753 |
+
"loss": 0.2714,
|
| 289754 |
+
"step": 149640
|
| 289755 |
+
},
|
| 289756 |
+
{
|
| 289757 |
+
"epoch": 1206.49,
|
| 289758 |
+
"learning_rate": 7.599579967689823e-06,
|
| 289759 |
+
"loss": 0.2534,
|
| 289760 |
+
"step": 149645
|
| 289761 |
+
},
|
| 289762 |
+
{
|
| 289763 |
+
"epoch": 1206.53,
|
| 289764 |
+
"learning_rate": 7.599499192245558e-06,
|
| 289765 |
+
"loss": 0.3401,
|
| 289766 |
+
"step": 149650
|
| 289767 |
+
},
|
| 289768 |
+
{
|
| 289769 |
+
"epoch": 1206.57,
|
| 289770 |
+
"learning_rate": 7.599418416801293e-06,
|
| 289771 |
+
"loss": 0.5341,
|
| 289772 |
+
"step": 149655
|
| 289773 |
+
},
|
| 289774 |
+
{
|
| 289775 |
+
"epoch": 1206.61,
|
| 289776 |
+
"learning_rate": 7.5993376413570285e-06,
|
| 289777 |
+
"loss": 1.1729,
|
| 289778 |
+
"step": 149660
|
| 289779 |
+
},
|
| 289780 |
+
{
|
| 289781 |
+
"epoch": 1206.65,
|
| 289782 |
+
"learning_rate": 7.599256865912763e-06,
|
| 289783 |
+
"loss": 0.2943,
|
| 289784 |
+
"step": 149665
|
| 289785 |
+
},
|
| 289786 |
+
{
|
| 289787 |
+
"epoch": 1206.69,
|
| 289788 |
+
"learning_rate": 7.5991760904684985e-06,
|
| 289789 |
+
"loss": 0.3001,
|
| 289790 |
+
"step": 149670
|
| 289791 |
+
},
|
| 289792 |
+
{
|
| 289793 |
+
"epoch": 1206.73,
|
| 289794 |
+
"learning_rate": 7.599095315024233e-06,
|
| 289795 |
+
"loss": 0.3681,
|
| 289796 |
+
"step": 149675
|
| 289797 |
+
},
|
| 289798 |
+
{
|
| 289799 |
+
"epoch": 1206.77,
|
| 289800 |
+
"learning_rate": 7.5990145395799685e-06,
|
| 289801 |
+
"loss": 0.54,
|
| 289802 |
+
"step": 149680
|
| 289803 |
+
},
|
| 289804 |
+
{
|
| 289805 |
+
"epoch": 1206.81,
|
| 289806 |
+
"learning_rate": 7.5989337641357034e-06,
|
| 289807 |
+
"loss": 1.0087,
|
| 289808 |
+
"step": 149685
|
| 289809 |
+
},
|
| 289810 |
+
{
|
| 289811 |
+
"epoch": 1206.85,
|
| 289812 |
+
"learning_rate": 7.5988529886914384e-06,
|
| 289813 |
+
"loss": 0.2777,
|
| 289814 |
+
"step": 149690
|
| 289815 |
+
},
|
| 289816 |
+
{
|
| 289817 |
+
"epoch": 1206.89,
|
| 289818 |
+
"learning_rate": 7.598772213247173e-06,
|
| 289819 |
+
"loss": 0.2576,
|
| 289820 |
+
"step": 149695
|
| 289821 |
+
},
|
| 289822 |
+
{
|
| 289823 |
+
"epoch": 1206.93,
|
| 289824 |
+
"learning_rate": 7.598691437802908e-06,
|
| 289825 |
+
"loss": 0.3037,
|
| 289826 |
+
"step": 149700
|
| 289827 |
+
},
|
| 289828 |
+
{
|
| 289829 |
+
"epoch": 1206.97,
|
| 289830 |
+
"learning_rate": 7.598610662358643e-06,
|
| 289831 |
+
"loss": 0.5127,
|
| 289832 |
+
"step": 149705
|
| 289833 |
+
},
|
| 289834 |
+
{
|
| 289835 |
+
"epoch": 1207.0,
|
| 289836 |
+
"eval_loss": 0.40235501527786255,
|
| 289837 |
+
"eval_runtime": 42.5624,
|
| 289838 |
+
"eval_samples_per_second": 19.736,
|
| 289839 |
+
"eval_steps_per_second": 0.634,
|
| 289840 |
+
"eval_wer": 0.17382999044890163,
|
| 289841 |
+
"step": 149708
|
| 289842 |
+
},
|
| 289843 |
+
{
|
| 289844 |
+
"epoch": 1207.02,
|
| 289845 |
+
"learning_rate": 7.598529886914378e-06,
|
| 289846 |
+
"loss": 0.3289,
|
| 289847 |
+
"step": 149710
|
| 289848 |
+
},
|
| 289849 |
+
{
|
| 289850 |
+
"epoch": 1207.06,
|
| 289851 |
+
"learning_rate": 7.598449111470114e-06,
|
| 289852 |
+
"loss": 0.258,
|
| 289853 |
+
"step": 149715
|
| 289854 |
+
},
|
| 289855 |
+
{
|
| 289856 |
+
"epoch": 1207.1,
|
| 289857 |
+
"learning_rate": 7.598368336025848e-06,
|
| 289858 |
+
"loss": 0.322,
|
| 289859 |
+
"step": 149720
|
| 289860 |
+
},
|
| 289861 |
+
{
|
| 289862 |
+
"epoch": 1207.14,
|
| 289863 |
+
"learning_rate": 7.598287560581584e-06,
|
| 289864 |
+
"loss": 0.377,
|
| 289865 |
+
"step": 149725
|
| 289866 |
+
},
|
| 289867 |
+
{
|
| 289868 |
+
"epoch": 1207.18,
|
| 289869 |
+
"learning_rate": 7.598206785137318e-06,
|
| 289870 |
+
"loss": 0.6225,
|
| 289871 |
+
"step": 149730
|
| 289872 |
+
},
|
| 289873 |
+
{
|
| 289874 |
+
"epoch": 1207.22,
|
| 289875 |
+
"learning_rate": 7.598126009693054e-06,
|
| 289876 |
+
"loss": 0.8557,
|
| 289877 |
+
"step": 149735
|
| 289878 |
+
},
|
| 289879 |
+
{
|
| 289880 |
+
"epoch": 1207.26,
|
| 289881 |
+
"learning_rate": 7.598045234248788e-06,
|
| 289882 |
+
"loss": 0.2603,
|
| 289883 |
+
"step": 149740
|
| 289884 |
+
},
|
| 289885 |
+
{
|
| 289886 |
+
"epoch": 1207.3,
|
| 289887 |
+
"learning_rate": 7.597964458804524e-06,
|
| 289888 |
+
"loss": 0.2477,
|
| 289889 |
+
"step": 149745
|
| 289890 |
+
},
|
| 289891 |
+
{
|
| 289892 |
+
"epoch": 1207.34,
|
| 289893 |
+
"learning_rate": 7.597883683360259e-06,
|
| 289894 |
+
"loss": 0.304,
|
| 289895 |
+
"step": 149750
|
| 289896 |
+
},
|
| 289897 |
+
{
|
| 289898 |
+
"epoch": 1207.38,
|
| 289899 |
+
"learning_rate": 7.597802907915994e-06,
|
| 289900 |
+
"loss": 0.6338,
|
| 289901 |
+
"step": 149755
|
| 289902 |
+
},
|
| 289903 |
+
{
|
| 289904 |
+
"epoch": 1207.42,
|
| 289905 |
+
"learning_rate": 7.597722132471729e-06,
|
| 289906 |
+
"loss": 0.9901,
|
| 289907 |
+
"step": 149760
|
| 289908 |
+
},
|
| 289909 |
+
{
|
| 289910 |
+
"epoch": 1207.46,
|
| 289911 |
+
"learning_rate": 7.597641357027464e-06,
|
| 289912 |
+
"loss": 0.3567,
|
| 289913 |
+
"step": 149765
|
| 289914 |
+
},
|
| 289915 |
+
{
|
| 289916 |
+
"epoch": 1207.5,
|
| 289917 |
+
"learning_rate": 7.5975605815832e-06,
|
| 289918 |
+
"loss": 0.2428,
|
| 289919 |
+
"step": 149770
|
| 289920 |
+
},
|
| 289921 |
+
{
|
| 289922 |
+
"epoch": 1207.54,
|
| 289923 |
+
"learning_rate": 7.597479806138934e-06,
|
| 289924 |
+
"loss": 0.3402,
|
| 289925 |
+
"step": 149775
|
| 289926 |
+
},
|
| 289927 |
+
{
|
| 289928 |
+
"epoch": 1207.58,
|
| 289929 |
+
"learning_rate": 7.59739903069467e-06,
|
| 289930 |
+
"loss": 0.4664,
|
| 289931 |
+
"step": 149780
|
| 289932 |
+
},
|
| 289933 |
+
{
|
| 289934 |
+
"epoch": 1207.62,
|
| 289935 |
+
"learning_rate": 7.597318255250404e-06,
|
| 289936 |
+
"loss": 0.8885,
|
| 289937 |
+
"step": 149785
|
| 289938 |
+
},
|
| 289939 |
+
{
|
| 289940 |
+
"epoch": 1207.66,
|
| 289941 |
+
"learning_rate": 7.59723747980614e-06,
|
| 289942 |
+
"loss": 0.2265,
|
| 289943 |
+
"step": 149790
|
| 289944 |
+
},
|
| 289945 |
+
{
|
| 289946 |
+
"epoch": 1207.7,
|
| 289947 |
+
"learning_rate": 7.597156704361874e-06,
|
| 289948 |
+
"loss": 0.3005,
|
| 289949 |
+
"step": 149795
|
| 289950 |
+
},
|
| 289951 |
+
{
|
| 289952 |
+
"epoch": 1207.74,
|
| 289953 |
+
"learning_rate": 7.59707592891761e-06,
|
| 289954 |
+
"loss": 0.3373,
|
| 289955 |
+
"step": 149800
|
| 289956 |
+
},
|
| 289957 |
+
{
|
| 289958 |
+
"epoch": 1207.78,
|
| 289959 |
+
"learning_rate": 7.596995153473344e-06,
|
| 289960 |
+
"loss": 0.6167,
|
| 289961 |
+
"step": 149805
|
| 289962 |
+
},
|
| 289963 |
+
{
|
| 289964 |
+
"epoch": 1207.82,
|
| 289965 |
+
"learning_rate": 7.59691437802908e-06,
|
| 289966 |
+
"loss": 0.9362,
|
| 289967 |
+
"step": 149810
|
| 289968 |
+
},
|
| 289969 |
+
{
|
| 289970 |
+
"epoch": 1207.86,
|
| 289971 |
+
"learning_rate": 7.596833602584815e-06,
|
| 289972 |
+
"loss": 0.2685,
|
| 289973 |
+
"step": 149815
|
| 289974 |
+
},
|
| 289975 |
+
{
|
| 289976 |
+
"epoch": 1207.9,
|
| 289977 |
+
"learning_rate": 7.59675282714055e-06,
|
| 289978 |
+
"loss": 0.275,
|
| 289979 |
+
"step": 149820
|
| 289980 |
+
},
|
| 289981 |
+
{
|
| 289982 |
+
"epoch": 1207.94,
|
| 289983 |
+
"learning_rate": 7.596672051696285e-06,
|
| 289984 |
+
"loss": 0.3284,
|
| 289985 |
+
"step": 149825
|
| 289986 |
+
},
|
| 289987 |
+
{
|
| 289988 |
+
"epoch": 1207.98,
|
| 289989 |
+
"learning_rate": 7.59659127625202e-06,
|
| 289990 |
+
"loss": 0.8461,
|
| 289991 |
+
"step": 149830
|
| 289992 |
+
},
|
| 289993 |
+
{
|
| 289994 |
+
"epoch": 1208.0,
|
| 289995 |
+
"eval_loss": 0.370609313249588,
|
| 289996 |
+
"eval_runtime": 41.783,
|
| 289997 |
+
"eval_samples_per_second": 20.104,
|
| 289998 |
+
"eval_steps_per_second": 0.646,
|
| 289999 |
+
"eval_wer": 0.17362270450751252,
|
| 290000 |
+
"step": 149832
|
| 290001 |
+
},
|
| 290002 |
+
{
|
| 290003 |
+
"epoch": 1208.02,
|
| 290004 |
+
"learning_rate": 7.596510500807756e-06,
|
| 290005 |
+
"loss": 0.2909,
|
| 290006 |
+
"step": 149835
|
| 290007 |
+
},
|
| 290008 |
+
{
|
| 290009 |
+
"epoch": 1208.06,
|
| 290010 |
+
"learning_rate": 7.59642972536349e-06,
|
| 290011 |
+
"loss": 0.2565,
|
| 290012 |
+
"step": 149840
|
| 290013 |
+
},
|
| 290014 |
+
{
|
| 290015 |
+
"epoch": 1208.1,
|
| 290016 |
+
"learning_rate": 7.596348949919226e-06,
|
| 290017 |
+
"loss": 0.2966,
|
| 290018 |
+
"step": 149845
|
| 290019 |
+
},
|
| 290020 |
+
{
|
| 290021 |
+
"epoch": 1208.14,
|
| 290022 |
+
"learning_rate": 7.59626817447496e-06,
|
| 290023 |
+
"loss": 0.3049,
|
| 290024 |
+
"step": 149850
|
| 290025 |
+
},
|
| 290026 |
+
{
|
| 290027 |
+
"epoch": 1208.18,
|
| 290028 |
+
"learning_rate": 7.5961873990306956e-06,
|
| 290029 |
+
"loss": 0.7274,
|
| 290030 |
+
"step": 149855
|
| 290031 |
+
},
|
| 290032 |
+
{
|
| 290033 |
+
"epoch": 1208.22,
|
| 290034 |
+
"learning_rate": 7.59610662358643e-06,
|
| 290035 |
+
"loss": 0.7799,
|
| 290036 |
+
"step": 149860
|
| 290037 |
+
},
|
| 290038 |
+
{
|
| 290039 |
+
"epoch": 1208.27,
|
| 290040 |
+
"learning_rate": 7.5960258481421655e-06,
|
| 290041 |
+
"loss": 0.2418,
|
| 290042 |
+
"step": 149865
|
| 290043 |
+
},
|
| 290044 |
+
{
|
| 290045 |
+
"epoch": 1208.31,
|
| 290046 |
+
"learning_rate": 7.5959450726979e-06,
|
| 290047 |
+
"loss": 0.2956,
|
| 290048 |
+
"step": 149870
|
| 290049 |
+
},
|
| 290050 |
+
{
|
| 290051 |
+
"epoch": 1208.35,
|
| 290052 |
+
"learning_rate": 7.5958642972536355e-06,
|
| 290053 |
+
"loss": 0.3961,
|
| 290054 |
+
"step": 149875
|
| 290055 |
+
},
|
| 290056 |
+
{
|
| 290057 |
+
"epoch": 1208.39,
|
| 290058 |
+
"learning_rate": 7.59578352180937e-06,
|
| 290059 |
+
"loss": 0.7961,
|
| 290060 |
+
"step": 149880
|
| 290061 |
+
},
|
| 290062 |
+
{
|
| 290063 |
+
"epoch": 1208.43,
|
| 290064 |
+
"learning_rate": 7.5957027463651055e-06,
|
| 290065 |
+
"loss": 0.88,
|
| 290066 |
+
"step": 149885
|
| 290067 |
+
},
|
| 290068 |
+
{
|
| 290069 |
+
"epoch": 1208.47,
|
| 290070 |
+
"learning_rate": 7.595621970920841e-06,
|
| 290071 |
+
"loss": 0.337,
|
| 290072 |
+
"step": 149890
|
| 290073 |
+
},
|
| 290074 |
+
{
|
| 290075 |
+
"epoch": 1208.51,
|
| 290076 |
+
"learning_rate": 7.5955411954765755e-06,
|
| 290077 |
+
"loss": 0.2649,
|
| 290078 |
+
"step": 149895
|
| 290079 |
+
},
|
| 290080 |
+
{
|
| 290081 |
+
"epoch": 1208.55,
|
| 290082 |
+
"learning_rate": 7.595460420032311e-06,
|
| 290083 |
+
"loss": 0.2962,
|
| 290084 |
+
"step": 149900
|
| 290085 |
+
},
|
| 290086 |
+
{
|
| 290087 |
+
"epoch": 1208.59,
|
| 290088 |
+
"learning_rate": 7.5953796445880454e-06,
|
| 290089 |
+
"loss": 0.7462,
|
| 290090 |
+
"step": 149905
|
| 290091 |
+
},
|
| 290092 |
+
{
|
| 290093 |
+
"epoch": 1208.63,
|
| 290094 |
+
"learning_rate": 7.595298869143781e-06,
|
| 290095 |
+
"loss": 1.0426,
|
| 290096 |
+
"step": 149910
|
| 290097 |
+
},
|
| 290098 |
+
{
|
| 290099 |
+
"epoch": 1208.67,
|
| 290100 |
+
"learning_rate": 7.595218093699515e-06,
|
| 290101 |
+
"loss": 0.2865,
|
| 290102 |
+
"step": 149915
|
| 290103 |
+
},
|
| 290104 |
+
{
|
| 290105 |
+
"epoch": 1208.71,
|
| 290106 |
+
"learning_rate": 7.595137318255251e-06,
|
| 290107 |
+
"loss": 0.2116,
|
| 290108 |
+
"step": 149920
|
| 290109 |
+
},
|
| 290110 |
+
{
|
| 290111 |
+
"epoch": 1208.75,
|
| 290112 |
+
"learning_rate": 7.595056542810985e-06,
|
| 290113 |
+
"loss": 0.4093,
|
| 290114 |
+
"step": 149925
|
| 290115 |
+
},
|
| 290116 |
+
{
|
| 290117 |
+
"epoch": 1208.79,
|
| 290118 |
+
"learning_rate": 7.594975767366721e-06,
|
| 290119 |
+
"loss": 0.7037,
|
| 290120 |
+
"step": 149930
|
| 290121 |
+
},
|
| 290122 |
+
{
|
| 290123 |
+
"epoch": 1208.83,
|
| 290124 |
+
"learning_rate": 7.594894991922455e-06,
|
| 290125 |
+
"loss": 0.8359,
|
| 290126 |
+
"step": 149935
|
| 290127 |
+
},
|
| 290128 |
+
{
|
| 290129 |
+
"epoch": 1208.87,
|
| 290130 |
+
"learning_rate": 7.594814216478191e-06,
|
| 290131 |
+
"loss": 0.2893,
|
| 290132 |
+
"step": 149940
|
| 290133 |
+
},
|
| 290134 |
+
{
|
| 290135 |
+
"epoch": 1208.91,
|
| 290136 |
+
"learning_rate": 7.594733441033927e-06,
|
| 290137 |
+
"loss": 0.293,
|
| 290138 |
+
"step": 149945
|
| 290139 |
+
},
|
| 290140 |
+
{
|
| 290141 |
+
"epoch": 1208.95,
|
| 290142 |
+
"learning_rate": 7.594652665589661e-06,
|
| 290143 |
+
"loss": 0.3372,
|
| 290144 |
+
"step": 149950
|
| 290145 |
+
},
|
| 290146 |
+
{
|
| 290147 |
+
"epoch": 1208.99,
|
| 290148 |
+
"learning_rate": 7.594571890145397e-06,
|
| 290149 |
+
"loss": 0.8728,
|
| 290150 |
+
"step": 149955
|
| 290151 |
+
},
|
| 290152 |
+
{
|
| 290153 |
+
"epoch": 1209.0,
|
| 290154 |
+
"eval_loss": 0.3431926369667053,
|
| 290155 |
+
"eval_runtime": 42.3561,
|
| 290156 |
+
"eval_samples_per_second": 19.855,
|
| 290157 |
+
"eval_steps_per_second": 0.637,
|
| 290158 |
+
"eval_wer": 0.17415083477259644,
|
| 290159 |
+
"step": 149956
|
| 290160 |
+
},
|
| 290161 |
+
{
|
| 290162 |
+
"epoch": 1199.03,
|
| 290163 |
+
"learning_rate": 7.594491114701131e-06,
|
| 290164 |
+
"loss": 0.2865,
|
| 290165 |
+
"step": 149960
|
| 290166 |
+
},
|
| 290167 |
+
{
|
| 290168 |
+
"epoch": 1199.07,
|
| 290169 |
+
"learning_rate": 7.594410339256867e-06,
|
| 290170 |
+
"loss": 0.324,
|
| 290171 |
+
"step": 149965
|
| 290172 |
+
},
|
| 290173 |
+
{
|
| 290174 |
+
"epoch": 1199.11,
|
| 290175 |
+
"learning_rate": 7.594329563812601e-06,
|
| 290176 |
+
"loss": 0.3592,
|
| 290177 |
+
"step": 149970
|
| 290178 |
+
},
|
| 290179 |
+
{
|
| 290180 |
+
"epoch": 1199.15,
|
| 290181 |
+
"learning_rate": 7.594248788368337e-06,
|
| 290182 |
+
"loss": 0.4291,
|
| 290183 |
+
"step": 149975
|
| 290184 |
+
},
|
| 290185 |
+
{
|
| 290186 |
+
"epoch": 1199.19,
|
| 290187 |
+
"learning_rate": 7.594168012924071e-06,
|
| 290188 |
+
"loss": 0.8308,
|
| 290189 |
+
"step": 149980
|
| 290190 |
+
},
|
| 290191 |
+
{
|
| 290192 |
+
"epoch": 1199.23,
|
| 290193 |
+
"learning_rate": 7.594087237479807e-06,
|
| 290194 |
+
"loss": 0.6,
|
| 290195 |
+
"step": 149985
|
| 290196 |
+
},
|
| 290197 |
+
{
|
| 290198 |
+
"epoch": 1199.27,
|
| 290199 |
+
"learning_rate": 7.594006462035541e-06,
|
| 290200 |
+
"loss": 0.2941,
|
| 290201 |
+
"step": 149990
|
| 290202 |
+
},
|
| 290203 |
+
{
|
| 290204 |
+
"epoch": 1199.31,
|
| 290205 |
+
"learning_rate": 7.593925686591277e-06,
|
| 290206 |
+
"loss": 0.2726,
|
| 290207 |
+
"step": 149995
|
| 290208 |
+
},
|
| 290209 |
+
{
|
| 290210 |
+
"epoch": 1199.35,
|
| 290211 |
+
"learning_rate": 7.593844911147011e-06,
|
| 290212 |
+
"loss": 0.349,
|
| 290213 |
+
"step": 150000
|
| 290214 |
+
},
|
| 290215 |
+
{
|
| 290216 |
+
"epoch": 1199.39,
|
| 290217 |
+
"learning_rate": 7.593764135702747e-06,
|
| 290218 |
+
"loss": 0.8815,
|
| 290219 |
+
"step": 150005
|
| 290220 |
+
},
|
| 290221 |
+
{
|
| 290222 |
+
"epoch": 1199.43,
|
| 290223 |
+
"learning_rate": 7.593683360258483e-06,
|
| 290224 |
+
"loss": 0.7146,
|
| 290225 |
+
"step": 150010
|
| 290226 |
+
},
|
| 290227 |
+
{
|
| 290228 |
+
"epoch": 1199.47,
|
| 290229 |
+
"learning_rate": 7.593602584814217e-06,
|
| 290230 |
+
"loss": 0.2515,
|
| 290231 |
+
"step": 150015
|
| 290232 |
+
},
|
| 290233 |
+
{
|
| 290234 |
+
"epoch": 1199.51,
|
| 290235 |
+
"learning_rate": 7.593521809369953e-06,
|
| 290236 |
+
"loss": 0.271,
|
| 290237 |
+
"step": 150020
|
| 290238 |
+
},
|
| 290239 |
+
{
|
| 290240 |
+
"epoch": 1199.55,
|
| 290241 |
+
"learning_rate": 7.593441033925687e-06,
|
| 290242 |
+
"loss": 0.3378,
|
| 290243 |
+
"step": 150025
|
| 290244 |
+
},
|
| 290245 |
+
{
|
| 290246 |
+
"epoch": 1199.59,
|
| 290247 |
+
"learning_rate": 7.593360258481423e-06,
|
| 290248 |
+
"loss": 0.8722,
|
| 290249 |
+
"step": 150030
|
| 290250 |
+
},
|
| 290251 |
+
{
|
| 290252 |
+
"epoch": 1199.63,
|
| 290253 |
+
"learning_rate": 7.593279483037157e-06,
|
| 290254 |
+
"loss": 0.5991,
|
| 290255 |
+
"step": 150035
|
| 290256 |
+
},
|
| 290257 |
+
{
|
| 290258 |
+
"epoch": 1199.67,
|
| 290259 |
+
"learning_rate": 7.593198707592893e-06,
|
| 290260 |
+
"loss": 0.3126,
|
| 290261 |
+
"step": 150040
|
| 290262 |
+
},
|
| 290263 |
+
{
|
| 290264 |
+
"epoch": 1199.71,
|
| 290265 |
+
"learning_rate": 7.593117932148627e-06,
|
| 290266 |
+
"loss": 0.2889,
|
| 290267 |
+
"step": 150045
|
| 290268 |
+
},
|
| 290269 |
+
{
|
| 290270 |
+
"epoch": 1199.75,
|
| 290271 |
+
"learning_rate": 7.593037156704363e-06,
|
| 290272 |
+
"loss": 0.4038,
|
| 290273 |
+
"step": 150050
|
| 290274 |
+
},
|
| 290275 |
+
{
|
| 290276 |
+
"epoch": 1199.79,
|
| 290277 |
+
"learning_rate": 7.592956381260097e-06,
|
| 290278 |
+
"loss": 0.9201,
|
| 290279 |
+
"step": 150055
|
| 290280 |
+
},
|
| 290281 |
+
{
|
| 290282 |
+
"epoch": 1199.83,
|
| 290283 |
+
"learning_rate": 7.592875605815833e-06,
|
| 290284 |
+
"loss": 0.6433,
|
| 290285 |
+
"step": 150060
|
| 290286 |
+
},
|
| 290287 |
+
{
|
| 290288 |
+
"epoch": 1199.87,
|
| 290289 |
+
"learning_rate": 7.592794830371568e-06,
|
| 290290 |
+
"loss": 0.2426,
|
| 290291 |
+
"step": 150065
|
| 290292 |
+
},
|
| 290293 |
+
{
|
| 290294 |
+
"epoch": 1199.91,
|
| 290295 |
+
"learning_rate": 7.5927140549273026e-06,
|
| 290296 |
+
"loss": 0.2838,
|
| 290297 |
+
"step": 150070
|
| 290298 |
+
},
|
| 290299 |
+
{
|
| 290300 |
+
"epoch": 1199.95,
|
| 290301 |
+
"learning_rate": 7.592633279483038e-06,
|
| 290302 |
+
"loss": 0.3511,
|
| 290303 |
+
"step": 150075
|
| 290304 |
+
},
|
| 290305 |
+
{
|
| 290306 |
+
"epoch": 1199.99,
|
| 290307 |
+
"learning_rate": 7.5925525040387725e-06,
|
| 290308 |
+
"loss": 1.0652,
|
| 290309 |
+
"step": 150080
|
| 290310 |
+
},
|
| 290311 |
+
{
|
| 290312 |
+
"epoch": 1200.0,
|
| 290313 |
+
"eval_loss": 0.38891738653182983,
|
| 290314 |
+
"eval_runtime": 42.3061,
|
| 290315 |
+
"eval_samples_per_second": 19.879,
|
| 290316 |
+
"eval_steps_per_second": 0.638,
|
| 290317 |
+
"eval_wer": 0.1781133181851499,
|
| 290318 |
+
"step": 150081
|
| 290319 |
+
},
|
| 290320 |
+
{
|
| 290321 |
+
"epoch": 1210.03,
|
| 290322 |
+
"learning_rate": 7.592471728594508e-06,
|
| 290323 |
+
"loss": 0.2999,
|
| 290324 |
+
"step": 150085
|
| 290325 |
+
},
|
| 290326 |
+
{
|
| 290327 |
+
"epoch": 1210.07,
|
| 290328 |
+
"learning_rate": 7.5923909531502425e-06,
|
| 290329 |
+
"loss": 0.2679,
|
| 290330 |
+
"step": 150090
|
| 290331 |
+
},
|
| 290332 |
+
{
|
| 290333 |
+
"epoch": 1210.11,
|
| 290334 |
+
"learning_rate": 7.592310177705978e-06,
|
| 290335 |
+
"loss": 0.2975,
|
| 290336 |
+
"step": 150095
|
| 290337 |
+
},
|
| 290338 |
+
{
|
| 290339 |
+
"epoch": 1210.15,
|
| 290340 |
+
"learning_rate": 7.5922294022617125e-06,
|
| 290341 |
+
"loss": 0.3568,
|
| 290342 |
+
"step": 150100
|
| 290343 |
+
},
|
| 290344 |
+
{
|
| 290345 |
+
"epoch": 1210.19,
|
| 290346 |
+
"learning_rate": 7.592148626817448e-06,
|
| 290347 |
+
"loss": 0.8083,
|
| 290348 |
+
"step": 150105
|
| 290349 |
+
},
|
| 290350 |
+
{
|
| 290351 |
+
"epoch": 1210.23,
|
| 290352 |
+
"learning_rate": 7.5920678513731825e-06,
|
| 290353 |
+
"loss": 0.6369,
|
| 290354 |
+
"step": 150110
|
| 290355 |
+
},
|
| 290356 |
+
{
|
| 290357 |
+
"epoch": 1210.27,
|
| 290358 |
+
"learning_rate": 7.591987075928918e-06,
|
| 290359 |
+
"loss": 0.2588,
|
| 290360 |
+
"step": 150115
|
| 290361 |
+
},
|
| 290362 |
+
{
|
| 290363 |
+
"epoch": 1210.31,
|
| 290364 |
+
"learning_rate": 7.591906300484654e-06,
|
| 290365 |
+
"loss": 0.2907,
|
| 290366 |
+
"step": 150120
|
| 290367 |
+
},
|
| 290368 |
+
{
|
| 290369 |
+
"epoch": 1210.35,
|
| 290370 |
+
"learning_rate": 7.591825525040388e-06,
|
| 290371 |
+
"loss": 0.332,
|
| 290372 |
+
"step": 150125
|
| 290373 |
+
},
|
| 290374 |
+
{
|
| 290375 |
+
"epoch": 1210.4,
|
| 290376 |
+
"learning_rate": 7.591744749596124e-06,
|
| 290377 |
+
"loss": 0.8868,
|
| 290378 |
+
"step": 150130
|
| 290379 |
+
},
|
| 290380 |
+
{
|
| 290381 |
+
"epoch": 1210.44,
|
| 290382 |
+
"learning_rate": 7.591663974151858e-06,
|
| 290383 |
+
"loss": 0.5735,
|
| 290384 |
+
"step": 150135
|
| 290385 |
+
},
|
| 290386 |
+
{
|
| 290387 |
+
"epoch": 1210.48,
|
| 290388 |
+
"learning_rate": 7.591583198707594e-06,
|
| 290389 |
+
"loss": 0.2798,
|
| 290390 |
+
"step": 150140
|
| 290391 |
+
},
|
| 290392 |
+
{
|
| 290393 |
+
"epoch": 1210.52,
|
| 290394 |
+
"learning_rate": 7.591502423263328e-06,
|
| 290395 |
+
"loss": 0.3032,
|
| 290396 |
+
"step": 150145
|
| 290397 |
+
},
|
| 290398 |
+
{
|
| 290399 |
+
"epoch": 1210.56,
|
| 290400 |
+
"learning_rate": 7.591421647819064e-06,
|
| 290401 |
+
"loss": 0.42,
|
| 290402 |
+
"step": 150150
|
| 290403 |
+
},
|
| 290404 |
+
{
|
| 290405 |
+
"epoch": 1210.6,
|
| 290406 |
+
"learning_rate": 7.591340872374798e-06,
|
| 290407 |
+
"loss": 0.8859,
|
| 290408 |
+
"step": 150155
|
| 290409 |
+
},
|
| 290410 |
+
{
|
| 290411 |
+
"epoch": 1210.64,
|
| 290412 |
+
"learning_rate": 7.591260096930534e-06,
|
| 290413 |
+
"loss": 0.593,
|
| 290414 |
+
"step": 150160
|
| 290415 |
+
},
|
| 290416 |
+
{
|
| 290417 |
+
"epoch": 1210.68,
|
| 290418 |
+
"learning_rate": 7.591179321486268e-06,
|
| 290419 |
+
"loss": 0.2568,
|
| 290420 |
+
"step": 150165
|
| 290421 |
+
},
|
| 290422 |
+
{
|
| 290423 |
+
"epoch": 1210.72,
|
| 290424 |
+
"learning_rate": 7.591098546042004e-06,
|
| 290425 |
+
"loss": 0.295,
|
| 290426 |
+
"step": 150170
|
| 290427 |
+
},
|
| 290428 |
+
{
|
| 290429 |
+
"epoch": 1210.76,
|
| 290430 |
+
"learning_rate": 7.591017770597738e-06,
|
| 290431 |
+
"loss": 0.3517,
|
| 290432 |
+
"step": 150175
|
| 290433 |
+
},
|
| 290434 |
+
{
|
| 290435 |
+
"epoch": 1210.8,
|
| 290436 |
+
"learning_rate": 7.590936995153474e-06,
|
| 290437 |
+
"loss": 0.9385,
|
| 290438 |
+
"step": 150180
|
| 290439 |
+
},
|
| 290440 |
+
{
|
| 290441 |
+
"epoch": 1210.84,
|
| 290442 |
+
"learning_rate": 7.59085621970921e-06,
|
| 290443 |
+
"loss": 0.6471,
|
| 290444 |
+
"step": 150185
|
| 290445 |
+
},
|
| 290446 |
+
{
|
| 290447 |
+
"epoch": 1210.88,
|
| 290448 |
+
"learning_rate": 7.590775444264944e-06,
|
| 290449 |
+
"loss": 0.2162,
|
| 290450 |
+
"step": 150190
|
| 290451 |
+
},
|
| 290452 |
+
{
|
| 290453 |
+
"epoch": 1210.92,
|
| 290454 |
+
"learning_rate": 7.59069466882068e-06,
|
| 290455 |
+
"loss": 0.2966,
|
| 290456 |
+
"step": 150195
|
| 290457 |
+
},
|
| 290458 |
+
{
|
| 290459 |
+
"epoch": 1210.96,
|
| 290460 |
+
"learning_rate": 7.590613893376414e-06,
|
| 290461 |
+
"loss": 0.4473,
|
| 290462 |
+
"step": 150200
|
| 290463 |
+
},
|
| 290464 |
+
{
|
| 290465 |
+
"epoch": 1211.0,
|
| 290466 |
+
"learning_rate": 7.59053311793215e-06,
|
| 290467 |
+
"loss": 1.2014,
|
| 290468 |
+
"step": 150205
|
| 290469 |
+
},
|
| 290470 |
+
{
|
| 290471 |
+
"epoch": 1211.0,
|
| 290472 |
+
"eval_loss": 0.3155873119831085,
|
| 290473 |
+
"eval_runtime": 41.9259,
|
| 290474 |
+
"eval_samples_per_second": 20.059,
|
| 290475 |
+
"eval_steps_per_second": 0.644,
|
| 290476 |
+
"eval_wer": 0.17267648552564754,
|
| 290477 |
+
"step": 150205
|
| 290478 |
}
|
| 290479 |
],
|
| 290480 |
+
"max_steps": 620000,
|
| 290481 |
"num_train_epochs": 5000,
|
| 290482 |
+
"total_flos": 4.226898230914657e+20,
|
| 290483 |
"trial_name": null,
|
| 290484 |
"trial_params": null
|
| 290485 |
}
|
model-bin/finetune/base/{checkpoint-149584 β checkpoint-150205}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1630241766.1674721/events.out.tfevents.1630241766.cc93b136ebf5.1086.169
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3a8e839a63f1044cdd3733e6d1902b9a273438784bee6f325e386f0b3586fc66
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630242210.9386408/events.out.tfevents.1630242210.cc93b136ebf5.1086.171
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:84cec864ad98be8751dd931572737d3082faa656b62041daab2c9a739ce3b6ea
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630242652.9526577/events.out.tfevents.1630242652.cc93b136ebf5.1086.173
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9f065a191ea7a66fef5f98843dd7ebd9e1d383540d5907ac328552393f3bf6ea
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630243093.5265884/events.out.tfevents.1630243093.cc93b136ebf5.1086.175
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:98cb2b24b76763ede43b591c8a759dd0542a095509a8f38257e99523b07e9bef
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630243533.7269483/events.out.tfevents.1630243533.cc93b136ebf5.1086.177
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0bf56e0a8edf9659a6ca25eee11f9791f25f55e4e3afbfee3befb4f9cc0dc0d2
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630241766.cc93b136ebf5.1086.168
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5d37d442dac7faeb1f442af710999b3add3e950fda2f57ffd9ffc909c07dd4b3
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630242210.cc93b136ebf5.1086.170
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bac685984e447ab85c1e5e90ffaf8956acfa0a048401576ac463f1c896e28726
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630242652.cc93b136ebf5.1086.172
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1469d3ae4ad89e67e6efdb48f9759769145fb48aedebdc5644e6d86e7f741cc7
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630243093.cc93b136ebf5.1086.174
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:088b40b88599462f943492215da07c29f168d2b553b786062366b001377e6f31
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630243533.cc93b136ebf5.1086.176
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c77cd55108a8f9d2cb1905ea48e27f7e0979770117b25faa9ec77ac6e38f6997
|
| 3 |
+
size 8622
|