"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/trainer_state.json +799 -4
- model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630250753.668995/events.out.tfevents.1630250753.cc93b136ebf5.1086.209 +3 -0
- model-bin/finetune/base/log/1630251188.88167/events.out.tfevents.1630251188.cc93b136ebf5.1086.211 +3 -0
- model-bin/finetune/base/log/1630251628.1797204/events.out.tfevents.1630251628.cc93b136ebf5.1086.213 +3 -0
- model-bin/finetune/base/log/1630252065.7514985/events.out.tfevents.1630252065.cc93b136ebf5.1086.215 +3 -0
- model-bin/finetune/base/log/1630252500.5741549/events.out.tfevents.1630252500.cc93b136ebf5.1086.217 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630250753.cc93b136ebf5.1086.208 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630251188.cc93b136ebf5.1086.210 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630251628.cc93b136ebf5.1086.212 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630252065.cc93b136ebf5.1086.214 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630252500.cc93b136ebf5.1086.216 +3 -0
model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:013e7596c9833c0f02cb6d858875c8b42145994a60410ff7acc22c55ed3a3f78
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:358bbb0a31f3eada3dcf3518c389f01e7260b174069b7a039e533c23adcbc695
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c64726b931c7ed69c0aa3b500fe65f0f56bb20ca55efd46d255f359a9d38051f
|
| 3 |
size 14503
|
model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e38d467b638b9bb9004c2051e9035f041fc23da780dcb2db9c2dbc1bb4f1e23d
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1740c53bf7ef796ec727899064cf449f45e9e8026922bf22f64ed1a0b45bd32d
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1668011442822563,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -292842,11 +292842,806 @@
|
|
| 292842 |
"eval_steps_per_second": 0.667,
|
| 292843 |
"eval_wer": 0.1764970733903647,
|
| 292844 |
"step": 152068
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 292845 |
}
|
| 292846 |
],
|
| 292847 |
-
"max_steps":
|
| 292848 |
"num_train_epochs": 5000,
|
| 292849 |
-
"total_flos": 4.
|
| 292850 |
"trial_name": null,
|
| 292851 |
"trial_params": null
|
| 292852 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1668011442822563,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
| 4 |
+
"epoch": 1221.0,
|
| 5 |
+
"global_step": 152690,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 292842 |
"eval_steps_per_second": 0.667,
|
| 292843 |
"eval_wer": 0.1764970733903647,
|
| 292844 |
"step": 152068
|
| 292845 |
+
},
|
| 292846 |
+
{
|
| 292847 |
+
"epoch": 1226.02,
|
| 292848 |
+
"learning_rate": 7.560403877221325e-06,
|
| 292849 |
+
"loss": 0.3566,
|
| 292850 |
+
"step": 152070
|
| 292851 |
+
},
|
| 292852 |
+
{
|
| 292853 |
+
"epoch": 1226.06,
|
| 292854 |
+
"learning_rate": 7.5603231017770605e-06,
|
| 292855 |
+
"loss": 0.2873,
|
| 292856 |
+
"step": 152075
|
| 292857 |
+
},
|
| 292858 |
+
{
|
| 292859 |
+
"epoch": 1226.1,
|
| 292860 |
+
"learning_rate": 7.5602423263327955e-06,
|
| 292861 |
+
"loss": 0.2621,
|
| 292862 |
+
"step": 152080
|
| 292863 |
+
},
|
| 292864 |
+
{
|
| 292865 |
+
"epoch": 1226.14,
|
| 292866 |
+
"learning_rate": 7.5601615508885305e-06,
|
| 292867 |
+
"loss": 0.3207,
|
| 292868 |
+
"step": 152085
|
| 292869 |
+
},
|
| 292870 |
+
{
|
| 292871 |
+
"epoch": 1226.18,
|
| 292872 |
+
"learning_rate": 7.5600807754442655e-06,
|
| 292873 |
+
"loss": 0.7188,
|
| 292874 |
+
"step": 152090
|
| 292875 |
+
},
|
| 292876 |
+
{
|
| 292877 |
+
"epoch": 1226.22,
|
| 292878 |
+
"learning_rate": 7.5600000000000005e-06,
|
| 292879 |
+
"loss": 1.0223,
|
| 292880 |
+
"step": 152095
|
| 292881 |
+
},
|
| 292882 |
+
{
|
| 292883 |
+
"epoch": 1226.26,
|
| 292884 |
+
"learning_rate": 7.5599192245557355e-06,
|
| 292885 |
+
"loss": 0.2622,
|
| 292886 |
+
"step": 152100
|
| 292887 |
+
},
|
| 292888 |
+
{
|
| 292889 |
+
"epoch": 1226.3,
|
| 292890 |
+
"learning_rate": 7.5598384491114704e-06,
|
| 292891 |
+
"loss": 0.2759,
|
| 292892 |
+
"step": 152105
|
| 292893 |
+
},
|
| 292894 |
+
{
|
| 292895 |
+
"epoch": 1226.34,
|
| 292896 |
+
"learning_rate": 7.559757673667206e-06,
|
| 292897 |
+
"loss": 0.3349,
|
| 292898 |
+
"step": 152110
|
| 292899 |
+
},
|
| 292900 |
+
{
|
| 292901 |
+
"epoch": 1226.38,
|
| 292902 |
+
"learning_rate": 7.55967689822294e-06,
|
| 292903 |
+
"loss": 0.5856,
|
| 292904 |
+
"step": 152115
|
| 292905 |
+
},
|
| 292906 |
+
{
|
| 292907 |
+
"epoch": 1226.42,
|
| 292908 |
+
"learning_rate": 7.559596122778676e-06,
|
| 292909 |
+
"loss": 0.9452,
|
| 292910 |
+
"step": 152120
|
| 292911 |
+
},
|
| 292912 |
+
{
|
| 292913 |
+
"epoch": 1226.46,
|
| 292914 |
+
"learning_rate": 7.55951534733441e-06,
|
| 292915 |
+
"loss": 0.2844,
|
| 292916 |
+
"step": 152125
|
| 292917 |
+
},
|
| 292918 |
+
{
|
| 292919 |
+
"epoch": 1226.5,
|
| 292920 |
+
"learning_rate": 7.559434571890146e-06,
|
| 292921 |
+
"loss": 0.2637,
|
| 292922 |
+
"step": 152130
|
| 292923 |
+
},
|
| 292924 |
+
{
|
| 292925 |
+
"epoch": 1226.54,
|
| 292926 |
+
"learning_rate": 7.55935379644588e-06,
|
| 292927 |
+
"loss": 0.2986,
|
| 292928 |
+
"step": 152135
|
| 292929 |
+
},
|
| 292930 |
+
{
|
| 292931 |
+
"epoch": 1226.58,
|
| 292932 |
+
"learning_rate": 7.559273021001616e-06,
|
| 292933 |
+
"loss": 0.6334,
|
| 292934 |
+
"step": 152140
|
| 292935 |
+
},
|
| 292936 |
+
{
|
| 292937 |
+
"epoch": 1226.62,
|
| 292938 |
+
"learning_rate": 7.55919224555735e-06,
|
| 292939 |
+
"loss": 0.9182,
|
| 292940 |
+
"step": 152145
|
| 292941 |
+
},
|
| 292942 |
+
{
|
| 292943 |
+
"epoch": 1226.66,
|
| 292944 |
+
"learning_rate": 7.559111470113086e-06,
|
| 292945 |
+
"loss": 0.2254,
|
| 292946 |
+
"step": 152150
|
| 292947 |
+
},
|
| 292948 |
+
{
|
| 292949 |
+
"epoch": 1226.7,
|
| 292950 |
+
"learning_rate": 7.559030694668821e-06,
|
| 292951 |
+
"loss": 0.2388,
|
| 292952 |
+
"step": 152155
|
| 292953 |
+
},
|
| 292954 |
+
{
|
| 292955 |
+
"epoch": 1226.74,
|
| 292956 |
+
"learning_rate": 7.558949919224556e-06,
|
| 292957 |
+
"loss": 0.3118,
|
| 292958 |
+
"step": 152160
|
| 292959 |
+
},
|
| 292960 |
+
{
|
| 292961 |
+
"epoch": 1226.78,
|
| 292962 |
+
"learning_rate": 7.558869143780292e-06,
|
| 292963 |
+
"loss": 0.6253,
|
| 292964 |
+
"step": 152165
|
| 292965 |
+
},
|
| 292966 |
+
{
|
| 292967 |
+
"epoch": 1226.82,
|
| 292968 |
+
"learning_rate": 7.558788368336026e-06,
|
| 292969 |
+
"loss": 0.9901,
|
| 292970 |
+
"step": 152170
|
| 292971 |
+
},
|
| 292972 |
+
{
|
| 292973 |
+
"epoch": 1226.86,
|
| 292974 |
+
"learning_rate": 7.558707592891762e-06,
|
| 292975 |
+
"loss": 0.2865,
|
| 292976 |
+
"step": 152175
|
| 292977 |
+
},
|
| 292978 |
+
{
|
| 292979 |
+
"epoch": 1226.9,
|
| 292980 |
+
"learning_rate": 7.558626817447496e-06,
|
| 292981 |
+
"loss": 0.3002,
|
| 292982 |
+
"step": 152180
|
| 292983 |
+
},
|
| 292984 |
+
{
|
| 292985 |
+
"epoch": 1226.94,
|
| 292986 |
+
"learning_rate": 7.558546042003232e-06,
|
| 292987 |
+
"loss": 0.3956,
|
| 292988 |
+
"step": 152185
|
| 292989 |
+
},
|
| 292990 |
+
{
|
| 292991 |
+
"epoch": 1226.98,
|
| 292992 |
+
"learning_rate": 7.558465266558966e-06,
|
| 292993 |
+
"loss": 0.687,
|
| 292994 |
+
"step": 152190
|
| 292995 |
+
},
|
| 292996 |
+
{
|
| 292997 |
+
"epoch": 1227.0,
|
| 292998 |
+
"eval_loss": 0.39228349924087524,
|
| 292999 |
+
"eval_runtime": 41.5829,
|
| 293000 |
+
"eval_samples_per_second": 20.104,
|
| 293001 |
+
"eval_steps_per_second": 0.649,
|
| 293002 |
+
"eval_wer": 0.18111669214530102,
|
| 293003 |
+
"step": 152192
|
| 293004 |
+
},
|
| 293005 |
+
{
|
| 293006 |
+
"epoch": 1227.02,
|
| 293007 |
+
"learning_rate": 7.558384491114702e-06,
|
| 293008 |
+
"loss": 0.2944,
|
| 293009 |
+
"step": 152195
|
| 293010 |
+
},
|
| 293011 |
+
{
|
| 293012 |
+
"epoch": 1227.06,
|
| 293013 |
+
"learning_rate": 7.558303715670436e-06,
|
| 293014 |
+
"loss": 0.2883,
|
| 293015 |
+
"step": 152200
|
| 293016 |
+
},
|
| 293017 |
+
{
|
| 293018 |
+
"epoch": 1227.1,
|
| 293019 |
+
"learning_rate": 7.558222940226172e-06,
|
| 293020 |
+
"loss": 0.299,
|
| 293021 |
+
"step": 152205
|
| 293022 |
+
},
|
| 293023 |
+
{
|
| 293024 |
+
"epoch": 1227.14,
|
| 293025 |
+
"learning_rate": 7.558142164781906e-06,
|
| 293026 |
+
"loss": 0.2715,
|
| 293027 |
+
"step": 152210
|
| 293028 |
+
},
|
| 293029 |
+
{
|
| 293030 |
+
"epoch": 1227.18,
|
| 293031 |
+
"learning_rate": 7.558061389337642e-06,
|
| 293032 |
+
"loss": 0.7032,
|
| 293033 |
+
"step": 152215
|
| 293034 |
+
},
|
| 293035 |
+
{
|
| 293036 |
+
"epoch": 1227.22,
|
| 293037 |
+
"learning_rate": 7.557980613893378e-06,
|
| 293038 |
+
"loss": 0.889,
|
| 293039 |
+
"step": 152220
|
| 293040 |
+
},
|
| 293041 |
+
{
|
| 293042 |
+
"epoch": 1227.27,
|
| 293043 |
+
"learning_rate": 7.557899838449112e-06,
|
| 293044 |
+
"loss": 0.2777,
|
| 293045 |
+
"step": 152225
|
| 293046 |
+
},
|
| 293047 |
+
{
|
| 293048 |
+
"epoch": 1227.31,
|
| 293049 |
+
"learning_rate": 7.557819063004848e-06,
|
| 293050 |
+
"loss": 0.3006,
|
| 293051 |
+
"step": 152230
|
| 293052 |
+
},
|
| 293053 |
+
{
|
| 293054 |
+
"epoch": 1227.35,
|
| 293055 |
+
"learning_rate": 7.557738287560582e-06,
|
| 293056 |
+
"loss": 0.3308,
|
| 293057 |
+
"step": 152235
|
| 293058 |
+
},
|
| 293059 |
+
{
|
| 293060 |
+
"epoch": 1227.39,
|
| 293061 |
+
"learning_rate": 7.557657512116318e-06,
|
| 293062 |
+
"loss": 0.7281,
|
| 293063 |
+
"step": 152240
|
| 293064 |
+
},
|
| 293065 |
+
{
|
| 293066 |
+
"epoch": 1227.43,
|
| 293067 |
+
"learning_rate": 7.557576736672052e-06,
|
| 293068 |
+
"loss": 0.8714,
|
| 293069 |
+
"step": 152245
|
| 293070 |
+
},
|
| 293071 |
+
{
|
| 293072 |
+
"epoch": 1227.47,
|
| 293073 |
+
"learning_rate": 7.557495961227788e-06,
|
| 293074 |
+
"loss": 0.2962,
|
| 293075 |
+
"step": 152250
|
| 293076 |
+
},
|
| 293077 |
+
{
|
| 293078 |
+
"epoch": 1227.51,
|
| 293079 |
+
"learning_rate": 7.557415185783522e-06,
|
| 293080 |
+
"loss": 0.2503,
|
| 293081 |
+
"step": 152255
|
| 293082 |
+
},
|
| 293083 |
+
{
|
| 293084 |
+
"epoch": 1227.55,
|
| 293085 |
+
"learning_rate": 7.557334410339258e-06,
|
| 293086 |
+
"loss": 0.3374,
|
| 293087 |
+
"step": 152260
|
| 293088 |
+
},
|
| 293089 |
+
{
|
| 293090 |
+
"epoch": 1227.59,
|
| 293091 |
+
"learning_rate": 7.557253634894992e-06,
|
| 293092 |
+
"loss": 0.6683,
|
| 293093 |
+
"step": 152265
|
| 293094 |
+
},
|
| 293095 |
+
{
|
| 293096 |
+
"epoch": 1227.63,
|
| 293097 |
+
"learning_rate": 7.5571728594507276e-06,
|
| 293098 |
+
"loss": 0.7008,
|
| 293099 |
+
"step": 152270
|
| 293100 |
+
},
|
| 293101 |
+
{
|
| 293102 |
+
"epoch": 1227.67,
|
| 293103 |
+
"learning_rate": 7.557092084006462e-06,
|
| 293104 |
+
"loss": 0.3346,
|
| 293105 |
+
"step": 152275
|
| 293106 |
+
},
|
| 293107 |
+
{
|
| 293108 |
+
"epoch": 1227.71,
|
| 293109 |
+
"learning_rate": 7.5570113085621975e-06,
|
| 293110 |
+
"loss": 0.2314,
|
| 293111 |
+
"step": 152280
|
| 293112 |
+
},
|
| 293113 |
+
{
|
| 293114 |
+
"epoch": 1227.75,
|
| 293115 |
+
"learning_rate": 7.556930533117933e-06,
|
| 293116 |
+
"loss": 0.4819,
|
| 293117 |
+
"step": 152285
|
| 293118 |
+
},
|
| 293119 |
+
{
|
| 293120 |
+
"epoch": 1227.79,
|
| 293121 |
+
"learning_rate": 7.5568497576736675e-06,
|
| 293122 |
+
"loss": 0.5945,
|
| 293123 |
+
"step": 152290
|
| 293124 |
+
},
|
| 293125 |
+
{
|
| 293126 |
+
"epoch": 1227.83,
|
| 293127 |
+
"learning_rate": 7.556768982229403e-06,
|
| 293128 |
+
"loss": 0.8223,
|
| 293129 |
+
"step": 152295
|
| 293130 |
+
},
|
| 293131 |
+
{
|
| 293132 |
+
"epoch": 1227.87,
|
| 293133 |
+
"learning_rate": 7.5566882067851375e-06,
|
| 293134 |
+
"loss": 0.2747,
|
| 293135 |
+
"step": 152300
|
| 293136 |
+
},
|
| 293137 |
+
{
|
| 293138 |
+
"epoch": 1227.91,
|
| 293139 |
+
"learning_rate": 7.556607431340873e-06,
|
| 293140 |
+
"loss": 0.3093,
|
| 293141 |
+
"step": 152305
|
| 293142 |
+
},
|
| 293143 |
+
{
|
| 293144 |
+
"epoch": 1227.95,
|
| 293145 |
+
"learning_rate": 7.5565266558966075e-06,
|
| 293146 |
+
"loss": 0.4192,
|
| 293147 |
+
"step": 152310
|
| 293148 |
+
},
|
| 293149 |
+
{
|
| 293150 |
+
"epoch": 1227.99,
|
| 293151 |
+
"learning_rate": 7.556445880452343e-06,
|
| 293152 |
+
"loss": 0.7994,
|
| 293153 |
+
"step": 152315
|
| 293154 |
+
},
|
| 293155 |
+
{
|
| 293156 |
+
"epoch": 1228.0,
|
| 293157 |
+
"eval_loss": 0.4167693257331848,
|
| 293158 |
+
"eval_runtime": 41.5016,
|
| 293159 |
+
"eval_samples_per_second": 20.144,
|
| 293160 |
+
"eval_steps_per_second": 0.651,
|
| 293161 |
+
"eval_wer": 0.18405892219062203,
|
| 293162 |
+
"step": 152316
|
| 293163 |
+
},
|
| 293164 |
+
{
|
| 293165 |
+
"epoch": 1218.03,
|
| 293166 |
+
"learning_rate": 7.5563651050080774e-06,
|
| 293167 |
+
"loss": 0.2815,
|
| 293168 |
+
"step": 152320
|
| 293169 |
+
},
|
| 293170 |
+
{
|
| 293171 |
+
"epoch": 1218.07,
|
| 293172 |
+
"learning_rate": 7.556284329563813e-06,
|
| 293173 |
+
"loss": 0.2471,
|
| 293174 |
+
"step": 152325
|
| 293175 |
+
},
|
| 293176 |
+
{
|
| 293177 |
+
"epoch": 1218.11,
|
| 293178 |
+
"learning_rate": 7.556203554119547e-06,
|
| 293179 |
+
"loss": 0.2546,
|
| 293180 |
+
"step": 152330
|
| 293181 |
+
},
|
| 293182 |
+
{
|
| 293183 |
+
"epoch": 1218.15,
|
| 293184 |
+
"learning_rate": 7.556122778675283e-06,
|
| 293185 |
+
"loss": 0.339,
|
| 293186 |
+
"step": 152335
|
| 293187 |
+
},
|
| 293188 |
+
{
|
| 293189 |
+
"epoch": 1218.19,
|
| 293190 |
+
"learning_rate": 7.556042003231019e-06,
|
| 293191 |
+
"loss": 0.7478,
|
| 293192 |
+
"step": 152340
|
| 293193 |
+
},
|
| 293194 |
+
{
|
| 293195 |
+
"epoch": 1218.23,
|
| 293196 |
+
"learning_rate": 7.555961227786753e-06,
|
| 293197 |
+
"loss": 0.6217,
|
| 293198 |
+
"step": 152345
|
| 293199 |
+
},
|
| 293200 |
+
{
|
| 293201 |
+
"epoch": 1218.27,
|
| 293202 |
+
"learning_rate": 7.555880452342489e-06,
|
| 293203 |
+
"loss": 0.2577,
|
| 293204 |
+
"step": 152350
|
| 293205 |
+
},
|
| 293206 |
+
{
|
| 293207 |
+
"epoch": 1218.31,
|
| 293208 |
+
"learning_rate": 7.555799676898223e-06,
|
| 293209 |
+
"loss": 0.2436,
|
| 293210 |
+
"step": 152355
|
| 293211 |
+
},
|
| 293212 |
+
{
|
| 293213 |
+
"epoch": 1218.35,
|
| 293214 |
+
"learning_rate": 7.555718901453959e-06,
|
| 293215 |
+
"loss": 0.437,
|
| 293216 |
+
"step": 152360
|
| 293217 |
+
},
|
| 293218 |
+
{
|
| 293219 |
+
"epoch": 1218.39,
|
| 293220 |
+
"learning_rate": 7.555638126009693e-06,
|
| 293221 |
+
"loss": 0.829,
|
| 293222 |
+
"step": 152365
|
| 293223 |
+
},
|
| 293224 |
+
{
|
| 293225 |
+
"epoch": 1218.43,
|
| 293226 |
+
"learning_rate": 7.555557350565429e-06,
|
| 293227 |
+
"loss": 0.5973,
|
| 293228 |
+
"step": 152370
|
| 293229 |
+
},
|
| 293230 |
+
{
|
| 293231 |
+
"epoch": 1218.47,
|
| 293232 |
+
"learning_rate": 7.555476575121163e-06,
|
| 293233 |
+
"loss": 0.2641,
|
| 293234 |
+
"step": 152375
|
| 293235 |
+
},
|
| 293236 |
+
{
|
| 293237 |
+
"epoch": 1218.51,
|
| 293238 |
+
"learning_rate": 7.555395799676899e-06,
|
| 293239 |
+
"loss": 0.2739,
|
| 293240 |
+
"step": 152380
|
| 293241 |
+
},
|
| 293242 |
+
{
|
| 293243 |
+
"epoch": 1218.55,
|
| 293244 |
+
"learning_rate": 7.555315024232633e-06,
|
| 293245 |
+
"loss": 0.4165,
|
| 293246 |
+
"step": 152385
|
| 293247 |
+
},
|
| 293248 |
+
{
|
| 293249 |
+
"epoch": 1218.59,
|
| 293250 |
+
"learning_rate": 7.555234248788369e-06,
|
| 293251 |
+
"loss": 0.826,
|
| 293252 |
+
"step": 152390
|
| 293253 |
+
},
|
| 293254 |
+
{
|
| 293255 |
+
"epoch": 1218.63,
|
| 293256 |
+
"learning_rate": 7.555153473344105e-06,
|
| 293257 |
+
"loss": 0.6447,
|
| 293258 |
+
"step": 152395
|
| 293259 |
+
},
|
| 293260 |
+
{
|
| 293261 |
+
"epoch": 1218.67,
|
| 293262 |
+
"learning_rate": 7.555072697899839e-06,
|
| 293263 |
+
"loss": 0.2597,
|
| 293264 |
+
"step": 152400
|
| 293265 |
+
},
|
| 293266 |
+
{
|
| 293267 |
+
"epoch": 1218.71,
|
| 293268 |
+
"learning_rate": 7.554991922455575e-06,
|
| 293269 |
+
"loss": 0.2887,
|
| 293270 |
+
"step": 152405
|
| 293271 |
+
},
|
| 293272 |
+
{
|
| 293273 |
+
"epoch": 1218.75,
|
| 293274 |
+
"learning_rate": 7.554911147011309e-06,
|
| 293275 |
+
"loss": 0.4304,
|
| 293276 |
+
"step": 152410
|
| 293277 |
+
},
|
| 293278 |
+
{
|
| 293279 |
+
"epoch": 1218.79,
|
| 293280 |
+
"learning_rate": 7.554830371567045e-06,
|
| 293281 |
+
"loss": 0.9143,
|
| 293282 |
+
"step": 152415
|
| 293283 |
+
},
|
| 293284 |
+
{
|
| 293285 |
+
"epoch": 1218.83,
|
| 293286 |
+
"learning_rate": 7.554749596122779e-06,
|
| 293287 |
+
"loss": 0.6371,
|
| 293288 |
+
"step": 152420
|
| 293289 |
+
},
|
| 293290 |
+
{
|
| 293291 |
+
"epoch": 1218.87,
|
| 293292 |
+
"learning_rate": 7.554668820678515e-06,
|
| 293293 |
+
"loss": 0.2591,
|
| 293294 |
+
"step": 152425
|
| 293295 |
+
},
|
| 293296 |
+
{
|
| 293297 |
+
"epoch": 1218.91,
|
| 293298 |
+
"learning_rate": 7.554588045234249e-06,
|
| 293299 |
+
"loss": 0.3687,
|
| 293300 |
+
"step": 152430
|
| 293301 |
+
},
|
| 293302 |
+
{
|
| 293303 |
+
"epoch": 1218.95,
|
| 293304 |
+
"learning_rate": 7.554507269789985e-06,
|
| 293305 |
+
"loss": 0.4046,
|
| 293306 |
+
"step": 152435
|
| 293307 |
+
},
|
| 293308 |
+
{
|
| 293309 |
+
"epoch": 1218.99,
|
| 293310 |
+
"learning_rate": 7.554426494345719e-06,
|
| 293311 |
+
"loss": 0.902,
|
| 293312 |
+
"step": 152440
|
| 293313 |
+
},
|
| 293314 |
+
{
|
| 293315 |
+
"epoch": 1219.0,
|
| 293316 |
+
"eval_loss": 0.35546669363975525,
|
| 293317 |
+
"eval_runtime": 42.7299,
|
| 293318 |
+
"eval_samples_per_second": 19.565,
|
| 293319 |
+
"eval_steps_per_second": 0.632,
|
| 293320 |
+
"eval_wer": 0.17738750176329524,
|
| 293321 |
+
"step": 152441
|
| 293322 |
+
},
|
| 293323 |
+
{
|
| 293324 |
+
"epoch": 1229.03,
|
| 293325 |
+
"learning_rate": 7.554345718901455e-06,
|
| 293326 |
+
"loss": 0.2988,
|
| 293327 |
+
"step": 152445
|
| 293328 |
+
},
|
| 293329 |
+
{
|
| 293330 |
+
"epoch": 1229.07,
|
| 293331 |
+
"learning_rate": 7.554264943457189e-06,
|
| 293332 |
+
"loss": 0.2575,
|
| 293333 |
+
"step": 152450
|
| 293334 |
+
},
|
| 293335 |
+
{
|
| 293336 |
+
"epoch": 1229.11,
|
| 293337 |
+
"learning_rate": 7.554184168012925e-06,
|
| 293338 |
+
"loss": 0.3322,
|
| 293339 |
+
"step": 152455
|
| 293340 |
+
},
|
| 293341 |
+
{
|
| 293342 |
+
"epoch": 1229.15,
|
| 293343 |
+
"learning_rate": 7.5541033925686605e-06,
|
| 293344 |
+
"loss": 0.4137,
|
| 293345 |
+
"step": 152460
|
| 293346 |
+
},
|
| 293347 |
+
{
|
| 293348 |
+
"epoch": 1229.19,
|
| 293349 |
+
"learning_rate": 7.554022617124395e-06,
|
| 293350 |
+
"loss": 0.8767,
|
| 293351 |
+
"step": 152465
|
| 293352 |
+
},
|
| 293353 |
+
{
|
| 293354 |
+
"epoch": 1229.23,
|
| 293355 |
+
"learning_rate": 7.5539418416801305e-06,
|
| 293356 |
+
"loss": 0.625,
|
| 293357 |
+
"step": 152470
|
| 293358 |
+
},
|
| 293359 |
+
{
|
| 293360 |
+
"epoch": 1229.27,
|
| 293361 |
+
"learning_rate": 7.553861066235865e-06,
|
| 293362 |
+
"loss": 0.272,
|
| 293363 |
+
"step": 152475
|
| 293364 |
+
},
|
| 293365 |
+
{
|
| 293366 |
+
"epoch": 1229.31,
|
| 293367 |
+
"learning_rate": 7.5537802907916004e-06,
|
| 293368 |
+
"loss": 0.3351,
|
| 293369 |
+
"step": 152480
|
| 293370 |
+
},
|
| 293371 |
+
{
|
| 293372 |
+
"epoch": 1229.35,
|
| 293373 |
+
"learning_rate": 7.5536995153473346e-06,
|
| 293374 |
+
"loss": 0.356,
|
| 293375 |
+
"step": 152485
|
| 293376 |
+
},
|
| 293377 |
+
{
|
| 293378 |
+
"epoch": 1229.39,
|
| 293379 |
+
"learning_rate": 7.55361873990307e-06,
|
| 293380 |
+
"loss": 0.8533,
|
| 293381 |
+
"step": 152490
|
| 293382 |
+
},
|
| 293383 |
+
{
|
| 293384 |
+
"epoch": 1229.43,
|
| 293385 |
+
"learning_rate": 7.5535379644588045e-06,
|
| 293386 |
+
"loss": 0.6315,
|
| 293387 |
+
"step": 152495
|
| 293388 |
+
},
|
| 293389 |
+
{
|
| 293390 |
+
"epoch": 1229.47,
|
| 293391 |
+
"learning_rate": 7.55345718901454e-06,
|
| 293392 |
+
"loss": 0.2547,
|
| 293393 |
+
"step": 152500
|
| 293394 |
+
},
|
| 293395 |
+
{
|
| 293396 |
+
"epoch": 1229.51,
|
| 293397 |
+
"learning_rate": 7.5533764135702745e-06,
|
| 293398 |
+
"loss": 0.2782,
|
| 293399 |
+
"step": 152505
|
| 293400 |
+
},
|
| 293401 |
+
{
|
| 293402 |
+
"epoch": 1229.55,
|
| 293403 |
+
"learning_rate": 7.55329563812601e-06,
|
| 293404 |
+
"loss": 0.3844,
|
| 293405 |
+
"step": 152510
|
| 293406 |
+
},
|
| 293407 |
+
{
|
| 293408 |
+
"epoch": 1229.59,
|
| 293409 |
+
"learning_rate": 7.553214862681746e-06,
|
| 293410 |
+
"loss": 1.0582,
|
| 293411 |
+
"step": 152515
|
| 293412 |
+
},
|
| 293413 |
+
{
|
| 293414 |
+
"epoch": 1229.63,
|
| 293415 |
+
"learning_rate": 7.55313408723748e-06,
|
| 293416 |
+
"loss": 0.6986,
|
| 293417 |
+
"step": 152520
|
| 293418 |
+
},
|
| 293419 |
+
{
|
| 293420 |
+
"epoch": 1229.67,
|
| 293421 |
+
"learning_rate": 7.553053311793216e-06,
|
| 293422 |
+
"loss": 0.2519,
|
| 293423 |
+
"step": 152525
|
| 293424 |
+
},
|
| 293425 |
+
{
|
| 293426 |
+
"epoch": 1229.71,
|
| 293427 |
+
"learning_rate": 7.55297253634895e-06,
|
| 293428 |
+
"loss": 0.3056,
|
| 293429 |
+
"step": 152530
|
| 293430 |
+
},
|
| 293431 |
+
{
|
| 293432 |
+
"epoch": 1229.76,
|
| 293433 |
+
"learning_rate": 7.552891760904686e-06,
|
| 293434 |
+
"loss": 0.3928,
|
| 293435 |
+
"step": 152535
|
| 293436 |
+
},
|
| 293437 |
+
{
|
| 293438 |
+
"epoch": 1229.8,
|
| 293439 |
+
"learning_rate": 7.55281098546042e-06,
|
| 293440 |
+
"loss": 0.8044,
|
| 293441 |
+
"step": 152540
|
| 293442 |
+
},
|
| 293443 |
+
{
|
| 293444 |
+
"epoch": 1229.84,
|
| 293445 |
+
"learning_rate": 7.552730210016156e-06,
|
| 293446 |
+
"loss": 0.651,
|
| 293447 |
+
"step": 152545
|
| 293448 |
+
},
|
| 293449 |
+
{
|
| 293450 |
+
"epoch": 1229.88,
|
| 293451 |
+
"learning_rate": 7.55264943457189e-06,
|
| 293452 |
+
"loss": 0.2421,
|
| 293453 |
+
"step": 152550
|
| 293454 |
+
},
|
| 293455 |
+
{
|
| 293456 |
+
"epoch": 1229.92,
|
| 293457 |
+
"learning_rate": 7.552568659127626e-06,
|
| 293458 |
+
"loss": 0.3304,
|
| 293459 |
+
"step": 152555
|
| 293460 |
+
},
|
| 293461 |
+
{
|
| 293462 |
+
"epoch": 1229.96,
|
| 293463 |
+
"learning_rate": 7.55248788368336e-06,
|
| 293464 |
+
"loss": 0.4315,
|
| 293465 |
+
"step": 152560
|
| 293466 |
+
},
|
| 293467 |
+
{
|
| 293468 |
+
"epoch": 1230.0,
|
| 293469 |
+
"learning_rate": 7.552407108239096e-06,
|
| 293470 |
+
"loss": 1.0926,
|
| 293471 |
+
"step": 152565
|
| 293472 |
+
},
|
| 293473 |
+
{
|
| 293474 |
+
"epoch": 1230.0,
|
| 293475 |
+
"eval_loss": 0.5449385643005371,
|
| 293476 |
+
"eval_runtime": 40.681,
|
| 293477 |
+
"eval_samples_per_second": 20.55,
|
| 293478 |
+
"eval_steps_per_second": 0.664,
|
| 293479 |
+
"eval_wer": 0.18418858485560524,
|
| 293480 |
+
"step": 152565
|
| 293481 |
+
},
|
| 293482 |
+
{
|
| 293483 |
+
"epoch": 1220.04,
|
| 293484 |
+
"learning_rate": 7.552326332794832e-06,
|
| 293485 |
+
"loss": 0.2789,
|
| 293486 |
+
"step": 152570
|
| 293487 |
+
},
|
| 293488 |
+
{
|
| 293489 |
+
"epoch": 1220.08,
|
| 293490 |
+
"learning_rate": 7.552245557350566e-06,
|
| 293491 |
+
"loss": 0.271,
|
| 293492 |
+
"step": 152575
|
| 293493 |
+
},
|
| 293494 |
+
{
|
| 293495 |
+
"epoch": 1220.12,
|
| 293496 |
+
"learning_rate": 7.552164781906302e-06,
|
| 293497 |
+
"loss": 0.243,
|
| 293498 |
+
"step": 152580
|
| 293499 |
+
},
|
| 293500 |
+
{
|
| 293501 |
+
"epoch": 1220.16,
|
| 293502 |
+
"learning_rate": 7.552084006462036e-06,
|
| 293503 |
+
"loss": 0.4346,
|
| 293504 |
+
"step": 152585
|
| 293505 |
+
},
|
| 293506 |
+
{
|
| 293507 |
+
"epoch": 1220.2,
|
| 293508 |
+
"learning_rate": 7.552003231017772e-06,
|
| 293509 |
+
"loss": 1.2121,
|
| 293510 |
+
"step": 152590
|
| 293511 |
+
},
|
| 293512 |
+
{
|
| 293513 |
+
"epoch": 1220.24,
|
| 293514 |
+
"learning_rate": 7.551922455573506e-06,
|
| 293515 |
+
"loss": 0.373,
|
| 293516 |
+
"step": 152595
|
| 293517 |
+
},
|
| 293518 |
+
{
|
| 293519 |
+
"epoch": 1220.28,
|
| 293520 |
+
"learning_rate": 7.551841680129242e-06,
|
| 293521 |
+
"loss": 0.2537,
|
| 293522 |
+
"step": 152600
|
| 293523 |
+
},
|
| 293524 |
+
{
|
| 293525 |
+
"epoch": 1220.32,
|
| 293526 |
+
"learning_rate": 7.551760904684976e-06,
|
| 293527 |
+
"loss": 0.2496,
|
| 293528 |
+
"step": 152605
|
| 293529 |
+
},
|
| 293530 |
+
{
|
| 293531 |
+
"epoch": 1220.36,
|
| 293532 |
+
"learning_rate": 7.551680129240712e-06,
|
| 293533 |
+
"loss": 0.5073,
|
| 293534 |
+
"step": 152610
|
| 293535 |
+
},
|
| 293536 |
+
{
|
| 293537 |
+
"epoch": 1220.4,
|
| 293538 |
+
"learning_rate": 7.551599353796446e-06,
|
| 293539 |
+
"loss": 1.2567,
|
| 293540 |
+
"step": 152615
|
| 293541 |
+
},
|
| 293542 |
+
{
|
| 293543 |
+
"epoch": 1220.44,
|
| 293544 |
+
"learning_rate": 7.551518578352182e-06,
|
| 293545 |
+
"loss": 0.3609,
|
| 293546 |
+
"step": 152620
|
| 293547 |
+
},
|
| 293548 |
+
{
|
| 293549 |
+
"epoch": 1220.48,
|
| 293550 |
+
"learning_rate": 7.551437802907916e-06,
|
| 293551 |
+
"loss": 0.2748,
|
| 293552 |
+
"step": 152625
|
| 293553 |
+
},
|
| 293554 |
+
{
|
| 293555 |
+
"epoch": 1220.52,
|
| 293556 |
+
"learning_rate": 7.551357027463652e-06,
|
| 293557 |
+
"loss": 0.3177,
|
| 293558 |
+
"step": 152630
|
| 293559 |
+
},
|
| 293560 |
+
{
|
| 293561 |
+
"epoch": 1220.56,
|
| 293562 |
+
"learning_rate": 7.551276252019388e-06,
|
| 293563 |
+
"loss": 0.437,
|
| 293564 |
+
"step": 152635
|
| 293565 |
+
},
|
| 293566 |
+
{
|
| 293567 |
+
"epoch": 1220.6,
|
| 293568 |
+
"learning_rate": 7.551195476575122e-06,
|
| 293569 |
+
"loss": 1.3232,
|
| 293570 |
+
"step": 152640
|
| 293571 |
+
},
|
| 293572 |
+
{
|
| 293573 |
+
"epoch": 1220.64,
|
| 293574 |
+
"learning_rate": 7.5511147011308576e-06,
|
| 293575 |
+
"loss": 0.2868,
|
| 293576 |
+
"step": 152645
|
| 293577 |
+
},
|
| 293578 |
+
{
|
| 293579 |
+
"epoch": 1220.68,
|
| 293580 |
+
"learning_rate": 7.551033925686592e-06,
|
| 293581 |
+
"loss": 0.3602,
|
| 293582 |
+
"step": 152650
|
| 293583 |
+
},
|
| 293584 |
+
{
|
| 293585 |
+
"epoch": 1220.72,
|
| 293586 |
+
"learning_rate": 7.5509531502423275e-06,
|
| 293587 |
+
"loss": 0.2706,
|
| 293588 |
+
"step": 152655
|
| 293589 |
+
},
|
| 293590 |
+
{
|
| 293591 |
+
"epoch": 1220.76,
|
| 293592 |
+
"learning_rate": 7.550872374798062e-06,
|
| 293593 |
+
"loss": 0.4301,
|
| 293594 |
+
"step": 152660
|
| 293595 |
+
},
|
| 293596 |
+
{
|
| 293597 |
+
"epoch": 1220.8,
|
| 293598 |
+
"learning_rate": 7.5507915993537975e-06,
|
| 293599 |
+
"loss": 1.0362,
|
| 293600 |
+
"step": 152665
|
| 293601 |
+
},
|
| 293602 |
+
{
|
| 293603 |
+
"epoch": 1220.84,
|
| 293604 |
+
"learning_rate": 7.550710823909532e-06,
|
| 293605 |
+
"loss": 0.2926,
|
| 293606 |
+
"step": 152670
|
| 293607 |
+
},
|
| 293608 |
+
{
|
| 293609 |
+
"epoch": 1220.88,
|
| 293610 |
+
"learning_rate": 7.5506300484652675e-06,
|
| 293611 |
+
"loss": 0.2753,
|
| 293612 |
+
"step": 152675
|
| 293613 |
+
},
|
| 293614 |
+
{
|
| 293615 |
+
"epoch": 1220.92,
|
| 293616 |
+
"learning_rate": 7.550549273021002e-06,
|
| 293617 |
+
"loss": 0.3452,
|
| 293618 |
+
"step": 152680
|
| 293619 |
+
},
|
| 293620 |
+
{
|
| 293621 |
+
"epoch": 1220.96,
|
| 293622 |
+
"learning_rate": 7.5504684975767375e-06,
|
| 293623 |
+
"loss": 0.4678,
|
| 293624 |
+
"step": 152685
|
| 293625 |
+
},
|
| 293626 |
+
{
|
| 293627 |
+
"epoch": 1221.0,
|
| 293628 |
+
"learning_rate": 7.5503877221324724e-06,
|
| 293629 |
+
"loss": 1.3258,
|
| 293630 |
+
"step": 152690
|
| 293631 |
+
},
|
| 293632 |
+
{
|
| 293633 |
+
"epoch": 1221.0,
|
| 293634 |
+
"eval_loss": 0.3559703230857849,
|
| 293635 |
+
"eval_runtime": 43.0859,
|
| 293636 |
+
"eval_samples_per_second": 19.403,
|
| 293637 |
+
"eval_steps_per_second": 0.627,
|
| 293638 |
+
"eval_wer": 0.17048071387449626,
|
| 293639 |
+
"step": 152690
|
| 293640 |
}
|
| 293641 |
],
|
| 293642 |
+
"max_steps": 625000,
|
| 293643 |
"num_train_epochs": 5000,
|
| 293644 |
+
"total_flos": 4.2968328943638807e+20,
|
| 293645 |
"trial_name": null,
|
| 293646 |
"trial_params": null
|
| 293647 |
}
|
model-bin/finetune/base/{checkpoint-152068 β checkpoint-152690}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1630250753.668995/events.out.tfevents.1630250753.cc93b136ebf5.1086.209
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d9c0aee1ce7801ac75359895fcb1668e7c7233a293e1e85b17034353ff0003ee
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630251188.88167/events.out.tfevents.1630251188.cc93b136ebf5.1086.211
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fc10d5878cb756554085f94d5e325de11d0130822fcaaad5d044e2996b03f139
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630251628.1797204/events.out.tfevents.1630251628.cc93b136ebf5.1086.213
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5a75e79f95a32dbd01c2a774a76e9bb2824f0d5b1551051b1f8085c6c884e1d3
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630252065.7514985/events.out.tfevents.1630252065.cc93b136ebf5.1086.215
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e323b0936deda44bf91875918182857d58ba5428f8afdebac01580255cb0e269
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630252500.5741549/events.out.tfevents.1630252500.cc93b136ebf5.1086.217
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:cb9ebf7e8137afa957d4d2026869e20b8bd962c5bce171c25618696ce1316d6e
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630250753.cc93b136ebf5.1086.208
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2c6b329615b997a29dc2441fdc952ec848b8f5de5513d43042f86ea38b90097d
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630251188.cc93b136ebf5.1086.210
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:28f4b9a76a7866c02a50de0f72d113f2770ef0fb3e160da1f4087186466043da
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630251628.cc93b136ebf5.1086.212
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5c24d8523270a58fc6c0a96845b48c37ed90191d876506302e6a8a7e1c204bf0
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630252065.cc93b136ebf5.1086.214
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9ad901e5a01c00724eaa3e2c6d1064f0725f8468b8af7d9f4449a542c3c3cd99
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630252500.cc93b136ebf5.1086.216
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f1c9e6fc6b8beb6d7bda1c768e2664d3a7732f3851f11540d7cf4be9cc49d64f
|
| 3 |
+
size 8622
|