"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-134775 β checkpoint-135397}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-134775 β checkpoint-135397}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-134775 β checkpoint-135397}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-134775 β checkpoint-135397}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-134775 β checkpoint-135397}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-134775 β checkpoint-135397}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-134775 β checkpoint-135397}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-134775 β checkpoint-135397}/trainer_state.json +793 -4
- model-bin/finetune/base/{checkpoint-134775 β checkpoint-135397}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630177407.9573364/events.out.tfevents.1630177407.86bb0ddabf9b.4092.271 +3 -0
- model-bin/finetune/base/log/1630177794.4337196/events.out.tfevents.1630177794.86bb0ddabf9b.4092.273 +3 -0
- model-bin/finetune/base/log/1630178315.3923218/events.out.tfevents.1630178315.86bb0ddabf9b.4092.275 +3 -0
- model-bin/finetune/base/log/1630178704.6355941/events.out.tfevents.1630178704.86bb0ddabf9b.4092.277 +3 -0
- model-bin/finetune/base/log/1630179100.382477/events.out.tfevents.1630179100.86bb0ddabf9b.4092.279 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630177407.86bb0ddabf9b.4092.270 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630177794.86bb0ddabf9b.4092.272 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630178315.86bb0ddabf9b.4092.274 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630178704.86bb0ddabf9b.4092.276 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630179100.86bb0ddabf9b.4092.278 +3 -0
model-bin/finetune/base/{checkpoint-134775 β checkpoint-135397}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-134775 β checkpoint-135397}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:263b32b67e9cf11f3e509a6180c4f9195a74262b5e6dd3d340d03a28daebcd4f
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-134775 β checkpoint-135397}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-134775 β checkpoint-135397}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:69906934b17f75f9108d164b0605da8aa57a189704d642f1bd48b13830394bdb
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-134775 β checkpoint-135397}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bcf03b393f34c957f32a4a548de51987df0b57643c3a5e679cc655ed78b5cdb6
|
| 3 |
size 14503
|
model-bin/finetune/base/{checkpoint-134775 β checkpoint-135397}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:242b9fc28ce4768baff9a3973c2264ec433790bcad1670cb995632a3f788ff11
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-134775 β checkpoint-135397}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:86054fc88dc8ccd2ec65945e420ecf3412e52c6f6e0a12d48d5f9e9b7b926e1d
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-134775 β checkpoint-135397}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1689111747851003,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-132910",
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -270843,11 +270843,800 @@
|
|
| 270843 |
"eval_steps_per_second": 0.756,
|
| 270844 |
"eval_wer": 0.18336918626214674,
|
| 270845 |
"step": 134775
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 270846 |
}
|
| 270847 |
],
|
| 270848 |
-
"max_steps":
|
| 270849 |
"num_train_epochs": 5000,
|
| 270850 |
-
"total_flos": 3.
|
| 270851 |
"trial_name": null,
|
| 270852 |
"trial_params": null
|
| 270853 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1689111747851003,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-132910",
|
| 4 |
+
"epoch": 1090.995983935743,
|
| 5 |
+
"global_step": 135397,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 270843 |
"eval_steps_per_second": 0.756,
|
| 270844 |
"eval_wer": 0.18336918626214674,
|
| 270845 |
"step": 134775
|
| 270846 |
+
},
|
| 270847 |
+
{
|
| 270848 |
+
"epoch": 1078.04,
|
| 270849 |
+
"learning_rate": 7.856939102564103e-06,
|
| 270850 |
+
"loss": 0.3143,
|
| 270851 |
+
"step": 134780
|
| 270852 |
+
},
|
| 270853 |
+
{
|
| 270854 |
+
"epoch": 1078.08,
|
| 270855 |
+
"learning_rate": 7.856858974358975e-06,
|
| 270856 |
+
"loss": 0.2961,
|
| 270857 |
+
"step": 134785
|
| 270858 |
+
},
|
| 270859 |
+
{
|
| 270860 |
+
"epoch": 1078.12,
|
| 270861 |
+
"learning_rate": 7.856778846153846e-06,
|
| 270862 |
+
"loss": 0.2859,
|
| 270863 |
+
"step": 134790
|
| 270864 |
+
},
|
| 270865 |
+
{
|
| 270866 |
+
"epoch": 1078.16,
|
| 270867 |
+
"learning_rate": 7.85669871794872e-06,
|
| 270868 |
+
"loss": 0.4711,
|
| 270869 |
+
"step": 134795
|
| 270870 |
+
},
|
| 270871 |
+
{
|
| 270872 |
+
"epoch": 1078.2,
|
| 270873 |
+
"learning_rate": 7.85661858974359e-06,
|
| 270874 |
+
"loss": 1.2223,
|
| 270875 |
+
"step": 134800
|
| 270876 |
+
},
|
| 270877 |
+
{
|
| 270878 |
+
"epoch": 1078.24,
|
| 270879 |
+
"learning_rate": 7.856538461538462e-06,
|
| 270880 |
+
"loss": 0.2998,
|
| 270881 |
+
"step": 134805
|
| 270882 |
+
},
|
| 270883 |
+
{
|
| 270884 |
+
"epoch": 1078.28,
|
| 270885 |
+
"learning_rate": 7.856458333333335e-06,
|
| 270886 |
+
"loss": 0.2899,
|
| 270887 |
+
"step": 134810
|
| 270888 |
+
},
|
| 270889 |
+
{
|
| 270890 |
+
"epoch": 1078.32,
|
| 270891 |
+
"learning_rate": 7.856378205128206e-06,
|
| 270892 |
+
"loss": 0.2958,
|
| 270893 |
+
"step": 134815
|
| 270894 |
+
},
|
| 270895 |
+
{
|
| 270896 |
+
"epoch": 1078.36,
|
| 270897 |
+
"learning_rate": 7.856298076923078e-06,
|
| 270898 |
+
"loss": 0.4927,
|
| 270899 |
+
"step": 134820
|
| 270900 |
+
},
|
| 270901 |
+
{
|
| 270902 |
+
"epoch": 1078.4,
|
| 270903 |
+
"learning_rate": 7.856217948717949e-06,
|
| 270904 |
+
"loss": 1.141,
|
| 270905 |
+
"step": 134825
|
| 270906 |
+
},
|
| 270907 |
+
{
|
| 270908 |
+
"epoch": 1078.44,
|
| 270909 |
+
"learning_rate": 7.856137820512822e-06,
|
| 270910 |
+
"loss": 0.3416,
|
| 270911 |
+
"step": 134830
|
| 270912 |
+
},
|
| 270913 |
+
{
|
| 270914 |
+
"epoch": 1078.48,
|
| 270915 |
+
"learning_rate": 7.856057692307692e-06,
|
| 270916 |
+
"loss": 0.2529,
|
| 270917 |
+
"step": 134835
|
| 270918 |
+
},
|
| 270919 |
+
{
|
| 270920 |
+
"epoch": 1078.52,
|
| 270921 |
+
"learning_rate": 7.855977564102565e-06,
|
| 270922 |
+
"loss": 0.2934,
|
| 270923 |
+
"step": 134840
|
| 270924 |
+
},
|
| 270925 |
+
{
|
| 270926 |
+
"epoch": 1078.56,
|
| 270927 |
+
"learning_rate": 7.855897435897436e-06,
|
| 270928 |
+
"loss": 0.4133,
|
| 270929 |
+
"step": 134845
|
| 270930 |
+
},
|
| 270931 |
+
{
|
| 270932 |
+
"epoch": 1078.6,
|
| 270933 |
+
"learning_rate": 7.855817307692308e-06,
|
| 270934 |
+
"loss": 1.2142,
|
| 270935 |
+
"step": 134850
|
| 270936 |
+
},
|
| 270937 |
+
{
|
| 270938 |
+
"epoch": 1078.64,
|
| 270939 |
+
"learning_rate": 7.85573717948718e-06,
|
| 270940 |
+
"loss": 0.3058,
|
| 270941 |
+
"step": 134855
|
| 270942 |
+
},
|
| 270943 |
+
{
|
| 270944 |
+
"epoch": 1078.68,
|
| 270945 |
+
"learning_rate": 7.855657051282052e-06,
|
| 270946 |
+
"loss": 0.2776,
|
| 270947 |
+
"step": 134860
|
| 270948 |
+
},
|
| 270949 |
+
{
|
| 270950 |
+
"epoch": 1078.72,
|
| 270951 |
+
"learning_rate": 7.855576923076923e-06,
|
| 270952 |
+
"loss": 0.3003,
|
| 270953 |
+
"step": 134865
|
| 270954 |
+
},
|
| 270955 |
+
{
|
| 270956 |
+
"epoch": 1078.76,
|
| 270957 |
+
"learning_rate": 7.855496794871795e-06,
|
| 270958 |
+
"loss": 0.4309,
|
| 270959 |
+
"step": 134870
|
| 270960 |
+
},
|
| 270961 |
+
{
|
| 270962 |
+
"epoch": 1078.8,
|
| 270963 |
+
"learning_rate": 7.855416666666668e-06,
|
| 270964 |
+
"loss": 1.1745,
|
| 270965 |
+
"step": 134875
|
| 270966 |
+
},
|
| 270967 |
+
{
|
| 270968 |
+
"epoch": 1078.84,
|
| 270969 |
+
"learning_rate": 7.855336538461539e-06,
|
| 270970 |
+
"loss": 0.3052,
|
| 270971 |
+
"step": 134880
|
| 270972 |
+
},
|
| 270973 |
+
{
|
| 270974 |
+
"epoch": 1078.88,
|
| 270975 |
+
"learning_rate": 7.85525641025641e-06,
|
| 270976 |
+
"loss": 0.285,
|
| 270977 |
+
"step": 134885
|
| 270978 |
+
},
|
| 270979 |
+
{
|
| 270980 |
+
"epoch": 1078.92,
|
| 270981 |
+
"learning_rate": 7.855176282051282e-06,
|
| 270982 |
+
"loss": 0.296,
|
| 270983 |
+
"step": 134890
|
| 270984 |
+
},
|
| 270985 |
+
{
|
| 270986 |
+
"epoch": 1078.96,
|
| 270987 |
+
"learning_rate": 7.855096153846155e-06,
|
| 270988 |
+
"loss": 0.4263,
|
| 270989 |
+
"step": 134895
|
| 270990 |
+
},
|
| 270991 |
+
{
|
| 270992 |
+
"epoch": 1079.0,
|
| 270993 |
+
"learning_rate": 7.855016025641026e-06,
|
| 270994 |
+
"loss": 1.5733,
|
| 270995 |
+
"step": 134900
|
| 270996 |
+
},
|
| 270997 |
+
{
|
| 270998 |
+
"epoch": 1079.0,
|
| 270999 |
+
"eval_loss": 0.4056174159049988,
|
| 271000 |
+
"eval_runtime": 36.4834,
|
| 271001 |
+
"eval_samples_per_second": 22.887,
|
| 271002 |
+
"eval_steps_per_second": 0.74,
|
| 271003 |
+
"eval_wer": 0.17084527220630372,
|
| 271004 |
+
"step": 134900
|
| 271005 |
+
},
|
| 271006 |
+
{
|
| 271007 |
+
"epoch": 1087.04,
|
| 271008 |
+
"learning_rate": 7.854935897435898e-06,
|
| 271009 |
+
"loss": 0.2978,
|
| 271010 |
+
"step": 134905
|
| 271011 |
+
},
|
| 271012 |
+
{
|
| 271013 |
+
"epoch": 1087.08,
|
| 271014 |
+
"learning_rate": 7.85485576923077e-06,
|
| 271015 |
+
"loss": 0.2587,
|
| 271016 |
+
"step": 134910
|
| 271017 |
+
},
|
| 271018 |
+
{
|
| 271019 |
+
"epoch": 1087.12,
|
| 271020 |
+
"learning_rate": 7.854775641025642e-06,
|
| 271021 |
+
"loss": 0.2738,
|
| 271022 |
+
"step": 134915
|
| 271023 |
+
},
|
| 271024 |
+
{
|
| 271025 |
+
"epoch": 1087.16,
|
| 271026 |
+
"learning_rate": 7.854695512820513e-06,
|
| 271027 |
+
"loss": 0.4799,
|
| 271028 |
+
"step": 134920
|
| 271029 |
+
},
|
| 271030 |
+
{
|
| 271031 |
+
"epoch": 1087.2,
|
| 271032 |
+
"learning_rate": 7.854615384615385e-06,
|
| 271033 |
+
"loss": 1.1296,
|
| 271034 |
+
"step": 134925
|
| 271035 |
+
},
|
| 271036 |
+
{
|
| 271037 |
+
"epoch": 1087.24,
|
| 271038 |
+
"learning_rate": 7.854535256410258e-06,
|
| 271039 |
+
"loss": 0.2773,
|
| 271040 |
+
"step": 134930
|
| 271041 |
+
},
|
| 271042 |
+
{
|
| 271043 |
+
"epoch": 1087.28,
|
| 271044 |
+
"learning_rate": 7.854455128205129e-06,
|
| 271045 |
+
"loss": 0.292,
|
| 271046 |
+
"step": 134935
|
| 271047 |
+
},
|
| 271048 |
+
{
|
| 271049 |
+
"epoch": 1087.32,
|
| 271050 |
+
"learning_rate": 7.854375e-06,
|
| 271051 |
+
"loss": 0.2944,
|
| 271052 |
+
"step": 134940
|
| 271053 |
+
},
|
| 271054 |
+
{
|
| 271055 |
+
"epoch": 1087.36,
|
| 271056 |
+
"learning_rate": 7.854294871794872e-06,
|
| 271057 |
+
"loss": 0.4595,
|
| 271058 |
+
"step": 134945
|
| 271059 |
+
},
|
| 271060 |
+
{
|
| 271061 |
+
"epoch": 1087.4,
|
| 271062 |
+
"learning_rate": 7.854214743589745e-06,
|
| 271063 |
+
"loss": 1.1867,
|
| 271064 |
+
"step": 134950
|
| 271065 |
+
},
|
| 271066 |
+
{
|
| 271067 |
+
"epoch": 1087.44,
|
| 271068 |
+
"learning_rate": 7.854134615384616e-06,
|
| 271069 |
+
"loss": 0.3082,
|
| 271070 |
+
"step": 134955
|
| 271071 |
+
},
|
| 271072 |
+
{
|
| 271073 |
+
"epoch": 1087.48,
|
| 271074 |
+
"learning_rate": 7.854054487179488e-06,
|
| 271075 |
+
"loss": 0.4056,
|
| 271076 |
+
"step": 134960
|
| 271077 |
+
},
|
| 271078 |
+
{
|
| 271079 |
+
"epoch": 1087.52,
|
| 271080 |
+
"learning_rate": 7.85397435897436e-06,
|
| 271081 |
+
"loss": 0.2779,
|
| 271082 |
+
"step": 134965
|
| 271083 |
+
},
|
| 271084 |
+
{
|
| 271085 |
+
"epoch": 1087.56,
|
| 271086 |
+
"learning_rate": 7.85389423076923e-06,
|
| 271087 |
+
"loss": 0.5577,
|
| 271088 |
+
"step": 134970
|
| 271089 |
+
},
|
| 271090 |
+
{
|
| 271091 |
+
"epoch": 1087.6,
|
| 271092 |
+
"learning_rate": 7.853814102564103e-06,
|
| 271093 |
+
"loss": 1.1997,
|
| 271094 |
+
"step": 134975
|
| 271095 |
+
},
|
| 271096 |
+
{
|
| 271097 |
+
"epoch": 1087.64,
|
| 271098 |
+
"learning_rate": 7.853733974358975e-06,
|
| 271099 |
+
"loss": 0.3067,
|
| 271100 |
+
"step": 134980
|
| 271101 |
+
},
|
| 271102 |
+
{
|
| 271103 |
+
"epoch": 1087.68,
|
| 271104 |
+
"learning_rate": 7.853653846153846e-06,
|
| 271105 |
+
"loss": 0.2451,
|
| 271106 |
+
"step": 134985
|
| 271107 |
+
},
|
| 271108 |
+
{
|
| 271109 |
+
"epoch": 1087.72,
|
| 271110 |
+
"learning_rate": 7.853573717948717e-06,
|
| 271111 |
+
"loss": 0.2979,
|
| 271112 |
+
"step": 134990
|
| 271113 |
+
},
|
| 271114 |
+
{
|
| 271115 |
+
"epoch": 1087.76,
|
| 271116 |
+
"learning_rate": 7.85349358974359e-06,
|
| 271117 |
+
"loss": 0.4681,
|
| 271118 |
+
"step": 134995
|
| 271119 |
+
},
|
| 271120 |
+
{
|
| 271121 |
+
"epoch": 1087.8,
|
| 271122 |
+
"learning_rate": 7.853413461538464e-06,
|
| 271123 |
+
"loss": 1.2237,
|
| 271124 |
+
"step": 135000
|
| 271125 |
+
},
|
| 271126 |
+
{
|
| 271127 |
+
"epoch": 1087.84,
|
| 271128 |
+
"learning_rate": 7.853333333333333e-06,
|
| 271129 |
+
"loss": 0.4052,
|
| 271130 |
+
"step": 135005
|
| 271131 |
+
},
|
| 271132 |
+
{
|
| 271133 |
+
"epoch": 1087.88,
|
| 271134 |
+
"learning_rate": 7.853253205128206e-06,
|
| 271135 |
+
"loss": 0.3075,
|
| 271136 |
+
"step": 135010
|
| 271137 |
+
},
|
| 271138 |
+
{
|
| 271139 |
+
"epoch": 1087.92,
|
| 271140 |
+
"learning_rate": 7.853173076923078e-06,
|
| 271141 |
+
"loss": 0.2888,
|
| 271142 |
+
"step": 135015
|
| 271143 |
+
},
|
| 271144 |
+
{
|
| 271145 |
+
"epoch": 1087.96,
|
| 271146 |
+
"learning_rate": 7.853092948717949e-06,
|
| 271147 |
+
"loss": 0.473,
|
| 271148 |
+
"step": 135020
|
| 271149 |
+
},
|
| 271150 |
+
{
|
| 271151 |
+
"epoch": 1088.0,
|
| 271152 |
+
"eval_loss": 0.3412569761276245,
|
| 271153 |
+
"eval_runtime": 47.4376,
|
| 271154 |
+
"eval_samples_per_second": 17.602,
|
| 271155 |
+
"eval_steps_per_second": 0.569,
|
| 271156 |
+
"eval_wer": 0.1722349803689162,
|
| 271157 |
+
"step": 135024
|
| 271158 |
+
},
|
| 271159 |
+
{
|
| 271160 |
+
"epoch": 1088.01,
|
| 271161 |
+
"learning_rate": 7.85301282051282e-06,
|
| 271162 |
+
"loss": 0.3541,
|
| 271163 |
+
"step": 135025
|
| 271164 |
+
},
|
| 271165 |
+
{
|
| 271166 |
+
"epoch": 1088.05,
|
| 271167 |
+
"learning_rate": 7.852932692307693e-06,
|
| 271168 |
+
"loss": 0.2841,
|
| 271169 |
+
"step": 135030
|
| 271170 |
+
},
|
| 271171 |
+
{
|
| 271172 |
+
"epoch": 1088.09,
|
| 271173 |
+
"learning_rate": 7.852852564102565e-06,
|
| 271174 |
+
"loss": 0.2678,
|
| 271175 |
+
"step": 135035
|
| 271176 |
+
},
|
| 271177 |
+
{
|
| 271178 |
+
"epoch": 1088.13,
|
| 271179 |
+
"learning_rate": 7.852772435897436e-06,
|
| 271180 |
+
"loss": 0.3088,
|
| 271181 |
+
"step": 135040
|
| 271182 |
+
},
|
| 271183 |
+
{
|
| 271184 |
+
"epoch": 1088.17,
|
| 271185 |
+
"learning_rate": 7.852692307692308e-06,
|
| 271186 |
+
"loss": 0.4679,
|
| 271187 |
+
"step": 135045
|
| 271188 |
+
},
|
| 271189 |
+
{
|
| 271190 |
+
"epoch": 1088.21,
|
| 271191 |
+
"learning_rate": 7.85261217948718e-06,
|
| 271192 |
+
"loss": 1.0785,
|
| 271193 |
+
"step": 135050
|
| 271194 |
+
},
|
| 271195 |
+
{
|
| 271196 |
+
"epoch": 1088.25,
|
| 271197 |
+
"learning_rate": 7.852532051282052e-06,
|
| 271198 |
+
"loss": 0.3021,
|
| 271199 |
+
"step": 135055
|
| 271200 |
+
},
|
| 271201 |
+
{
|
| 271202 |
+
"epoch": 1088.29,
|
| 271203 |
+
"learning_rate": 7.852451923076923e-06,
|
| 271204 |
+
"loss": 0.2812,
|
| 271205 |
+
"step": 135060
|
| 271206 |
+
},
|
| 271207 |
+
{
|
| 271208 |
+
"epoch": 1088.33,
|
| 271209 |
+
"learning_rate": 7.852371794871796e-06,
|
| 271210 |
+
"loss": 0.2677,
|
| 271211 |
+
"step": 135065
|
| 271212 |
+
},
|
| 271213 |
+
{
|
| 271214 |
+
"epoch": 1088.37,
|
| 271215 |
+
"learning_rate": 7.852291666666668e-06,
|
| 271216 |
+
"loss": 0.5161,
|
| 271217 |
+
"step": 135070
|
| 271218 |
+
},
|
| 271219 |
+
{
|
| 271220 |
+
"epoch": 1088.41,
|
| 271221 |
+
"learning_rate": 7.852211538461539e-06,
|
| 271222 |
+
"loss": 1.0857,
|
| 271223 |
+
"step": 135075
|
| 271224 |
+
},
|
| 271225 |
+
{
|
| 271226 |
+
"epoch": 1088.45,
|
| 271227 |
+
"learning_rate": 7.85213141025641e-06,
|
| 271228 |
+
"loss": 0.3308,
|
| 271229 |
+
"step": 135080
|
| 271230 |
+
},
|
| 271231 |
+
{
|
| 271232 |
+
"epoch": 1088.49,
|
| 271233 |
+
"learning_rate": 7.852051282051283e-06,
|
| 271234 |
+
"loss": 0.2679,
|
| 271235 |
+
"step": 135085
|
| 271236 |
+
},
|
| 271237 |
+
{
|
| 271238 |
+
"epoch": 1088.53,
|
| 271239 |
+
"learning_rate": 7.851971153846155e-06,
|
| 271240 |
+
"loss": 0.2971,
|
| 271241 |
+
"step": 135090
|
| 271242 |
+
},
|
| 271243 |
+
{
|
| 271244 |
+
"epoch": 1088.57,
|
| 271245 |
+
"learning_rate": 7.851891025641026e-06,
|
| 271246 |
+
"loss": 0.5472,
|
| 271247 |
+
"step": 135095
|
| 271248 |
+
},
|
| 271249 |
+
{
|
| 271250 |
+
"epoch": 1088.61,
|
| 271251 |
+
"learning_rate": 7.851810897435898e-06,
|
| 271252 |
+
"loss": 1.0136,
|
| 271253 |
+
"step": 135100
|
| 271254 |
+
},
|
| 271255 |
+
{
|
| 271256 |
+
"epoch": 1088.65,
|
| 271257 |
+
"learning_rate": 7.85173076923077e-06,
|
| 271258 |
+
"loss": 0.2525,
|
| 271259 |
+
"step": 135105
|
| 271260 |
+
},
|
| 271261 |
+
{
|
| 271262 |
+
"epoch": 1088.69,
|
| 271263 |
+
"learning_rate": 7.851650641025642e-06,
|
| 271264 |
+
"loss": 0.3529,
|
| 271265 |
+
"step": 135110
|
| 271266 |
+
},
|
| 271267 |
+
{
|
| 271268 |
+
"epoch": 1088.73,
|
| 271269 |
+
"learning_rate": 7.851570512820513e-06,
|
| 271270 |
+
"loss": 0.3132,
|
| 271271 |
+
"step": 135115
|
| 271272 |
+
},
|
| 271273 |
+
{
|
| 271274 |
+
"epoch": 1088.77,
|
| 271275 |
+
"learning_rate": 7.851490384615386e-06,
|
| 271276 |
+
"loss": 0.5094,
|
| 271277 |
+
"step": 135120
|
| 271278 |
+
},
|
| 271279 |
+
{
|
| 271280 |
+
"epoch": 1088.81,
|
| 271281 |
+
"learning_rate": 7.851410256410256e-06,
|
| 271282 |
+
"loss": 1.2523,
|
| 271283 |
+
"step": 135125
|
| 271284 |
+
},
|
| 271285 |
+
{
|
| 271286 |
+
"epoch": 1088.85,
|
| 271287 |
+
"learning_rate": 7.851330128205129e-06,
|
| 271288 |
+
"loss": 0.2699,
|
| 271289 |
+
"step": 135130
|
| 271290 |
+
},
|
| 271291 |
+
{
|
| 271292 |
+
"epoch": 1088.89,
|
| 271293 |
+
"learning_rate": 7.85125e-06,
|
| 271294 |
+
"loss": 0.2912,
|
| 271295 |
+
"step": 135135
|
| 271296 |
+
},
|
| 271297 |
+
{
|
| 271298 |
+
"epoch": 1088.93,
|
| 271299 |
+
"learning_rate": 7.851169871794872e-06,
|
| 271300 |
+
"loss": 0.3102,
|
| 271301 |
+
"step": 135140
|
| 271302 |
+
},
|
| 271303 |
+
{
|
| 271304 |
+
"epoch": 1088.97,
|
| 271305 |
+
"learning_rate": 7.851089743589743e-06,
|
| 271306 |
+
"loss": 0.5502,
|
| 271307 |
+
"step": 135145
|
| 271308 |
+
},
|
| 271309 |
+
{
|
| 271310 |
+
"epoch": 1089.0,
|
| 271311 |
+
"eval_loss": 0.31616154313087463,
|
| 271312 |
+
"eval_runtime": 37.6277,
|
| 271313 |
+
"eval_samples_per_second": 22.404,
|
| 271314 |
+
"eval_steps_per_second": 0.718,
|
| 271315 |
+
"eval_wer": 0.17189166133503944,
|
| 271316 |
+
"step": 135148
|
| 271317 |
+
},
|
| 271318 |
+
{
|
| 271319 |
+
"epoch": 1081.02,
|
| 271320 |
+
"learning_rate": 7.851009615384616e-06,
|
| 271321 |
+
"loss": 0.3547,
|
| 271322 |
+
"step": 135150
|
| 271323 |
+
},
|
| 271324 |
+
{
|
| 271325 |
+
"epoch": 1081.06,
|
| 271326 |
+
"learning_rate": 7.850929487179488e-06,
|
| 271327 |
+
"loss": 0.2758,
|
| 271328 |
+
"step": 135155
|
| 271329 |
+
},
|
| 271330 |
+
{
|
| 271331 |
+
"epoch": 1081.1,
|
| 271332 |
+
"learning_rate": 7.850849358974359e-06,
|
| 271333 |
+
"loss": 0.2526,
|
| 271334 |
+
"step": 135160
|
| 271335 |
+
},
|
| 271336 |
+
{
|
| 271337 |
+
"epoch": 1081.14,
|
| 271338 |
+
"learning_rate": 7.850769230769232e-06,
|
| 271339 |
+
"loss": 0.3548,
|
| 271340 |
+
"step": 135165
|
| 271341 |
+
},
|
| 271342 |
+
{
|
| 271343 |
+
"epoch": 1081.18,
|
| 271344 |
+
"learning_rate": 7.850689102564103e-06,
|
| 271345 |
+
"loss": 0.6463,
|
| 271346 |
+
"step": 135170
|
| 271347 |
+
},
|
| 271348 |
+
{
|
| 271349 |
+
"epoch": 1081.22,
|
| 271350 |
+
"learning_rate": 7.850608974358975e-06,
|
| 271351 |
+
"loss": 0.9952,
|
| 271352 |
+
"step": 135175
|
| 271353 |
+
},
|
| 271354 |
+
{
|
| 271355 |
+
"epoch": 1081.26,
|
| 271356 |
+
"learning_rate": 7.850528846153846e-06,
|
| 271357 |
+
"loss": 0.2859,
|
| 271358 |
+
"step": 135180
|
| 271359 |
+
},
|
| 271360 |
+
{
|
| 271361 |
+
"epoch": 1081.3,
|
| 271362 |
+
"learning_rate": 7.850448717948719e-06,
|
| 271363 |
+
"loss": 0.2531,
|
| 271364 |
+
"step": 135185
|
| 271365 |
+
},
|
| 271366 |
+
{
|
| 271367 |
+
"epoch": 1081.34,
|
| 271368 |
+
"learning_rate": 7.85036858974359e-06,
|
| 271369 |
+
"loss": 0.3461,
|
| 271370 |
+
"step": 135190
|
| 271371 |
+
},
|
| 271372 |
+
{
|
| 271373 |
+
"epoch": 1081.38,
|
| 271374 |
+
"learning_rate": 7.850288461538462e-06,
|
| 271375 |
+
"loss": 0.5408,
|
| 271376 |
+
"step": 135195
|
| 271377 |
+
},
|
| 271378 |
+
{
|
| 271379 |
+
"epoch": 1081.42,
|
| 271380 |
+
"learning_rate": 7.850208333333333e-06,
|
| 271381 |
+
"loss": 0.9077,
|
| 271382 |
+
"step": 135200
|
| 271383 |
+
},
|
| 271384 |
+
{
|
| 271385 |
+
"epoch": 1081.46,
|
| 271386 |
+
"learning_rate": 7.850128205128206e-06,
|
| 271387 |
+
"loss": 0.2544,
|
| 271388 |
+
"step": 135205
|
| 271389 |
+
},
|
| 271390 |
+
{
|
| 271391 |
+
"epoch": 1081.5,
|
| 271392 |
+
"learning_rate": 7.850048076923078e-06,
|
| 271393 |
+
"loss": 0.3896,
|
| 271394 |
+
"step": 135210
|
| 271395 |
+
},
|
| 271396 |
+
{
|
| 271397 |
+
"epoch": 1081.54,
|
| 271398 |
+
"learning_rate": 7.849967948717949e-06,
|
| 271399 |
+
"loss": 0.2894,
|
| 271400 |
+
"step": 135215
|
| 271401 |
+
},
|
| 271402 |
+
{
|
| 271403 |
+
"epoch": 1081.58,
|
| 271404 |
+
"learning_rate": 7.849887820512822e-06,
|
| 271405 |
+
"loss": 0.7234,
|
| 271406 |
+
"step": 135220
|
| 271407 |
+
},
|
| 271408 |
+
{
|
| 271409 |
+
"epoch": 1081.62,
|
| 271410 |
+
"learning_rate": 7.849807692307693e-06,
|
| 271411 |
+
"loss": 1.113,
|
| 271412 |
+
"step": 135225
|
| 271413 |
+
},
|
| 271414 |
+
{
|
| 271415 |
+
"epoch": 1081.66,
|
| 271416 |
+
"learning_rate": 7.849727564102565e-06,
|
| 271417 |
+
"loss": 0.3091,
|
| 271418 |
+
"step": 135230
|
| 271419 |
+
},
|
| 271420 |
+
{
|
| 271421 |
+
"epoch": 1081.7,
|
| 271422 |
+
"learning_rate": 7.849647435897436e-06,
|
| 271423 |
+
"loss": 0.2534,
|
| 271424 |
+
"step": 135235
|
| 271425 |
+
},
|
| 271426 |
+
{
|
| 271427 |
+
"epoch": 1081.74,
|
| 271428 |
+
"learning_rate": 7.849567307692309e-06,
|
| 271429 |
+
"loss": 0.3279,
|
| 271430 |
+
"step": 135240
|
| 271431 |
+
},
|
| 271432 |
+
{
|
| 271433 |
+
"epoch": 1081.78,
|
| 271434 |
+
"learning_rate": 7.84948717948718e-06,
|
| 271435 |
+
"loss": 0.5411,
|
| 271436 |
+
"step": 135245
|
| 271437 |
+
},
|
| 271438 |
+
{
|
| 271439 |
+
"epoch": 1081.82,
|
| 271440 |
+
"learning_rate": 7.849407051282052e-06,
|
| 271441 |
+
"loss": 0.9766,
|
| 271442 |
+
"step": 135250
|
| 271443 |
+
},
|
| 271444 |
+
{
|
| 271445 |
+
"epoch": 1081.86,
|
| 271446 |
+
"learning_rate": 7.849326923076925e-06,
|
| 271447 |
+
"loss": 0.2756,
|
| 271448 |
+
"step": 135255
|
| 271449 |
+
},
|
| 271450 |
+
{
|
| 271451 |
+
"epoch": 1081.9,
|
| 271452 |
+
"learning_rate": 7.849246794871796e-06,
|
| 271453 |
+
"loss": 0.2983,
|
| 271454 |
+
"step": 135260
|
| 271455 |
+
},
|
| 271456 |
+
{
|
| 271457 |
+
"epoch": 1081.94,
|
| 271458 |
+
"learning_rate": 7.849166666666668e-06,
|
| 271459 |
+
"loss": 0.3392,
|
| 271460 |
+
"step": 135265
|
| 271461 |
+
},
|
| 271462 |
+
{
|
| 271463 |
+
"epoch": 1081.98,
|
| 271464 |
+
"learning_rate": 7.849086538461539e-06,
|
| 271465 |
+
"loss": 0.5692,
|
| 271466 |
+
"step": 135270
|
| 271467 |
+
},
|
| 271468 |
+
{
|
| 271469 |
+
"epoch": 1082.0,
|
| 271470 |
+
"eval_loss": 0.3459949195384979,
|
| 271471 |
+
"eval_runtime": 35.3988,
|
| 271472 |
+
"eval_samples_per_second": 23.814,
|
| 271473 |
+
"eval_steps_per_second": 0.763,
|
| 271474 |
+
"eval_wer": 0.1782000448866612,
|
| 271475 |
+
"step": 135273
|
| 271476 |
+
},
|
| 271477 |
+
{
|
| 271478 |
+
"epoch": 1090.02,
|
| 271479 |
+
"learning_rate": 7.849006410256412e-06,
|
| 271480 |
+
"loss": 0.3767,
|
| 271481 |
+
"step": 135275
|
| 271482 |
+
},
|
| 271483 |
+
{
|
| 271484 |
+
"epoch": 1090.06,
|
| 271485 |
+
"learning_rate": 7.848926282051282e-06,
|
| 271486 |
+
"loss": 0.3471,
|
| 271487 |
+
"step": 135280
|
| 271488 |
+
},
|
| 271489 |
+
{
|
| 271490 |
+
"epoch": 1090.1,
|
| 271491 |
+
"learning_rate": 7.848846153846155e-06,
|
| 271492 |
+
"loss": 0.3465,
|
| 271493 |
+
"step": 135285
|
| 271494 |
+
},
|
| 271495 |
+
{
|
| 271496 |
+
"epoch": 1090.14,
|
| 271497 |
+
"learning_rate": 7.848766025641026e-06,
|
| 271498 |
+
"loss": 0.3366,
|
| 271499 |
+
"step": 135290
|
| 271500 |
+
},
|
| 271501 |
+
{
|
| 271502 |
+
"epoch": 1090.18,
|
| 271503 |
+
"learning_rate": 7.848685897435897e-06,
|
| 271504 |
+
"loss": 0.6436,
|
| 271505 |
+
"step": 135295
|
| 271506 |
+
},
|
| 271507 |
+
{
|
| 271508 |
+
"epoch": 1090.22,
|
| 271509 |
+
"learning_rate": 7.848605769230769e-06,
|
| 271510 |
+
"loss": 1.0347,
|
| 271511 |
+
"step": 135300
|
| 271512 |
+
},
|
| 271513 |
+
{
|
| 271514 |
+
"epoch": 1090.26,
|
| 271515 |
+
"learning_rate": 7.848525641025642e-06,
|
| 271516 |
+
"loss": 0.26,
|
| 271517 |
+
"step": 135305
|
| 271518 |
+
},
|
| 271519 |
+
{
|
| 271520 |
+
"epoch": 1090.3,
|
| 271521 |
+
"learning_rate": 7.848445512820513e-06,
|
| 271522 |
+
"loss": 0.3024,
|
| 271523 |
+
"step": 135310
|
| 271524 |
+
},
|
| 271525 |
+
{
|
| 271526 |
+
"epoch": 1090.34,
|
| 271527 |
+
"learning_rate": 7.848365384615385e-06,
|
| 271528 |
+
"loss": 0.3282,
|
| 271529 |
+
"step": 135315
|
| 271530 |
+
},
|
| 271531 |
+
{
|
| 271532 |
+
"epoch": 1090.38,
|
| 271533 |
+
"learning_rate": 7.848285256410258e-06,
|
| 271534 |
+
"loss": 0.6089,
|
| 271535 |
+
"step": 135320
|
| 271536 |
+
},
|
| 271537 |
+
{
|
| 271538 |
+
"epoch": 1090.42,
|
| 271539 |
+
"learning_rate": 7.848205128205129e-06,
|
| 271540 |
+
"loss": 1.1415,
|
| 271541 |
+
"step": 135325
|
| 271542 |
+
},
|
| 271543 |
+
{
|
| 271544 |
+
"epoch": 1090.46,
|
| 271545 |
+
"learning_rate": 7.848125e-06,
|
| 271546 |
+
"loss": 0.3656,
|
| 271547 |
+
"step": 135330
|
| 271548 |
+
},
|
| 271549 |
+
{
|
| 271550 |
+
"epoch": 1090.5,
|
| 271551 |
+
"learning_rate": 7.848044871794872e-06,
|
| 271552 |
+
"loss": 0.3207,
|
| 271553 |
+
"step": 135335
|
| 271554 |
+
},
|
| 271555 |
+
{
|
| 271556 |
+
"epoch": 1090.54,
|
| 271557 |
+
"learning_rate": 7.847964743589745e-06,
|
| 271558 |
+
"loss": 0.3475,
|
| 271559 |
+
"step": 135340
|
| 271560 |
+
},
|
| 271561 |
+
{
|
| 271562 |
+
"epoch": 1090.58,
|
| 271563 |
+
"learning_rate": 7.847884615384616e-06,
|
| 271564 |
+
"loss": 0.5692,
|
| 271565 |
+
"step": 135345
|
| 271566 |
+
},
|
| 271567 |
+
{
|
| 271568 |
+
"epoch": 1090.62,
|
| 271569 |
+
"learning_rate": 7.847804487179487e-06,
|
| 271570 |
+
"loss": 0.9884,
|
| 271571 |
+
"step": 135350
|
| 271572 |
+
},
|
| 271573 |
+
{
|
| 271574 |
+
"epoch": 1090.66,
|
| 271575 |
+
"learning_rate": 7.84772435897436e-06,
|
| 271576 |
+
"loss": 0.2562,
|
| 271577 |
+
"step": 135355
|
| 271578 |
+
},
|
| 271579 |
+
{
|
| 271580 |
+
"epoch": 1090.7,
|
| 271581 |
+
"learning_rate": 7.847644230769232e-06,
|
| 271582 |
+
"loss": 0.2487,
|
| 271583 |
+
"step": 135360
|
| 271584 |
+
},
|
| 271585 |
+
{
|
| 271586 |
+
"epoch": 1090.74,
|
| 271587 |
+
"learning_rate": 7.847564102564103e-06,
|
| 271588 |
+
"loss": 0.2955,
|
| 271589 |
+
"step": 135365
|
| 271590 |
+
},
|
| 271591 |
+
{
|
| 271592 |
+
"epoch": 1090.78,
|
| 271593 |
+
"learning_rate": 7.847483974358975e-06,
|
| 271594 |
+
"loss": 0.6426,
|
| 271595 |
+
"step": 135370
|
| 271596 |
+
},
|
| 271597 |
+
{
|
| 271598 |
+
"epoch": 1090.82,
|
| 271599 |
+
"learning_rate": 7.847403846153848e-06,
|
| 271600 |
+
"loss": 0.9612,
|
| 271601 |
+
"step": 135375
|
| 271602 |
+
},
|
| 271603 |
+
{
|
| 271604 |
+
"epoch": 1090.86,
|
| 271605 |
+
"learning_rate": 7.847323717948719e-06,
|
| 271606 |
+
"loss": 0.2593,
|
| 271607 |
+
"step": 135380
|
| 271608 |
+
},
|
| 271609 |
+
{
|
| 271610 |
+
"epoch": 1090.9,
|
| 271611 |
+
"learning_rate": 7.84724358974359e-06,
|
| 271612 |
+
"loss": 0.2933,
|
| 271613 |
+
"step": 135385
|
| 271614 |
+
},
|
| 271615 |
+
{
|
| 271616 |
+
"epoch": 1090.94,
|
| 271617 |
+
"learning_rate": 7.847163461538462e-06,
|
| 271618 |
+
"loss": 0.2808,
|
| 271619 |
+
"step": 135390
|
| 271620 |
+
},
|
| 271621 |
+
{
|
| 271622 |
+
"epoch": 1090.98,
|
| 271623 |
+
"learning_rate": 7.847083333333335e-06,
|
| 271624 |
+
"loss": 0.6968,
|
| 271625 |
+
"step": 135395
|
| 271626 |
+
},
|
| 271627 |
+
{
|
| 271628 |
+
"epoch": 1091.0,
|
| 271629 |
+
"eval_loss": 0.34540772438049316,
|
| 271630 |
+
"eval_runtime": 36.4518,
|
| 271631 |
+
"eval_samples_per_second": 23.126,
|
| 271632 |
+
"eval_steps_per_second": 0.741,
|
| 271633 |
+
"eval_wer": 0.1816446402349486,
|
| 271634 |
+
"step": 135397
|
| 271635 |
}
|
| 271636 |
],
|
| 271637 |
+
"max_steps": 620000,
|
| 271638 |
"num_train_epochs": 5000,
|
| 271639 |
+
"total_flos": 3.8100088804180695e+20,
|
| 271640 |
"trial_name": null,
|
| 271641 |
"trial_params": null
|
| 271642 |
}
|
model-bin/finetune/base/{checkpoint-134775 β checkpoint-135397}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1630177407.9573364/events.out.tfevents.1630177407.86bb0ddabf9b.4092.271
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1a3dda3568ab5f552119446cb3cbe35c0009c2821e8d07c77e1d2889d3761b69
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630177794.4337196/events.out.tfevents.1630177794.86bb0ddabf9b.4092.273
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c467f12b9e38ed64c9af29a80298de07aff56cf855d631c0e90b6f059c1ee768
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630178315.3923218/events.out.tfevents.1630178315.86bb0ddabf9b.4092.275
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1e3242fe1a17f9f89dfd0f467b76b6299b06ed50dc9f826bde91621473dd8e65
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630178704.6355941/events.out.tfevents.1630178704.86bb0ddabf9b.4092.277
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1d727c49e9a77ac0c60ef5c5c3b0f6ec125ad300234c34b529deee5b38173317
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630179100.382477/events.out.tfevents.1630179100.86bb0ddabf9b.4092.279
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d7ecb4cb435e563cc61e872dad96b16a5406ea5cd39f1b88c2e2ba233e9bd195
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630177407.86bb0ddabf9b.4092.270
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2e28e47456fb848b8e935cde2ee2f1f116631cc9b54d8b68c1920f582ee85114
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630177794.86bb0ddabf9b.4092.272
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b67964ffde4c1f31f76169b568249f46ccd1a91bac7d9006cafae444e7386f28
|
| 3 |
+
size 8462
|
model-bin/finetune/base/log/events.out.tfevents.1630178315.86bb0ddabf9b.4092.274
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0cff9e1011782159c7a91329877bbb6cd4d24baed9a89d99eac34ba78348b670
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630178704.86bb0ddabf9b.4092.276
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d44b259ceefb258da398ddcea6bea20fc03da55ece6595639d460c7307dfa2a3
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630179100.86bb0ddabf9b.4092.278
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ce4775844d8c6f45bba80a1a12304ff037c7ee877bc759898e14848c37337e46
|
| 3 |
+
size 8622
|