"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/trainer_state.json +793 -4
- model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630228288.1449773/events.out.tfevents.1630228288.cc93b136ebf5.1086.109 +3 -0
- model-bin/finetune/base/log/1630228721.9317095/events.out.tfevents.1630228721.cc93b136ebf5.1086.111 +3 -0
- model-bin/finetune/base/log/1630229152.9104557/events.out.tfevents.1630229152.cc93b136ebf5.1086.113 +3 -0
- model-bin/finetune/base/log/1630229586.5053954/events.out.tfevents.1630229586.cc93b136ebf5.1086.115 +3 -0
- model-bin/finetune/base/log/1630230014.08416/events.out.tfevents.1630230014.cc93b136ebf5.1086.117 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630228288.cc93b136ebf5.1086.108 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630228721.cc93b136ebf5.1086.110 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630229152.cc93b136ebf5.1086.112 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630229586.cc93b136ebf5.1086.114 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630230014.cc93b136ebf5.1086.116 +3 -0
model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6f7568a2befca944be90026427116813896320af24320ff0d6d6cce3345d3e75
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2a22ed2c49e0003da5e3bff5a28ee7e00afcf8ecb8920666d8355ae042273f4b
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:49db60bb7be014c03e20f099f8b5b41640dd5b10966a8d382a375eaacf0dba1f
|
| 3 |
size 14503
|
model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1e32e4b8d64538adb2e3a82fff0984d5e19137de747f31f2ccfe398a635069ee
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:75a35ac11ff6fdc1ff80807c5620a77c06912969c11e90901d0e650569700174
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.16716790141568252,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-145602",
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -284934,11 +284934,800 @@
|
|
| 284934 |
"eval_steps_per_second": 0.655,
|
| 284935 |
"eval_wer": 0.17012599228024178,
|
| 284936 |
"step": 145852
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 284937 |
}
|
| 284938 |
],
|
| 284939 |
-
"max_steps":
|
| 284940 |
"num_train_epochs": 5000,
|
| 284941 |
-
"total_flos": 4.
|
| 284942 |
"trial_name": null,
|
| 284943 |
"trial_params": null
|
| 284944 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.16716790141568252,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-145602",
|
| 4 |
+
"epoch": 1180.995983935743,
|
| 5 |
+
"global_step": 146472,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 284934 |
"eval_steps_per_second": 0.655,
|
| 284935 |
"eval_wer": 0.17012599228024178,
|
| 284936 |
"step": 145852
|
| 284937 |
+
},
|
| 284938 |
+
{
|
| 284939 |
+
"epoch": 1176.02,
|
| 284940 |
+
"learning_rate": 7.660775444264943e-06,
|
| 284941 |
+
"loss": 0.4127,
|
| 284942 |
+
"step": 145855
|
| 284943 |
+
},
|
| 284944 |
+
{
|
| 284945 |
+
"epoch": 1176.06,
|
| 284946 |
+
"learning_rate": 7.660694668820679e-06,
|
| 284947 |
+
"loss": 0.2818,
|
| 284948 |
+
"step": 145860
|
| 284949 |
+
},
|
| 284950 |
+
{
|
| 284951 |
+
"epoch": 1176.1,
|
| 284952 |
+
"learning_rate": 7.660613893376413e-06,
|
| 284953 |
+
"loss": 0.2966,
|
| 284954 |
+
"step": 145865
|
| 284955 |
+
},
|
| 284956 |
+
{
|
| 284957 |
+
"epoch": 1176.15,
|
| 284958 |
+
"learning_rate": 7.660533117932149e-06,
|
| 284959 |
+
"loss": 0.3262,
|
| 284960 |
+
"step": 145870
|
| 284961 |
+
},
|
| 284962 |
+
{
|
| 284963 |
+
"epoch": 1176.19,
|
| 284964 |
+
"learning_rate": 7.660452342487883e-06,
|
| 284965 |
+
"loss": 0.744,
|
| 284966 |
+
"step": 145875
|
| 284967 |
+
},
|
| 284968 |
+
{
|
| 284969 |
+
"epoch": 1176.23,
|
| 284970 |
+
"learning_rate": 7.660371567043619e-06,
|
| 284971 |
+
"loss": 0.8761,
|
| 284972 |
+
"step": 145880
|
| 284973 |
+
},
|
| 284974 |
+
{
|
| 284975 |
+
"epoch": 1176.27,
|
| 284976 |
+
"learning_rate": 7.660290791599355e-06,
|
| 284977 |
+
"loss": 0.2513,
|
| 284978 |
+
"step": 145885
|
| 284979 |
+
},
|
| 284980 |
+
{
|
| 284981 |
+
"epoch": 1176.31,
|
| 284982 |
+
"learning_rate": 7.660210016155089e-06,
|
| 284983 |
+
"loss": 0.2938,
|
| 284984 |
+
"step": 145890
|
| 284985 |
+
},
|
| 284986 |
+
{
|
| 284987 |
+
"epoch": 1176.35,
|
| 284988 |
+
"learning_rate": 7.660129240710825e-06,
|
| 284989 |
+
"loss": 0.3237,
|
| 284990 |
+
"step": 145895
|
| 284991 |
+
},
|
| 284992 |
+
{
|
| 284993 |
+
"epoch": 1176.39,
|
| 284994 |
+
"learning_rate": 7.660048465266559e-06,
|
| 284995 |
+
"loss": 0.7375,
|
| 284996 |
+
"step": 145900
|
| 284997 |
+
},
|
| 284998 |
+
{
|
| 284999 |
+
"epoch": 1176.43,
|
| 285000 |
+
"learning_rate": 7.659967689822295e-06,
|
| 285001 |
+
"loss": 0.8132,
|
| 285002 |
+
"step": 145905
|
| 285003 |
+
},
|
| 285004 |
+
{
|
| 285005 |
+
"epoch": 1176.47,
|
| 285006 |
+
"learning_rate": 7.659886914378029e-06,
|
| 285007 |
+
"loss": 0.3004,
|
| 285008 |
+
"step": 145910
|
| 285009 |
+
},
|
| 285010 |
+
{
|
| 285011 |
+
"epoch": 1176.51,
|
| 285012 |
+
"learning_rate": 7.659806138933765e-06,
|
| 285013 |
+
"loss": 0.2726,
|
| 285014 |
+
"step": 145915
|
| 285015 |
+
},
|
| 285016 |
+
{
|
| 285017 |
+
"epoch": 1176.55,
|
| 285018 |
+
"learning_rate": 7.659725363489499e-06,
|
| 285019 |
+
"loss": 0.3235,
|
| 285020 |
+
"step": 145920
|
| 285021 |
+
},
|
| 285022 |
+
{
|
| 285023 |
+
"epoch": 1176.59,
|
| 285024 |
+
"learning_rate": 7.659644588045235e-06,
|
| 285025 |
+
"loss": 0.7467,
|
| 285026 |
+
"step": 145925
|
| 285027 |
+
},
|
| 285028 |
+
{
|
| 285029 |
+
"epoch": 1176.63,
|
| 285030 |
+
"learning_rate": 7.659563812600969e-06,
|
| 285031 |
+
"loss": 0.7942,
|
| 285032 |
+
"step": 145930
|
| 285033 |
+
},
|
| 285034 |
+
{
|
| 285035 |
+
"epoch": 1176.67,
|
| 285036 |
+
"learning_rate": 7.659483037156705e-06,
|
| 285037 |
+
"loss": 0.2774,
|
| 285038 |
+
"step": 145935
|
| 285039 |
+
},
|
| 285040 |
+
{
|
| 285041 |
+
"epoch": 1176.71,
|
| 285042 |
+
"learning_rate": 7.659402261712439e-06,
|
| 285043 |
+
"loss": 0.2752,
|
| 285044 |
+
"step": 145940
|
| 285045 |
+
},
|
| 285046 |
+
{
|
| 285047 |
+
"epoch": 1176.75,
|
| 285048 |
+
"learning_rate": 7.659321486268175e-06,
|
| 285049 |
+
"loss": 0.3578,
|
| 285050 |
+
"step": 145945
|
| 285051 |
+
},
|
| 285052 |
+
{
|
| 285053 |
+
"epoch": 1176.79,
|
| 285054 |
+
"learning_rate": 7.659240710823909e-06,
|
| 285055 |
+
"loss": 0.6986,
|
| 285056 |
+
"step": 145950
|
| 285057 |
+
},
|
| 285058 |
+
{
|
| 285059 |
+
"epoch": 1176.83,
|
| 285060 |
+
"learning_rate": 7.659159935379645e-06,
|
| 285061 |
+
"loss": 0.7739,
|
| 285062 |
+
"step": 145955
|
| 285063 |
+
},
|
| 285064 |
+
{
|
| 285065 |
+
"epoch": 1176.87,
|
| 285066 |
+
"learning_rate": 7.65907915993538e-06,
|
| 285067 |
+
"loss": 0.2683,
|
| 285068 |
+
"step": 145960
|
| 285069 |
+
},
|
| 285070 |
+
{
|
| 285071 |
+
"epoch": 1176.91,
|
| 285072 |
+
"learning_rate": 7.658998384491115e-06,
|
| 285073 |
+
"loss": 0.3132,
|
| 285074 |
+
"step": 145965
|
| 285075 |
+
},
|
| 285076 |
+
{
|
| 285077 |
+
"epoch": 1176.95,
|
| 285078 |
+
"learning_rate": 7.65891760904685e-06,
|
| 285079 |
+
"loss": 0.373,
|
| 285080 |
+
"step": 145970
|
| 285081 |
+
},
|
| 285082 |
+
{
|
| 285083 |
+
"epoch": 1176.99,
|
| 285084 |
+
"learning_rate": 7.658836833602585e-06,
|
| 285085 |
+
"loss": 1.049,
|
| 285086 |
+
"step": 145975
|
| 285087 |
+
},
|
| 285088 |
+
{
|
| 285089 |
+
"epoch": 1177.0,
|
| 285090 |
+
"eval_loss": 0.3793235719203949,
|
| 285091 |
+
"eval_runtime": 40.3575,
|
| 285092 |
+
"eval_samples_per_second": 20.74,
|
| 285093 |
+
"eval_steps_per_second": 0.669,
|
| 285094 |
+
"eval_wer": 0.17174677608440797,
|
| 285095 |
+
"step": 145976
|
| 285096 |
+
},
|
| 285097 |
+
{
|
| 285098 |
+
"epoch": 1177.03,
|
| 285099 |
+
"learning_rate": 7.65875605815832e-06,
|
| 285100 |
+
"loss": 0.3211,
|
| 285101 |
+
"step": 145980
|
| 285102 |
+
},
|
| 285103 |
+
{
|
| 285104 |
+
"epoch": 1177.07,
|
| 285105 |
+
"learning_rate": 7.658675282714055e-06,
|
| 285106 |
+
"loss": 0.2587,
|
| 285107 |
+
"step": 145985
|
| 285108 |
+
},
|
| 285109 |
+
{
|
| 285110 |
+
"epoch": 1177.11,
|
| 285111 |
+
"learning_rate": 7.65859450726979e-06,
|
| 285112 |
+
"loss": 0.3043,
|
| 285113 |
+
"step": 145990
|
| 285114 |
+
},
|
| 285115 |
+
{
|
| 285116 |
+
"epoch": 1177.15,
|
| 285117 |
+
"learning_rate": 7.658513731825525e-06,
|
| 285118 |
+
"loss": 0.374,
|
| 285119 |
+
"step": 145995
|
| 285120 |
+
},
|
| 285121 |
+
{
|
| 285122 |
+
"epoch": 1177.19,
|
| 285123 |
+
"learning_rate": 7.65843295638126e-06,
|
| 285124 |
+
"loss": 0.9153,
|
| 285125 |
+
"step": 146000
|
| 285126 |
+
},
|
| 285127 |
+
{
|
| 285128 |
+
"epoch": 1177.23,
|
| 285129 |
+
"learning_rate": 7.658352180936995e-06,
|
| 285130 |
+
"loss": 0.7194,
|
| 285131 |
+
"step": 146005
|
| 285132 |
+
},
|
| 285133 |
+
{
|
| 285134 |
+
"epoch": 1177.27,
|
| 285135 |
+
"learning_rate": 7.65827140549273e-06,
|
| 285136 |
+
"loss": 0.2684,
|
| 285137 |
+
"step": 146010
|
| 285138 |
+
},
|
| 285139 |
+
{
|
| 285140 |
+
"epoch": 1177.31,
|
| 285141 |
+
"learning_rate": 7.658190630048465e-06,
|
| 285142 |
+
"loss": 0.2866,
|
| 285143 |
+
"step": 146015
|
| 285144 |
+
},
|
| 285145 |
+
{
|
| 285146 |
+
"epoch": 1177.35,
|
| 285147 |
+
"learning_rate": 7.6581098546042e-06,
|
| 285148 |
+
"loss": 0.3972,
|
| 285149 |
+
"step": 146020
|
| 285150 |
+
},
|
| 285151 |
+
{
|
| 285152 |
+
"epoch": 1177.39,
|
| 285153 |
+
"learning_rate": 7.658029079159936e-06,
|
| 285154 |
+
"loss": 1.0221,
|
| 285155 |
+
"step": 146025
|
| 285156 |
+
},
|
| 285157 |
+
{
|
| 285158 |
+
"epoch": 1177.43,
|
| 285159 |
+
"learning_rate": 7.65794830371567e-06,
|
| 285160 |
+
"loss": 0.5779,
|
| 285161 |
+
"step": 146030
|
| 285162 |
+
},
|
| 285163 |
+
{
|
| 285164 |
+
"epoch": 1177.47,
|
| 285165 |
+
"learning_rate": 7.657867528271406e-06,
|
| 285166 |
+
"loss": 0.3161,
|
| 285167 |
+
"step": 146035
|
| 285168 |
+
},
|
| 285169 |
+
{
|
| 285170 |
+
"epoch": 1177.51,
|
| 285171 |
+
"learning_rate": 7.65778675282714e-06,
|
| 285172 |
+
"loss": 0.2923,
|
| 285173 |
+
"step": 146040
|
| 285174 |
+
},
|
| 285175 |
+
{
|
| 285176 |
+
"epoch": 1177.55,
|
| 285177 |
+
"learning_rate": 7.657705977382876e-06,
|
| 285178 |
+
"loss": 0.4056,
|
| 285179 |
+
"step": 146045
|
| 285180 |
+
},
|
| 285181 |
+
{
|
| 285182 |
+
"epoch": 1177.59,
|
| 285183 |
+
"learning_rate": 7.65762520193861e-06,
|
| 285184 |
+
"loss": 0.8064,
|
| 285185 |
+
"step": 146050
|
| 285186 |
+
},
|
| 285187 |
+
{
|
| 285188 |
+
"epoch": 1177.63,
|
| 285189 |
+
"learning_rate": 7.657544426494346e-06,
|
| 285190 |
+
"loss": 0.6026,
|
| 285191 |
+
"step": 146055
|
| 285192 |
+
},
|
| 285193 |
+
{
|
| 285194 |
+
"epoch": 1177.67,
|
| 285195 |
+
"learning_rate": 7.65746365105008e-06,
|
| 285196 |
+
"loss": 0.2731,
|
| 285197 |
+
"step": 146060
|
| 285198 |
+
},
|
| 285199 |
+
{
|
| 285200 |
+
"epoch": 1177.71,
|
| 285201 |
+
"learning_rate": 7.657382875605816e-06,
|
| 285202 |
+
"loss": 0.2761,
|
| 285203 |
+
"step": 146065
|
| 285204 |
+
},
|
| 285205 |
+
{
|
| 285206 |
+
"epoch": 1177.76,
|
| 285207 |
+
"learning_rate": 7.65730210016155e-06,
|
| 285208 |
+
"loss": 0.3563,
|
| 285209 |
+
"step": 146070
|
| 285210 |
+
},
|
| 285211 |
+
{
|
| 285212 |
+
"epoch": 1177.8,
|
| 285213 |
+
"learning_rate": 7.657221324717286e-06,
|
| 285214 |
+
"loss": 0.8102,
|
| 285215 |
+
"step": 146075
|
| 285216 |
+
},
|
| 285217 |
+
{
|
| 285218 |
+
"epoch": 1177.84,
|
| 285219 |
+
"learning_rate": 7.65714054927302e-06,
|
| 285220 |
+
"loss": 0.5813,
|
| 285221 |
+
"step": 146080
|
| 285222 |
+
},
|
| 285223 |
+
{
|
| 285224 |
+
"epoch": 1177.88,
|
| 285225 |
+
"learning_rate": 7.657059773828756e-06,
|
| 285226 |
+
"loss": 0.3342,
|
| 285227 |
+
"step": 146085
|
| 285228 |
+
},
|
| 285229 |
+
{
|
| 285230 |
+
"epoch": 1177.92,
|
| 285231 |
+
"learning_rate": 7.656978998384492e-06,
|
| 285232 |
+
"loss": 0.3355,
|
| 285233 |
+
"step": 146090
|
| 285234 |
+
},
|
| 285235 |
+
{
|
| 285236 |
+
"epoch": 1177.96,
|
| 285237 |
+
"learning_rate": 7.656898222940226e-06,
|
| 285238 |
+
"loss": 0.4923,
|
| 285239 |
+
"step": 146095
|
| 285240 |
+
},
|
| 285241 |
+
{
|
| 285242 |
+
"epoch": 1178.0,
|
| 285243 |
+
"learning_rate": 7.656817447495962e-06,
|
| 285244 |
+
"loss": 1.2601,
|
| 285245 |
+
"step": 146100
|
| 285246 |
+
},
|
| 285247 |
+
{
|
| 285248 |
+
"epoch": 1178.0,
|
| 285249 |
+
"eval_loss": 0.4012053608894348,
|
| 285250 |
+
"eval_runtime": 40.577,
|
| 285251 |
+
"eval_samples_per_second": 20.627,
|
| 285252 |
+
"eval_steps_per_second": 0.665,
|
| 285253 |
+
"eval_wer": 0.1786023678891135,
|
| 285254 |
+
"step": 146100
|
| 285255 |
+
},
|
| 285256 |
+
{
|
| 285257 |
+
"epoch": 1178.04,
|
| 285258 |
+
"learning_rate": 7.656736672051696e-06,
|
| 285259 |
+
"loss": 0.2932,
|
| 285260 |
+
"step": 146105
|
| 285261 |
+
},
|
| 285262 |
+
{
|
| 285263 |
+
"epoch": 1178.08,
|
| 285264 |
+
"learning_rate": 7.656655896607432e-06,
|
| 285265 |
+
"loss": 0.2603,
|
| 285266 |
+
"step": 146110
|
| 285267 |
+
},
|
| 285268 |
+
{
|
| 285269 |
+
"epoch": 1178.12,
|
| 285270 |
+
"learning_rate": 7.656575121163166e-06,
|
| 285271 |
+
"loss": 0.2993,
|
| 285272 |
+
"step": 146115
|
| 285273 |
+
},
|
| 285274 |
+
{
|
| 285275 |
+
"epoch": 1178.16,
|
| 285276 |
+
"learning_rate": 7.656494345718902e-06,
|
| 285277 |
+
"loss": 0.4776,
|
| 285278 |
+
"step": 146120
|
| 285279 |
+
},
|
| 285280 |
+
{
|
| 285281 |
+
"epoch": 1178.2,
|
| 285282 |
+
"learning_rate": 7.656413570274636e-06,
|
| 285283 |
+
"loss": 1.1354,
|
| 285284 |
+
"step": 146125
|
| 285285 |
+
},
|
| 285286 |
+
{
|
| 285287 |
+
"epoch": 1178.24,
|
| 285288 |
+
"learning_rate": 7.656332794830372e-06,
|
| 285289 |
+
"loss": 0.2897,
|
| 285290 |
+
"step": 146130
|
| 285291 |
+
},
|
| 285292 |
+
{
|
| 285293 |
+
"epoch": 1178.28,
|
| 285294 |
+
"learning_rate": 7.656252019386106e-06,
|
| 285295 |
+
"loss": 0.2256,
|
| 285296 |
+
"step": 146135
|
| 285297 |
+
},
|
| 285298 |
+
{
|
| 285299 |
+
"epoch": 1178.32,
|
| 285300 |
+
"learning_rate": 7.656171243941842e-06,
|
| 285301 |
+
"loss": 0.2845,
|
| 285302 |
+
"step": 146140
|
| 285303 |
+
},
|
| 285304 |
+
{
|
| 285305 |
+
"epoch": 1178.36,
|
| 285306 |
+
"learning_rate": 7.656090468497578e-06,
|
| 285307 |
+
"loss": 0.4063,
|
| 285308 |
+
"step": 146145
|
| 285309 |
+
},
|
| 285310 |
+
{
|
| 285311 |
+
"epoch": 1178.4,
|
| 285312 |
+
"learning_rate": 7.656009693053312e-06,
|
| 285313 |
+
"loss": 1.1799,
|
| 285314 |
+
"step": 146150
|
| 285315 |
+
},
|
| 285316 |
+
{
|
| 285317 |
+
"epoch": 1178.44,
|
| 285318 |
+
"learning_rate": 7.655928917609048e-06,
|
| 285319 |
+
"loss": 0.3141,
|
| 285320 |
+
"step": 146155
|
| 285321 |
+
},
|
| 285322 |
+
{
|
| 285323 |
+
"epoch": 1178.48,
|
| 285324 |
+
"learning_rate": 7.655848142164782e-06,
|
| 285325 |
+
"loss": 0.2586,
|
| 285326 |
+
"step": 146160
|
| 285327 |
+
},
|
| 285328 |
+
{
|
| 285329 |
+
"epoch": 1178.52,
|
| 285330 |
+
"learning_rate": 7.655767366720518e-06,
|
| 285331 |
+
"loss": 0.3502,
|
| 285332 |
+
"step": 146165
|
| 285333 |
+
},
|
| 285334 |
+
{
|
| 285335 |
+
"epoch": 1178.56,
|
| 285336 |
+
"learning_rate": 7.655686591276252e-06,
|
| 285337 |
+
"loss": 0.4228,
|
| 285338 |
+
"step": 146170
|
| 285339 |
+
},
|
| 285340 |
+
{
|
| 285341 |
+
"epoch": 1178.6,
|
| 285342 |
+
"learning_rate": 7.655605815831988e-06,
|
| 285343 |
+
"loss": 1.2601,
|
| 285344 |
+
"step": 146175
|
| 285345 |
+
},
|
| 285346 |
+
{
|
| 285347 |
+
"epoch": 1178.64,
|
| 285348 |
+
"learning_rate": 7.655525040387722e-06,
|
| 285349 |
+
"loss": 0.2679,
|
| 285350 |
+
"step": 146180
|
| 285351 |
+
},
|
| 285352 |
+
{
|
| 285353 |
+
"epoch": 1178.68,
|
| 285354 |
+
"learning_rate": 7.655444264943458e-06,
|
| 285355 |
+
"loss": 0.2357,
|
| 285356 |
+
"step": 146185
|
| 285357 |
+
},
|
| 285358 |
+
{
|
| 285359 |
+
"epoch": 1178.72,
|
| 285360 |
+
"learning_rate": 7.655363489499192e-06,
|
| 285361 |
+
"loss": 0.3045,
|
| 285362 |
+
"step": 146190
|
| 285363 |
+
},
|
| 285364 |
+
{
|
| 285365 |
+
"epoch": 1178.76,
|
| 285366 |
+
"learning_rate": 7.655282714054928e-06,
|
| 285367 |
+
"loss": 0.5178,
|
| 285368 |
+
"step": 146195
|
| 285369 |
+
},
|
| 285370 |
+
{
|
| 285371 |
+
"epoch": 1178.8,
|
| 285372 |
+
"learning_rate": 7.655201938610663e-06,
|
| 285373 |
+
"loss": 1.2614,
|
| 285374 |
+
"step": 146200
|
| 285375 |
+
},
|
| 285376 |
+
{
|
| 285377 |
+
"epoch": 1178.84,
|
| 285378 |
+
"learning_rate": 7.655121163166397e-06,
|
| 285379 |
+
"loss": 0.3759,
|
| 285380 |
+
"step": 146205
|
| 285381 |
+
},
|
| 285382 |
+
{
|
| 285383 |
+
"epoch": 1178.88,
|
| 285384 |
+
"learning_rate": 7.655040387722133e-06,
|
| 285385 |
+
"loss": 0.2394,
|
| 285386 |
+
"step": 146210
|
| 285387 |
+
},
|
| 285388 |
+
{
|
| 285389 |
+
"epoch": 1178.92,
|
| 285390 |
+
"learning_rate": 7.654959612277867e-06,
|
| 285391 |
+
"loss": 0.275,
|
| 285392 |
+
"step": 146215
|
| 285393 |
+
},
|
| 285394 |
+
{
|
| 285395 |
+
"epoch": 1178.96,
|
| 285396 |
+
"learning_rate": 7.654878836833603e-06,
|
| 285397 |
+
"loss": 0.5791,
|
| 285398 |
+
"step": 146220
|
| 285399 |
+
},
|
| 285400 |
+
{
|
| 285401 |
+
"epoch": 1179.0,
|
| 285402 |
+
"eval_loss": 0.358010470867157,
|
| 285403 |
+
"eval_runtime": 40.5201,
|
| 285404 |
+
"eval_samples_per_second": 20.656,
|
| 285405 |
+
"eval_steps_per_second": 0.666,
|
| 285406 |
+
"eval_wer": 0.176530911221799,
|
| 285407 |
+
"step": 146224
|
| 285408 |
+
},
|
| 285409 |
+
{
|
| 285410 |
+
"epoch": 1179.01,
|
| 285411 |
+
"learning_rate": 7.654798061389337e-06,
|
| 285412 |
+
"loss": 0.5013,
|
| 285413 |
+
"step": 146225
|
| 285414 |
+
},
|
| 285415 |
+
{
|
| 285416 |
+
"epoch": 1179.05,
|
| 285417 |
+
"learning_rate": 7.654717285945073e-06,
|
| 285418 |
+
"loss": 0.2654,
|
| 285419 |
+
"step": 146230
|
| 285420 |
+
},
|
| 285421 |
+
{
|
| 285422 |
+
"epoch": 1179.09,
|
| 285423 |
+
"learning_rate": 7.654636510500807e-06,
|
| 285424 |
+
"loss": 0.2647,
|
| 285425 |
+
"step": 146235
|
| 285426 |
+
},
|
| 285427 |
+
{
|
| 285428 |
+
"epoch": 1179.13,
|
| 285429 |
+
"learning_rate": 7.654555735056543e-06,
|
| 285430 |
+
"loss": 0.3248,
|
| 285431 |
+
"step": 146240
|
| 285432 |
+
},
|
| 285433 |
+
{
|
| 285434 |
+
"epoch": 1179.17,
|
| 285435 |
+
"learning_rate": 7.654474959612277e-06,
|
| 285436 |
+
"loss": 0.4694,
|
| 285437 |
+
"step": 146245
|
| 285438 |
+
},
|
| 285439 |
+
{
|
| 285440 |
+
"epoch": 1179.21,
|
| 285441 |
+
"learning_rate": 7.654394184168013e-06,
|
| 285442 |
+
"loss": 1.0148,
|
| 285443 |
+
"step": 146250
|
| 285444 |
+
},
|
| 285445 |
+
{
|
| 285446 |
+
"epoch": 1179.25,
|
| 285447 |
+
"learning_rate": 7.654313408723747e-06,
|
| 285448 |
+
"loss": 0.2833,
|
| 285449 |
+
"step": 146255
|
| 285450 |
+
},
|
| 285451 |
+
{
|
| 285452 |
+
"epoch": 1179.29,
|
| 285453 |
+
"learning_rate": 7.654232633279483e-06,
|
| 285454 |
+
"loss": 0.2983,
|
| 285455 |
+
"step": 146260
|
| 285456 |
+
},
|
| 285457 |
+
{
|
| 285458 |
+
"epoch": 1179.33,
|
| 285459 |
+
"learning_rate": 7.654151857835219e-06,
|
| 285460 |
+
"loss": 0.3522,
|
| 285461 |
+
"step": 146265
|
| 285462 |
+
},
|
| 285463 |
+
{
|
| 285464 |
+
"epoch": 1179.37,
|
| 285465 |
+
"learning_rate": 7.654071082390953e-06,
|
| 285466 |
+
"loss": 0.5271,
|
| 285467 |
+
"step": 146270
|
| 285468 |
+
},
|
| 285469 |
+
{
|
| 285470 |
+
"epoch": 1179.41,
|
| 285471 |
+
"learning_rate": 7.653990306946689e-06,
|
| 285472 |
+
"loss": 1.1747,
|
| 285473 |
+
"step": 146275
|
| 285474 |
+
},
|
| 285475 |
+
{
|
| 285476 |
+
"epoch": 1179.45,
|
| 285477 |
+
"learning_rate": 7.653909531502423e-06,
|
| 285478 |
+
"loss": 0.2796,
|
| 285479 |
+
"step": 146280
|
| 285480 |
+
},
|
| 285481 |
+
{
|
| 285482 |
+
"epoch": 1179.49,
|
| 285483 |
+
"learning_rate": 7.653828756058159e-06,
|
| 285484 |
+
"loss": 0.2518,
|
| 285485 |
+
"step": 146285
|
| 285486 |
+
},
|
| 285487 |
+
{
|
| 285488 |
+
"epoch": 1179.53,
|
| 285489 |
+
"learning_rate": 7.653747980613893e-06,
|
| 285490 |
+
"loss": 0.307,
|
| 285491 |
+
"step": 146290
|
| 285492 |
+
},
|
| 285493 |
+
{
|
| 285494 |
+
"epoch": 1179.57,
|
| 285495 |
+
"learning_rate": 7.653667205169629e-06,
|
| 285496 |
+
"loss": 0.4902,
|
| 285497 |
+
"step": 146295
|
| 285498 |
+
},
|
| 285499 |
+
{
|
| 285500 |
+
"epoch": 1179.61,
|
| 285501 |
+
"learning_rate": 7.653586429725363e-06,
|
| 285502 |
+
"loss": 1.0019,
|
| 285503 |
+
"step": 146300
|
| 285504 |
+
},
|
| 285505 |
+
{
|
| 285506 |
+
"epoch": 1179.65,
|
| 285507 |
+
"learning_rate": 7.653505654281099e-06,
|
| 285508 |
+
"loss": 0.278,
|
| 285509 |
+
"step": 146305
|
| 285510 |
+
},
|
| 285511 |
+
{
|
| 285512 |
+
"epoch": 1179.69,
|
| 285513 |
+
"learning_rate": 7.653424878836833e-06,
|
| 285514 |
+
"loss": 0.3384,
|
| 285515 |
+
"step": 146310
|
| 285516 |
+
},
|
| 285517 |
+
{
|
| 285518 |
+
"epoch": 1179.73,
|
| 285519 |
+
"learning_rate": 7.653344103392569e-06,
|
| 285520 |
+
"loss": 0.3353,
|
| 285521 |
+
"step": 146315
|
| 285522 |
+
},
|
| 285523 |
+
{
|
| 285524 |
+
"epoch": 1179.77,
|
| 285525 |
+
"learning_rate": 7.653263327948305e-06,
|
| 285526 |
+
"loss": 0.5559,
|
| 285527 |
+
"step": 146320
|
| 285528 |
+
},
|
| 285529 |
+
{
|
| 285530 |
+
"epoch": 1179.81,
|
| 285531 |
+
"learning_rate": 7.653182552504039e-06,
|
| 285532 |
+
"loss": 1.4385,
|
| 285533 |
+
"step": 146325
|
| 285534 |
+
},
|
| 285535 |
+
{
|
| 285536 |
+
"epoch": 1179.85,
|
| 285537 |
+
"learning_rate": 7.653101777059775e-06,
|
| 285538 |
+
"loss": 0.2435,
|
| 285539 |
+
"step": 146330
|
| 285540 |
+
},
|
| 285541 |
+
{
|
| 285542 |
+
"epoch": 1179.9,
|
| 285543 |
+
"learning_rate": 7.653021001615509e-06,
|
| 285544 |
+
"loss": 0.2579,
|
| 285545 |
+
"step": 146335
|
| 285546 |
+
},
|
| 285547 |
+
{
|
| 285548 |
+
"epoch": 1179.94,
|
| 285549 |
+
"learning_rate": 7.652940226171245e-06,
|
| 285550 |
+
"loss": 0.345,
|
| 285551 |
+
"step": 146340
|
| 285552 |
+
},
|
| 285553 |
+
{
|
| 285554 |
+
"epoch": 1179.98,
|
| 285555 |
+
"learning_rate": 7.652859450726979e-06,
|
| 285556 |
+
"loss": 0.5621,
|
| 285557 |
+
"step": 146345
|
| 285558 |
+
},
|
| 285559 |
+
{
|
| 285560 |
+
"epoch": 1180.0,
|
| 285561 |
+
"eval_loss": 0.449453204870224,
|
| 285562 |
+
"eval_runtime": 40.0278,
|
| 285563 |
+
"eval_samples_per_second": 20.91,
|
| 285564 |
+
"eval_steps_per_second": 0.675,
|
| 285565 |
+
"eval_wer": 0.18728573557907288,
|
| 285566 |
+
"step": 146348
|
| 285567 |
+
},
|
| 285568 |
+
{
|
| 285569 |
+
"epoch": 1180.02,
|
| 285570 |
+
"learning_rate": 7.652778675282715e-06,
|
| 285571 |
+
"loss": 0.2882,
|
| 285572 |
+
"step": 146350
|
| 285573 |
+
},
|
| 285574 |
+
{
|
| 285575 |
+
"epoch": 1180.06,
|
| 285576 |
+
"learning_rate": 7.652697899838449e-06,
|
| 285577 |
+
"loss": 0.2908,
|
| 285578 |
+
"step": 146355
|
| 285579 |
+
},
|
| 285580 |
+
{
|
| 285581 |
+
"epoch": 1180.1,
|
| 285582 |
+
"learning_rate": 7.652617124394185e-06,
|
| 285583 |
+
"loss": 0.2317,
|
| 285584 |
+
"step": 146360
|
| 285585 |
+
},
|
| 285586 |
+
{
|
| 285587 |
+
"epoch": 1180.14,
|
| 285588 |
+
"learning_rate": 7.652536348949919e-06,
|
| 285589 |
+
"loss": 0.3416,
|
| 285590 |
+
"step": 146365
|
| 285591 |
+
},
|
| 285592 |
+
{
|
| 285593 |
+
"epoch": 1180.18,
|
| 285594 |
+
"learning_rate": 7.652455573505655e-06,
|
| 285595 |
+
"loss": 0.6812,
|
| 285596 |
+
"step": 146370
|
| 285597 |
+
},
|
| 285598 |
+
{
|
| 285599 |
+
"epoch": 1180.22,
|
| 285600 |
+
"learning_rate": 7.65237479806139e-06,
|
| 285601 |
+
"loss": 0.8339,
|
| 285602 |
+
"step": 146375
|
| 285603 |
+
},
|
| 285604 |
+
{
|
| 285605 |
+
"epoch": 1180.26,
|
| 285606 |
+
"learning_rate": 7.652294022617125e-06,
|
| 285607 |
+
"loss": 0.431,
|
| 285608 |
+
"step": 146380
|
| 285609 |
+
},
|
| 285610 |
+
{
|
| 285611 |
+
"epoch": 1180.3,
|
| 285612 |
+
"learning_rate": 7.65221324717286e-06,
|
| 285613 |
+
"loss": 0.3809,
|
| 285614 |
+
"step": 146385
|
| 285615 |
+
},
|
| 285616 |
+
{
|
| 285617 |
+
"epoch": 1180.34,
|
| 285618 |
+
"learning_rate": 7.652132471728595e-06,
|
| 285619 |
+
"loss": 0.3004,
|
| 285620 |
+
"step": 146390
|
| 285621 |
+
},
|
| 285622 |
+
{
|
| 285623 |
+
"epoch": 1180.38,
|
| 285624 |
+
"learning_rate": 7.65205169628433e-06,
|
| 285625 |
+
"loss": 0.578,
|
| 285626 |
+
"step": 146395
|
| 285627 |
+
},
|
| 285628 |
+
{
|
| 285629 |
+
"epoch": 1180.42,
|
| 285630 |
+
"learning_rate": 7.651970920840065e-06,
|
| 285631 |
+
"loss": 0.9797,
|
| 285632 |
+
"step": 146400
|
| 285633 |
+
},
|
| 285634 |
+
{
|
| 285635 |
+
"epoch": 1180.46,
|
| 285636 |
+
"learning_rate": 7.6518901453958e-06,
|
| 285637 |
+
"loss": 0.2545,
|
| 285638 |
+
"step": 146405
|
| 285639 |
+
},
|
| 285640 |
+
{
|
| 285641 |
+
"epoch": 1180.5,
|
| 285642 |
+
"learning_rate": 7.651809369951535e-06,
|
| 285643 |
+
"loss": 0.2621,
|
| 285644 |
+
"step": 146410
|
| 285645 |
+
},
|
| 285646 |
+
{
|
| 285647 |
+
"epoch": 1180.54,
|
| 285648 |
+
"learning_rate": 7.65172859450727e-06,
|
| 285649 |
+
"loss": 0.3193,
|
| 285650 |
+
"step": 146415
|
| 285651 |
+
},
|
| 285652 |
+
{
|
| 285653 |
+
"epoch": 1180.58,
|
| 285654 |
+
"learning_rate": 7.651647819063005e-06,
|
| 285655 |
+
"loss": 0.6539,
|
| 285656 |
+
"step": 146420
|
| 285657 |
+
},
|
| 285658 |
+
{
|
| 285659 |
+
"epoch": 1180.62,
|
| 285660 |
+
"learning_rate": 7.65156704361874e-06,
|
| 285661 |
+
"loss": 1.0436,
|
| 285662 |
+
"step": 146425
|
| 285663 |
+
},
|
| 285664 |
+
{
|
| 285665 |
+
"epoch": 1180.66,
|
| 285666 |
+
"learning_rate": 7.651486268174474e-06,
|
| 285667 |
+
"loss": 0.2563,
|
| 285668 |
+
"step": 146430
|
| 285669 |
+
},
|
| 285670 |
+
{
|
| 285671 |
+
"epoch": 1180.7,
|
| 285672 |
+
"learning_rate": 7.65140549273021e-06,
|
| 285673 |
+
"loss": 0.3142,
|
| 285674 |
+
"step": 146435
|
| 285675 |
+
},
|
| 285676 |
+
{
|
| 285677 |
+
"epoch": 1180.74,
|
| 285678 |
+
"learning_rate": 7.651324717285946e-06,
|
| 285679 |
+
"loss": 0.3735,
|
| 285680 |
+
"step": 146440
|
| 285681 |
+
},
|
| 285682 |
+
{
|
| 285683 |
+
"epoch": 1180.78,
|
| 285684 |
+
"learning_rate": 7.65124394184168e-06,
|
| 285685 |
+
"loss": 0.6211,
|
| 285686 |
+
"step": 146445
|
| 285687 |
+
},
|
| 285688 |
+
{
|
| 285689 |
+
"epoch": 1180.82,
|
| 285690 |
+
"learning_rate": 7.651163166397416e-06,
|
| 285691 |
+
"loss": 1.0659,
|
| 285692 |
+
"step": 146450
|
| 285693 |
+
},
|
| 285694 |
+
{
|
| 285695 |
+
"epoch": 1180.86,
|
| 285696 |
+
"learning_rate": 7.65108239095315e-06,
|
| 285697 |
+
"loss": 0.3007,
|
| 285698 |
+
"step": 146455
|
| 285699 |
+
},
|
| 285700 |
+
{
|
| 285701 |
+
"epoch": 1180.9,
|
| 285702 |
+
"learning_rate": 7.651001615508886e-06,
|
| 285703 |
+
"loss": 0.2761,
|
| 285704 |
+
"step": 146460
|
| 285705 |
+
},
|
| 285706 |
+
{
|
| 285707 |
+
"epoch": 1180.94,
|
| 285708 |
+
"learning_rate": 7.65092084006462e-06,
|
| 285709 |
+
"loss": 0.3487,
|
| 285710 |
+
"step": 146465
|
| 285711 |
+
},
|
| 285712 |
+
{
|
| 285713 |
+
"epoch": 1180.98,
|
| 285714 |
+
"learning_rate": 7.650840064620356e-06,
|
| 285715 |
+
"loss": 0.7013,
|
| 285716 |
+
"step": 146470
|
| 285717 |
+
},
|
| 285718 |
+
{
|
| 285719 |
+
"epoch": 1181.0,
|
| 285720 |
+
"eval_loss": 0.34985774755477905,
|
| 285721 |
+
"eval_runtime": 41.8541,
|
| 285722 |
+
"eval_samples_per_second": 19.998,
|
| 285723 |
+
"eval_steps_per_second": 0.645,
|
| 285724 |
+
"eval_wer": 0.1758169934640523,
|
| 285725 |
+
"step": 146472
|
| 285726 |
}
|
| 285727 |
],
|
| 285728 |
+
"max_steps": 620000,
|
| 285729 |
"num_train_epochs": 5000,
|
| 285730 |
+
"total_flos": 4.1218134700506854e+20,
|
| 285731 |
"trial_name": null,
|
| 285732 |
"trial_params": null
|
| 285733 |
}
|
model-bin/finetune/base/{checkpoint-145852 β checkpoint-146472}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1630228288.1449773/events.out.tfevents.1630228288.cc93b136ebf5.1086.109
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dfe38cec38c3d2adbc09d27a46a1e57a1f70fa126623f4f52627be155109468e
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630228721.9317095/events.out.tfevents.1630228721.cc93b136ebf5.1086.111
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:28112e380a1179c13caf1bfbf7221311ccc001134b60d565f8f56c89d707889c
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630229152.9104557/events.out.tfevents.1630229152.cc93b136ebf5.1086.113
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6748d316e70868842a9f08d249dc07f834ef93882afb2c7c7fd28f72375b2422
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630229586.5053954/events.out.tfevents.1630229586.cc93b136ebf5.1086.115
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:feb8a302b12fa256e6d48cbf4c25ff505338ca01049c4f58a77b6d4f4b35098b
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630230014.08416/events.out.tfevents.1630230014.cc93b136ebf5.1086.117
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5e6b0961bc142db049474b69ea745d4c3a673f78a77e8b18e020138d7fee8b38
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630228288.cc93b136ebf5.1086.108
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:630bfa7f41e9ac1296d8c100a479675d8878d9a89021acbe1d1e3fd54480eaec
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630228721.cc93b136ebf5.1086.110
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:cc375cfeb2958a56e334f99323d9a111298cd5b84f7f763f7c9f2e5c1b031302
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630229152.cc93b136ebf5.1086.112
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:716525fcb37cbd537930bceaeec45c754a919f6732b733a07571f7ff0a0ea616
|
| 3 |
+
size 8462
|
model-bin/finetune/base/log/events.out.tfevents.1630229586.cc93b136ebf5.1086.114
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:933a727cbd0e343f0d73aec1a0317b5741e8e77cd32fdda735bd47490f98c084
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630230014.cc93b136ebf5.1086.116
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:931819ecb40a82e206200184d2b3d92970a599c9a49de08dcace11c58b3ff38f
|
| 3 |
+
size 8622
|