"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-74539 β checkpoint-75162}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-74539 β checkpoint-75162}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-74539 β checkpoint-75162}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-74539 β checkpoint-75162}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-74539 β checkpoint-75162}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-74539 β checkpoint-75162}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-74539 β checkpoint-75162}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-74539 β checkpoint-75162}/trainer_state.json +799 -4
- model-bin/finetune/base/{checkpoint-74539 β checkpoint-75162}/training_args.bin +0 -0
- model-bin/finetune/base/log/1629879092.38512/events.out.tfevents.1629879092.7e498afd5545.905.83 +3 -0
- model-bin/finetune/base/log/1629879755.9543498/events.out.tfevents.1629879755.7e498afd5545.905.85 +3 -0
- model-bin/finetune/base/log/1629880403.5211415/events.out.tfevents.1629880403.7e498afd5545.905.87 +3 -0
- model-bin/finetune/base/log/1629881049.788205/events.out.tfevents.1629881049.7e498afd5545.905.89 +3 -0
- model-bin/finetune/base/log/1629881697.4184577/events.out.tfevents.1629881697.7e498afd5545.905.91 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629879092.7e498afd5545.905.82 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629879755.7e498afd5545.905.84 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629880403.7e498afd5545.905.86 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629881049.7e498afd5545.905.88 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629881697.7e498afd5545.905.90 +3 -0
model-bin/finetune/base/{checkpoint-74539 β checkpoint-75162}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-74539 β checkpoint-75162}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:54b35bdaaf4933af7fe3f9bbd9b1424d570c568987740f15bcd85cc870d83f51
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-74539 β checkpoint-75162}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-74539 β checkpoint-75162}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f70ffbe9923a30dbf8b87abb1106958feef550de2620a2d2b80cc91fc8a098e2
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-74539 β checkpoint-75162}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4be019330451dcdea1470053a969250ec2ac7591e9e422df7bf9734de62a6607
|
| 3 |
size 14503
|
model-bin/finetune/base/{checkpoint-74539 β checkpoint-75162}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e3d8eee4042019854a091a06285230ead9805a530d0a1e8630cf2695e05d1d3b
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-74539 β checkpoint-75162}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8aae4596cc6445369fecda3c5539d240321af736664343a1c7e72481ab8edb80
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-74539 β checkpoint-75162}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.18412114350410416,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-69565",
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -194199,11 +194199,806 @@
|
|
| 194199 |
"eval_steps_per_second": 0.666,
|
| 194200 |
"eval_wer": 0.18913761467889909,
|
| 194201 |
"step": 74539
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 194202 |
}
|
| 194203 |
],
|
| 194204 |
-
"max_steps":
|
| 194205 |
"num_train_epochs": 5000,
|
| 194206 |
-
"total_flos": 2.
|
| 194207 |
"trial_name": null,
|
| 194208 |
"trial_params": null
|
| 194209 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.18412114350410416,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-69565",
|
| 4 |
+
"epoch": 601.0,
|
| 5 |
+
"global_step": 75162,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 194199 |
"eval_steps_per_second": 0.666,
|
| 194200 |
"eval_wer": 0.18913761467889909,
|
| 194201 |
"step": 74539
|
| 194202 |
+
},
|
| 194203 |
+
{
|
| 194204 |
+
"epoch": 596.01,
|
| 194205 |
+
"learning_rate": 8.821939102564104e-06,
|
| 194206 |
+
"loss": 0.4295,
|
| 194207 |
+
"step": 74540
|
| 194208 |
+
},
|
| 194209 |
+
{
|
| 194210 |
+
"epoch": 596.05,
|
| 194211 |
+
"learning_rate": 8.821858974358975e-06,
|
| 194212 |
+
"loss": 0.3124,
|
| 194213 |
+
"step": 74545
|
| 194214 |
+
},
|
| 194215 |
+
{
|
| 194216 |
+
"epoch": 596.09,
|
| 194217 |
+
"learning_rate": 8.821778846153846e-06,
|
| 194218 |
+
"loss": 0.3017,
|
| 194219 |
+
"step": 74550
|
| 194220 |
+
},
|
| 194221 |
+
{
|
| 194222 |
+
"epoch": 596.13,
|
| 194223 |
+
"learning_rate": 8.82169871794872e-06,
|
| 194224 |
+
"loss": 0.3852,
|
| 194225 |
+
"step": 74555
|
| 194226 |
+
},
|
| 194227 |
+
{
|
| 194228 |
+
"epoch": 596.17,
|
| 194229 |
+
"learning_rate": 8.82161858974359e-06,
|
| 194230 |
+
"loss": 0.5428,
|
| 194231 |
+
"step": 74560
|
| 194232 |
+
},
|
| 194233 |
+
{
|
| 194234 |
+
"epoch": 596.21,
|
| 194235 |
+
"learning_rate": 8.821538461538462e-06,
|
| 194236 |
+
"loss": 1.2103,
|
| 194237 |
+
"step": 74565
|
| 194238 |
+
},
|
| 194239 |
+
{
|
| 194240 |
+
"epoch": 596.25,
|
| 194241 |
+
"learning_rate": 8.821458333333333e-06,
|
| 194242 |
+
"loss": 0.3342,
|
| 194243 |
+
"step": 74570
|
| 194244 |
+
},
|
| 194245 |
+
{
|
| 194246 |
+
"epoch": 596.29,
|
| 194247 |
+
"learning_rate": 8.821378205128206e-06,
|
| 194248 |
+
"loss": 0.279,
|
| 194249 |
+
"step": 74575
|
| 194250 |
+
},
|
| 194251 |
+
{
|
| 194252 |
+
"epoch": 596.33,
|
| 194253 |
+
"learning_rate": 8.821298076923078e-06,
|
| 194254 |
+
"loss": 0.3352,
|
| 194255 |
+
"step": 74580
|
| 194256 |
+
},
|
| 194257 |
+
{
|
| 194258 |
+
"epoch": 596.37,
|
| 194259 |
+
"learning_rate": 8.82121794871795e-06,
|
| 194260 |
+
"loss": 0.6326,
|
| 194261 |
+
"step": 74585
|
| 194262 |
+
},
|
| 194263 |
+
{
|
| 194264 |
+
"epoch": 596.41,
|
| 194265 |
+
"learning_rate": 8.821137820512822e-06,
|
| 194266 |
+
"loss": 1.2738,
|
| 194267 |
+
"step": 74590
|
| 194268 |
+
},
|
| 194269 |
+
{
|
| 194270 |
+
"epoch": 596.45,
|
| 194271 |
+
"learning_rate": 8.821057692307694e-06,
|
| 194272 |
+
"loss": 0.2898,
|
| 194273 |
+
"step": 74595
|
| 194274 |
+
},
|
| 194275 |
+
{
|
| 194276 |
+
"epoch": 596.49,
|
| 194277 |
+
"learning_rate": 8.820977564102565e-06,
|
| 194278 |
+
"loss": 0.3831,
|
| 194279 |
+
"step": 74600
|
| 194280 |
+
},
|
| 194281 |
+
{
|
| 194282 |
+
"epoch": 596.53,
|
| 194283 |
+
"learning_rate": 8.820897435897436e-06,
|
| 194284 |
+
"loss": 0.3918,
|
| 194285 |
+
"step": 74605
|
| 194286 |
+
},
|
| 194287 |
+
{
|
| 194288 |
+
"epoch": 596.57,
|
| 194289 |
+
"learning_rate": 8.82081730769231e-06,
|
| 194290 |
+
"loss": 0.4952,
|
| 194291 |
+
"step": 74610
|
| 194292 |
+
},
|
| 194293 |
+
{
|
| 194294 |
+
"epoch": 596.61,
|
| 194295 |
+
"learning_rate": 8.82073717948718e-06,
|
| 194296 |
+
"loss": 1.0689,
|
| 194297 |
+
"step": 74615
|
| 194298 |
+
},
|
| 194299 |
+
{
|
| 194300 |
+
"epoch": 596.65,
|
| 194301 |
+
"learning_rate": 8.820657051282052e-06,
|
| 194302 |
+
"loss": 0.3247,
|
| 194303 |
+
"step": 74620
|
| 194304 |
+
},
|
| 194305 |
+
{
|
| 194306 |
+
"epoch": 596.69,
|
| 194307 |
+
"learning_rate": 8.820576923076923e-06,
|
| 194308 |
+
"loss": 0.3352,
|
| 194309 |
+
"step": 74625
|
| 194310 |
+
},
|
| 194311 |
+
{
|
| 194312 |
+
"epoch": 596.73,
|
| 194313 |
+
"learning_rate": 8.820496794871796e-06,
|
| 194314 |
+
"loss": 0.3389,
|
| 194315 |
+
"step": 74630
|
| 194316 |
+
},
|
| 194317 |
+
{
|
| 194318 |
+
"epoch": 596.76,
|
| 194319 |
+
"learning_rate": 8.820416666666666e-06,
|
| 194320 |
+
"loss": 0.6059,
|
| 194321 |
+
"step": 74635
|
| 194322 |
+
},
|
| 194323 |
+
{
|
| 194324 |
+
"epoch": 596.8,
|
| 194325 |
+
"learning_rate": 8.82033653846154e-06,
|
| 194326 |
+
"loss": 1.2196,
|
| 194327 |
+
"step": 74640
|
| 194328 |
+
},
|
| 194329 |
+
{
|
| 194330 |
+
"epoch": 596.84,
|
| 194331 |
+
"learning_rate": 8.820256410256412e-06,
|
| 194332 |
+
"loss": 0.3291,
|
| 194333 |
+
"step": 74645
|
| 194334 |
+
},
|
| 194335 |
+
{
|
| 194336 |
+
"epoch": 596.88,
|
| 194337 |
+
"learning_rate": 8.820176282051282e-06,
|
| 194338 |
+
"loss": 0.3028,
|
| 194339 |
+
"step": 74650
|
| 194340 |
+
},
|
| 194341 |
+
{
|
| 194342 |
+
"epoch": 596.92,
|
| 194343 |
+
"learning_rate": 8.820096153846155e-06,
|
| 194344 |
+
"loss": 0.3241,
|
| 194345 |
+
"step": 74655
|
| 194346 |
+
},
|
| 194347 |
+
{
|
| 194348 |
+
"epoch": 596.96,
|
| 194349 |
+
"learning_rate": 8.820016025641026e-06,
|
| 194350 |
+
"loss": 0.5501,
|
| 194351 |
+
"step": 74660
|
| 194352 |
+
},
|
| 194353 |
+
{
|
| 194354 |
+
"epoch": 597.0,
|
| 194355 |
+
"eval_loss": 0.3461139500141144,
|
| 194356 |
+
"eval_runtime": 40.6764,
|
| 194357 |
+
"eval_samples_per_second": 20.651,
|
| 194358 |
+
"eval_steps_per_second": 0.664,
|
| 194359 |
+
"eval_wer": 0.18884180790960453,
|
| 194360 |
+
"step": 74664
|
| 194361 |
+
},
|
| 194362 |
+
{
|
| 194363 |
+
"epoch": 602.01,
|
| 194364 |
+
"learning_rate": 8.819935897435898e-06,
|
| 194365 |
+
"loss": 0.4039,
|
| 194366 |
+
"step": 74665
|
| 194367 |
+
},
|
| 194368 |
+
{
|
| 194369 |
+
"epoch": 602.05,
|
| 194370 |
+
"learning_rate": 8.819855769230769e-06,
|
| 194371 |
+
"loss": 0.3075,
|
| 194372 |
+
"step": 74670
|
| 194373 |
+
},
|
| 194374 |
+
{
|
| 194375 |
+
"epoch": 602.09,
|
| 194376 |
+
"learning_rate": 8.819775641025642e-06,
|
| 194377 |
+
"loss": 0.2638,
|
| 194378 |
+
"step": 74675
|
| 194379 |
+
},
|
| 194380 |
+
{
|
| 194381 |
+
"epoch": 602.13,
|
| 194382 |
+
"learning_rate": 8.819695512820513e-06,
|
| 194383 |
+
"loss": 0.346,
|
| 194384 |
+
"step": 74680
|
| 194385 |
+
},
|
| 194386 |
+
{
|
| 194387 |
+
"epoch": 602.17,
|
| 194388 |
+
"learning_rate": 8.819615384615385e-06,
|
| 194389 |
+
"loss": 0.6652,
|
| 194390 |
+
"step": 74685
|
| 194391 |
+
},
|
| 194392 |
+
{
|
| 194393 |
+
"epoch": 602.21,
|
| 194394 |
+
"learning_rate": 8.819535256410258e-06,
|
| 194395 |
+
"loss": 1.2922,
|
| 194396 |
+
"step": 74690
|
| 194397 |
+
},
|
| 194398 |
+
{
|
| 194399 |
+
"epoch": 602.25,
|
| 194400 |
+
"learning_rate": 8.81945512820513e-06,
|
| 194401 |
+
"loss": 0.307,
|
| 194402 |
+
"step": 74695
|
| 194403 |
+
},
|
| 194404 |
+
{
|
| 194405 |
+
"epoch": 602.29,
|
| 194406 |
+
"learning_rate": 8.819375e-06,
|
| 194407 |
+
"loss": 0.2766,
|
| 194408 |
+
"step": 74700
|
| 194409 |
+
},
|
| 194410 |
+
{
|
| 194411 |
+
"epoch": 602.33,
|
| 194412 |
+
"learning_rate": 8.819294871794872e-06,
|
| 194413 |
+
"loss": 0.3401,
|
| 194414 |
+
"step": 74705
|
| 194415 |
+
},
|
| 194416 |
+
{
|
| 194417 |
+
"epoch": 602.37,
|
| 194418 |
+
"learning_rate": 8.819214743589745e-06,
|
| 194419 |
+
"loss": 0.5566,
|
| 194420 |
+
"step": 74710
|
| 194421 |
+
},
|
| 194422 |
+
{
|
| 194423 |
+
"epoch": 602.41,
|
| 194424 |
+
"learning_rate": 8.819134615384616e-06,
|
| 194425 |
+
"loss": 1.1703,
|
| 194426 |
+
"step": 74715
|
| 194427 |
+
},
|
| 194428 |
+
{
|
| 194429 |
+
"epoch": 602.45,
|
| 194430 |
+
"learning_rate": 8.819054487179488e-06,
|
| 194431 |
+
"loss": 0.3373,
|
| 194432 |
+
"step": 74720
|
| 194433 |
+
},
|
| 194434 |
+
{
|
| 194435 |
+
"epoch": 602.49,
|
| 194436 |
+
"learning_rate": 8.818974358974359e-06,
|
| 194437 |
+
"loss": 0.3019,
|
| 194438 |
+
"step": 74725
|
| 194439 |
+
},
|
| 194440 |
+
{
|
| 194441 |
+
"epoch": 602.53,
|
| 194442 |
+
"learning_rate": 8.818894230769232e-06,
|
| 194443 |
+
"loss": 0.351,
|
| 194444 |
+
"step": 74730
|
| 194445 |
+
},
|
| 194446 |
+
{
|
| 194447 |
+
"epoch": 602.57,
|
| 194448 |
+
"learning_rate": 8.818814102564103e-06,
|
| 194449 |
+
"loss": 0.5587,
|
| 194450 |
+
"step": 74735
|
| 194451 |
+
},
|
| 194452 |
+
{
|
| 194453 |
+
"epoch": 602.61,
|
| 194454 |
+
"learning_rate": 8.818733974358975e-06,
|
| 194455 |
+
"loss": 1.1618,
|
| 194456 |
+
"step": 74740
|
| 194457 |
+
},
|
| 194458 |
+
{
|
| 194459 |
+
"epoch": 602.65,
|
| 194460 |
+
"learning_rate": 8.818653846153848e-06,
|
| 194461 |
+
"loss": 0.327,
|
| 194462 |
+
"step": 74745
|
| 194463 |
+
},
|
| 194464 |
+
{
|
| 194465 |
+
"epoch": 602.69,
|
| 194466 |
+
"learning_rate": 8.81857371794872e-06,
|
| 194467 |
+
"loss": 0.3309,
|
| 194468 |
+
"step": 74750
|
| 194469 |
+
},
|
| 194470 |
+
{
|
| 194471 |
+
"epoch": 602.73,
|
| 194472 |
+
"learning_rate": 8.81849358974359e-06,
|
| 194473 |
+
"loss": 0.4235,
|
| 194474 |
+
"step": 74755
|
| 194475 |
+
},
|
| 194476 |
+
{
|
| 194477 |
+
"epoch": 602.77,
|
| 194478 |
+
"learning_rate": 8.818413461538462e-06,
|
| 194479 |
+
"loss": 0.5577,
|
| 194480 |
+
"step": 74760
|
| 194481 |
+
},
|
| 194482 |
+
{
|
| 194483 |
+
"epoch": 602.81,
|
| 194484 |
+
"learning_rate": 8.818333333333335e-06,
|
| 194485 |
+
"loss": 1.1446,
|
| 194486 |
+
"step": 74765
|
| 194487 |
+
},
|
| 194488 |
+
{
|
| 194489 |
+
"epoch": 602.85,
|
| 194490 |
+
"learning_rate": 8.818253205128205e-06,
|
| 194491 |
+
"loss": 0.3165,
|
| 194492 |
+
"step": 74770
|
| 194493 |
+
},
|
| 194494 |
+
{
|
| 194495 |
+
"epoch": 602.9,
|
| 194496 |
+
"learning_rate": 8.818173076923078e-06,
|
| 194497 |
+
"loss": 0.2766,
|
| 194498 |
+
"step": 74775
|
| 194499 |
+
},
|
| 194500 |
+
{
|
| 194501 |
+
"epoch": 602.94,
|
| 194502 |
+
"learning_rate": 8.818092948717949e-06,
|
| 194503 |
+
"loss": 0.3203,
|
| 194504 |
+
"step": 74780
|
| 194505 |
+
},
|
| 194506 |
+
{
|
| 194507 |
+
"epoch": 602.98,
|
| 194508 |
+
"learning_rate": 8.81801282051282e-06,
|
| 194509 |
+
"loss": 0.5715,
|
| 194510 |
+
"step": 74785
|
| 194511 |
+
},
|
| 194512 |
+
{
|
| 194513 |
+
"epoch": 603.0,
|
| 194514 |
+
"eval_loss": 0.38998129963874817,
|
| 194515 |
+
"eval_runtime": 40.9192,
|
| 194516 |
+
"eval_samples_per_second": 20.528,
|
| 194517 |
+
"eval_steps_per_second": 0.66,
|
| 194518 |
+
"eval_wer": 0.1910048391688016,
|
| 194519 |
+
"step": 74788
|
| 194520 |
+
},
|
| 194521 |
+
{
|
| 194522 |
+
"epoch": 603.02,
|
| 194523 |
+
"learning_rate": 8.817932692307694e-06,
|
| 194524 |
+
"loss": 0.3641,
|
| 194525 |
+
"step": 74790
|
| 194526 |
+
},
|
| 194527 |
+
{
|
| 194528 |
+
"epoch": 603.06,
|
| 194529 |
+
"learning_rate": 8.817852564102565e-06,
|
| 194530 |
+
"loss": 0.3938,
|
| 194531 |
+
"step": 74795
|
| 194532 |
+
},
|
| 194533 |
+
{
|
| 194534 |
+
"epoch": 603.1,
|
| 194535 |
+
"learning_rate": 8.817772435897436e-06,
|
| 194536 |
+
"loss": 0.3374,
|
| 194537 |
+
"step": 74800
|
| 194538 |
+
},
|
| 194539 |
+
{
|
| 194540 |
+
"epoch": 603.14,
|
| 194541 |
+
"learning_rate": 8.817692307692308e-06,
|
| 194542 |
+
"loss": 0.36,
|
| 194543 |
+
"step": 74805
|
| 194544 |
+
},
|
| 194545 |
+
{
|
| 194546 |
+
"epoch": 603.18,
|
| 194547 |
+
"learning_rate": 8.81761217948718e-06,
|
| 194548 |
+
"loss": 0.7182,
|
| 194549 |
+
"step": 74810
|
| 194550 |
+
},
|
| 194551 |
+
{
|
| 194552 |
+
"epoch": 603.22,
|
| 194553 |
+
"learning_rate": 8.817532051282052e-06,
|
| 194554 |
+
"loss": 1.0542,
|
| 194555 |
+
"step": 74815
|
| 194556 |
+
},
|
| 194557 |
+
{
|
| 194558 |
+
"epoch": 603.26,
|
| 194559 |
+
"learning_rate": 8.817451923076923e-06,
|
| 194560 |
+
"loss": 0.3317,
|
| 194561 |
+
"step": 74820
|
| 194562 |
+
},
|
| 194563 |
+
{
|
| 194564 |
+
"epoch": 603.3,
|
| 194565 |
+
"learning_rate": 8.817371794871795e-06,
|
| 194566 |
+
"loss": 0.4227,
|
| 194567 |
+
"step": 74825
|
| 194568 |
+
},
|
| 194569 |
+
{
|
| 194570 |
+
"epoch": 603.34,
|
| 194571 |
+
"learning_rate": 8.817291666666668e-06,
|
| 194572 |
+
"loss": 0.4196,
|
| 194573 |
+
"step": 74830
|
| 194574 |
+
},
|
| 194575 |
+
{
|
| 194576 |
+
"epoch": 603.38,
|
| 194577 |
+
"learning_rate": 8.817211538461539e-06,
|
| 194578 |
+
"loss": 0.6258,
|
| 194579 |
+
"step": 74835
|
| 194580 |
+
},
|
| 194581 |
+
{
|
| 194582 |
+
"epoch": 603.42,
|
| 194583 |
+
"learning_rate": 8.81713141025641e-06,
|
| 194584 |
+
"loss": 1.0404,
|
| 194585 |
+
"step": 74840
|
| 194586 |
+
},
|
| 194587 |
+
{
|
| 194588 |
+
"epoch": 603.46,
|
| 194589 |
+
"learning_rate": 8.817051282051284e-06,
|
| 194590 |
+
"loss": 0.3388,
|
| 194591 |
+
"step": 74845
|
| 194592 |
+
},
|
| 194593 |
+
{
|
| 194594 |
+
"epoch": 603.5,
|
| 194595 |
+
"learning_rate": 8.816971153846155e-06,
|
| 194596 |
+
"loss": 0.4682,
|
| 194597 |
+
"step": 74850
|
| 194598 |
+
},
|
| 194599 |
+
{
|
| 194600 |
+
"epoch": 603.54,
|
| 194601 |
+
"learning_rate": 8.816891025641026e-06,
|
| 194602 |
+
"loss": 0.3918,
|
| 194603 |
+
"step": 74855
|
| 194604 |
+
},
|
| 194605 |
+
{
|
| 194606 |
+
"epoch": 603.58,
|
| 194607 |
+
"learning_rate": 8.816810897435898e-06,
|
| 194608 |
+
"loss": 0.6384,
|
| 194609 |
+
"step": 74860
|
| 194610 |
+
},
|
| 194611 |
+
{
|
| 194612 |
+
"epoch": 603.62,
|
| 194613 |
+
"learning_rate": 8.81673076923077e-06,
|
| 194614 |
+
"loss": 1.0819,
|
| 194615 |
+
"step": 74865
|
| 194616 |
+
},
|
| 194617 |
+
{
|
| 194618 |
+
"epoch": 603.66,
|
| 194619 |
+
"learning_rate": 8.816650641025642e-06,
|
| 194620 |
+
"loss": 0.2616,
|
| 194621 |
+
"step": 74870
|
| 194622 |
+
},
|
| 194623 |
+
{
|
| 194624 |
+
"epoch": 603.7,
|
| 194625 |
+
"learning_rate": 8.816570512820513e-06,
|
| 194626 |
+
"loss": 0.3247,
|
| 194627 |
+
"step": 74875
|
| 194628 |
+
},
|
| 194629 |
+
{
|
| 194630 |
+
"epoch": 603.74,
|
| 194631 |
+
"learning_rate": 8.816490384615385e-06,
|
| 194632 |
+
"loss": 0.318,
|
| 194633 |
+
"step": 74880
|
| 194634 |
+
},
|
| 194635 |
+
{
|
| 194636 |
+
"epoch": 603.78,
|
| 194637 |
+
"learning_rate": 8.816410256410258e-06,
|
| 194638 |
+
"loss": 0.6705,
|
| 194639 |
+
"step": 74885
|
| 194640 |
+
},
|
| 194641 |
+
{
|
| 194642 |
+
"epoch": 603.82,
|
| 194643 |
+
"learning_rate": 8.816330128205129e-06,
|
| 194644 |
+
"loss": 1.2115,
|
| 194645 |
+
"step": 74890
|
| 194646 |
+
},
|
| 194647 |
+
{
|
| 194648 |
+
"epoch": 603.86,
|
| 194649 |
+
"learning_rate": 8.81625e-06,
|
| 194650 |
+
"loss": 0.3773,
|
| 194651 |
+
"step": 74895
|
| 194652 |
+
},
|
| 194653 |
+
{
|
| 194654 |
+
"epoch": 603.9,
|
| 194655 |
+
"learning_rate": 8.816169871794874e-06,
|
| 194656 |
+
"loss": 0.3811,
|
| 194657 |
+
"step": 74900
|
| 194658 |
+
},
|
| 194659 |
+
{
|
| 194660 |
+
"epoch": 603.94,
|
| 194661 |
+
"learning_rate": 8.816089743589745e-06,
|
| 194662 |
+
"loss": 0.388,
|
| 194663 |
+
"step": 74905
|
| 194664 |
+
},
|
| 194665 |
+
{
|
| 194666 |
+
"epoch": 603.98,
|
| 194667 |
+
"learning_rate": 8.816009615384616e-06,
|
| 194668 |
+
"loss": 0.8204,
|
| 194669 |
+
"step": 74910
|
| 194670 |
+
},
|
| 194671 |
+
{
|
| 194672 |
+
"epoch": 604.0,
|
| 194673 |
+
"eval_loss": 0.4254598617553711,
|
| 194674 |
+
"eval_runtime": 39.7945,
|
| 194675 |
+
"eval_samples_per_second": 21.108,
|
| 194676 |
+
"eval_steps_per_second": 0.678,
|
| 194677 |
+
"eval_wer": 0.19010397463313916,
|
| 194678 |
+
"step": 74912
|
| 194679 |
+
},
|
| 194680 |
+
{
|
| 194681 |
+
"epoch": 599.02,
|
| 194682 |
+
"learning_rate": 8.815929487179488e-06,
|
| 194683 |
+
"loss": 0.3721,
|
| 194684 |
+
"step": 74915
|
| 194685 |
+
},
|
| 194686 |
+
{
|
| 194687 |
+
"epoch": 599.06,
|
| 194688 |
+
"learning_rate": 8.81584935897436e-06,
|
| 194689 |
+
"loss": 0.2658,
|
| 194690 |
+
"step": 74920
|
| 194691 |
+
},
|
| 194692 |
+
{
|
| 194693 |
+
"epoch": 599.1,
|
| 194694 |
+
"learning_rate": 8.81576923076923e-06,
|
| 194695 |
+
"loss": 0.3926,
|
| 194696 |
+
"step": 74925
|
| 194697 |
+
},
|
| 194698 |
+
{
|
| 194699 |
+
"epoch": 599.14,
|
| 194700 |
+
"learning_rate": 8.815689102564103e-06,
|
| 194701 |
+
"loss": 0.3716,
|
| 194702 |
+
"step": 74930
|
| 194703 |
+
},
|
| 194704 |
+
{
|
| 194705 |
+
"epoch": 599.18,
|
| 194706 |
+
"learning_rate": 8.815608974358975e-06,
|
| 194707 |
+
"loss": 0.7772,
|
| 194708 |
+
"step": 74935
|
| 194709 |
+
},
|
| 194710 |
+
{
|
| 194711 |
+
"epoch": 599.22,
|
| 194712 |
+
"learning_rate": 8.815528846153846e-06,
|
| 194713 |
+
"loss": 0.8761,
|
| 194714 |
+
"step": 74940
|
| 194715 |
+
},
|
| 194716 |
+
{
|
| 194717 |
+
"epoch": 599.26,
|
| 194718 |
+
"learning_rate": 8.81544871794872e-06,
|
| 194719 |
+
"loss": 0.3565,
|
| 194720 |
+
"step": 74945
|
| 194721 |
+
},
|
| 194722 |
+
{
|
| 194723 |
+
"epoch": 599.3,
|
| 194724 |
+
"learning_rate": 8.81536858974359e-06,
|
| 194725 |
+
"loss": 0.363,
|
| 194726 |
+
"step": 74950
|
| 194727 |
+
},
|
| 194728 |
+
{
|
| 194729 |
+
"epoch": 599.34,
|
| 194730 |
+
"learning_rate": 8.815288461538462e-06,
|
| 194731 |
+
"loss": 0.361,
|
| 194732 |
+
"step": 74955
|
| 194733 |
+
},
|
| 194734 |
+
{
|
| 194735 |
+
"epoch": 599.38,
|
| 194736 |
+
"learning_rate": 8.815208333333333e-06,
|
| 194737 |
+
"loss": 0.7271,
|
| 194738 |
+
"step": 74960
|
| 194739 |
+
},
|
| 194740 |
+
{
|
| 194741 |
+
"epoch": 599.42,
|
| 194742 |
+
"learning_rate": 8.815128205128206e-06,
|
| 194743 |
+
"loss": 0.9404,
|
| 194744 |
+
"step": 74965
|
| 194745 |
+
},
|
| 194746 |
+
{
|
| 194747 |
+
"epoch": 599.46,
|
| 194748 |
+
"learning_rate": 8.815048076923078e-06,
|
| 194749 |
+
"loss": 0.3435,
|
| 194750 |
+
"step": 74970
|
| 194751 |
+
},
|
| 194752 |
+
{
|
| 194753 |
+
"epoch": 599.5,
|
| 194754 |
+
"learning_rate": 8.814967948717949e-06,
|
| 194755 |
+
"loss": 0.3344,
|
| 194756 |
+
"step": 74975
|
| 194757 |
+
},
|
| 194758 |
+
{
|
| 194759 |
+
"epoch": 599.54,
|
| 194760 |
+
"learning_rate": 8.81488782051282e-06,
|
| 194761 |
+
"loss": 0.4215,
|
| 194762 |
+
"step": 74980
|
| 194763 |
+
},
|
| 194764 |
+
{
|
| 194765 |
+
"epoch": 599.58,
|
| 194766 |
+
"learning_rate": 8.814807692307693e-06,
|
| 194767 |
+
"loss": 0.7642,
|
| 194768 |
+
"step": 74985
|
| 194769 |
+
},
|
| 194770 |
+
{
|
| 194771 |
+
"epoch": 599.62,
|
| 194772 |
+
"learning_rate": 8.814727564102565e-06,
|
| 194773 |
+
"loss": 0.9729,
|
| 194774 |
+
"step": 74990
|
| 194775 |
+
},
|
| 194776 |
+
{
|
| 194777 |
+
"epoch": 599.66,
|
| 194778 |
+
"learning_rate": 8.814647435897436e-06,
|
| 194779 |
+
"loss": 0.3428,
|
| 194780 |
+
"step": 74995
|
| 194781 |
+
},
|
| 194782 |
+
{
|
| 194783 |
+
"epoch": 599.7,
|
| 194784 |
+
"learning_rate": 8.81456730769231e-06,
|
| 194785 |
+
"loss": 0.2913,
|
| 194786 |
+
"step": 75000
|
| 194787 |
+
},
|
| 194788 |
+
{
|
| 194789 |
+
"epoch": 599.74,
|
| 194790 |
+
"learning_rate": 8.81448717948718e-06,
|
| 194791 |
+
"loss": 0.3823,
|
| 194792 |
+
"step": 75005
|
| 194793 |
+
},
|
| 194794 |
+
{
|
| 194795 |
+
"epoch": 599.78,
|
| 194796 |
+
"learning_rate": 8.814407051282052e-06,
|
| 194797 |
+
"loss": 0.6696,
|
| 194798 |
+
"step": 75010
|
| 194799 |
+
},
|
| 194800 |
+
{
|
| 194801 |
+
"epoch": 599.82,
|
| 194802 |
+
"learning_rate": 8.814326923076923e-06,
|
| 194803 |
+
"loss": 0.9848,
|
| 194804 |
+
"step": 75015
|
| 194805 |
+
},
|
| 194806 |
+
{
|
| 194807 |
+
"epoch": 599.86,
|
| 194808 |
+
"learning_rate": 8.814246794871796e-06,
|
| 194809 |
+
"loss": 0.2948,
|
| 194810 |
+
"step": 75020
|
| 194811 |
+
},
|
| 194812 |
+
{
|
| 194813 |
+
"epoch": 599.9,
|
| 194814 |
+
"learning_rate": 8.814166666666668e-06,
|
| 194815 |
+
"loss": 0.3141,
|
| 194816 |
+
"step": 75025
|
| 194817 |
+
},
|
| 194818 |
+
{
|
| 194819 |
+
"epoch": 599.94,
|
| 194820 |
+
"learning_rate": 8.814086538461539e-06,
|
| 194821 |
+
"loss": 0.4113,
|
| 194822 |
+
"step": 75030
|
| 194823 |
+
},
|
| 194824 |
+
{
|
| 194825 |
+
"epoch": 599.98,
|
| 194826 |
+
"learning_rate": 8.81400641025641e-06,
|
| 194827 |
+
"loss": 0.7153,
|
| 194828 |
+
"step": 75035
|
| 194829 |
+
},
|
| 194830 |
+
{
|
| 194831 |
+
"epoch": 600.0,
|
| 194832 |
+
"eval_loss": 0.5015895962715149,
|
| 194833 |
+
"eval_runtime": 39.7322,
|
| 194834 |
+
"eval_samples_per_second": 21.142,
|
| 194835 |
+
"eval_steps_per_second": 0.68,
|
| 194836 |
+
"eval_wer": 0.20653854649878506,
|
| 194837 |
+
"step": 75037
|
| 194838 |
+
},
|
| 194839 |
+
{
|
| 194840 |
+
"epoch": 600.02,
|
| 194841 |
+
"learning_rate": 8.813926282051283e-06,
|
| 194842 |
+
"loss": 0.3683,
|
| 194843 |
+
"step": 75040
|
| 194844 |
+
},
|
| 194845 |
+
{
|
| 194846 |
+
"epoch": 600.06,
|
| 194847 |
+
"learning_rate": 8.813846153846155e-06,
|
| 194848 |
+
"loss": 0.27,
|
| 194849 |
+
"step": 75045
|
| 194850 |
+
},
|
| 194851 |
+
{
|
| 194852 |
+
"epoch": 600.1,
|
| 194853 |
+
"learning_rate": 8.813766025641026e-06,
|
| 194854 |
+
"loss": 0.3321,
|
| 194855 |
+
"step": 75050
|
| 194856 |
+
},
|
| 194857 |
+
{
|
| 194858 |
+
"epoch": 600.14,
|
| 194859 |
+
"learning_rate": 8.8136858974359e-06,
|
| 194860 |
+
"loss": 0.4144,
|
| 194861 |
+
"step": 75055
|
| 194862 |
+
},
|
| 194863 |
+
{
|
| 194864 |
+
"epoch": 600.18,
|
| 194865 |
+
"learning_rate": 8.813605769230769e-06,
|
| 194866 |
+
"loss": 0.9552,
|
| 194867 |
+
"step": 75060
|
| 194868 |
+
},
|
| 194869 |
+
{
|
| 194870 |
+
"epoch": 600.22,
|
| 194871 |
+
"learning_rate": 8.813525641025642e-06,
|
| 194872 |
+
"loss": 0.9213,
|
| 194873 |
+
"step": 75065
|
| 194874 |
+
},
|
| 194875 |
+
{
|
| 194876 |
+
"epoch": 600.26,
|
| 194877 |
+
"learning_rate": 8.813445512820513e-06,
|
| 194878 |
+
"loss": 0.3018,
|
| 194879 |
+
"step": 75070
|
| 194880 |
+
},
|
| 194881 |
+
{
|
| 194882 |
+
"epoch": 600.3,
|
| 194883 |
+
"learning_rate": 8.813365384615385e-06,
|
| 194884 |
+
"loss": 0.3287,
|
| 194885 |
+
"step": 75075
|
| 194886 |
+
},
|
| 194887 |
+
{
|
| 194888 |
+
"epoch": 600.34,
|
| 194889 |
+
"learning_rate": 8.813285256410256e-06,
|
| 194890 |
+
"loss": 0.4535,
|
| 194891 |
+
"step": 75080
|
| 194892 |
+
},
|
| 194893 |
+
{
|
| 194894 |
+
"epoch": 600.38,
|
| 194895 |
+
"learning_rate": 8.813205128205129e-06,
|
| 194896 |
+
"loss": 0.8355,
|
| 194897 |
+
"step": 75085
|
| 194898 |
+
},
|
| 194899 |
+
{
|
| 194900 |
+
"epoch": 600.42,
|
| 194901 |
+
"learning_rate": 8.813125e-06,
|
| 194902 |
+
"loss": 0.8693,
|
| 194903 |
+
"step": 75090
|
| 194904 |
+
},
|
| 194905 |
+
{
|
| 194906 |
+
"epoch": 600.46,
|
| 194907 |
+
"learning_rate": 8.813044871794872e-06,
|
| 194908 |
+
"loss": 0.2923,
|
| 194909 |
+
"step": 75095
|
| 194910 |
+
},
|
| 194911 |
+
{
|
| 194912 |
+
"epoch": 600.5,
|
| 194913 |
+
"learning_rate": 8.812964743589745e-06,
|
| 194914 |
+
"loss": 0.3067,
|
| 194915 |
+
"step": 75100
|
| 194916 |
+
},
|
| 194917 |
+
{
|
| 194918 |
+
"epoch": 600.54,
|
| 194919 |
+
"learning_rate": 8.812884615384616e-06,
|
| 194920 |
+
"loss": 0.3881,
|
| 194921 |
+
"step": 75105
|
| 194922 |
+
},
|
| 194923 |
+
{
|
| 194924 |
+
"epoch": 600.58,
|
| 194925 |
+
"learning_rate": 8.812804487179488e-06,
|
| 194926 |
+
"loss": 0.7686,
|
| 194927 |
+
"step": 75110
|
| 194928 |
+
},
|
| 194929 |
+
{
|
| 194930 |
+
"epoch": 600.62,
|
| 194931 |
+
"learning_rate": 8.812724358974359e-06,
|
| 194932 |
+
"loss": 0.9086,
|
| 194933 |
+
"step": 75115
|
| 194934 |
+
},
|
| 194935 |
+
{
|
| 194936 |
+
"epoch": 600.66,
|
| 194937 |
+
"learning_rate": 8.812644230769232e-06,
|
| 194938 |
+
"loss": 0.2994,
|
| 194939 |
+
"step": 75120
|
| 194940 |
+
},
|
| 194941 |
+
{
|
| 194942 |
+
"epoch": 600.7,
|
| 194943 |
+
"learning_rate": 8.812564102564103e-06,
|
| 194944 |
+
"loss": 0.3495,
|
| 194945 |
+
"step": 75125
|
| 194946 |
+
},
|
| 194947 |
+
{
|
| 194948 |
+
"epoch": 600.74,
|
| 194949 |
+
"learning_rate": 8.812483974358975e-06,
|
| 194950 |
+
"loss": 0.3605,
|
| 194951 |
+
"step": 75130
|
| 194952 |
+
},
|
| 194953 |
+
{
|
| 194954 |
+
"epoch": 600.78,
|
| 194955 |
+
"learning_rate": 8.812403846153846e-06,
|
| 194956 |
+
"loss": 0.8099,
|
| 194957 |
+
"step": 75135
|
| 194958 |
+
},
|
| 194959 |
+
{
|
| 194960 |
+
"epoch": 600.82,
|
| 194961 |
+
"learning_rate": 8.812323717948719e-06,
|
| 194962 |
+
"loss": 1.0115,
|
| 194963 |
+
"step": 75140
|
| 194964 |
+
},
|
| 194965 |
+
{
|
| 194966 |
+
"epoch": 600.86,
|
| 194967 |
+
"learning_rate": 8.81224358974359e-06,
|
| 194968 |
+
"loss": 0.3424,
|
| 194969 |
+
"step": 75145
|
| 194970 |
+
},
|
| 194971 |
+
{
|
| 194972 |
+
"epoch": 600.9,
|
| 194973 |
+
"learning_rate": 8.812163461538462e-06,
|
| 194974 |
+
"loss": 0.2935,
|
| 194975 |
+
"step": 75150
|
| 194976 |
+
},
|
| 194977 |
+
{
|
| 194978 |
+
"epoch": 600.94,
|
| 194979 |
+
"learning_rate": 8.812083333333335e-06,
|
| 194980 |
+
"loss": 0.3522,
|
| 194981 |
+
"step": 75155
|
| 194982 |
+
},
|
| 194983 |
+
{
|
| 194984 |
+
"epoch": 600.98,
|
| 194985 |
+
"learning_rate": 8.812003205128206e-06,
|
| 194986 |
+
"loss": 0.9099,
|
| 194987 |
+
"step": 75160
|
| 194988 |
+
},
|
| 194989 |
+
{
|
| 194990 |
+
"epoch": 601.0,
|
| 194991 |
+
"eval_loss": 0.41707369685173035,
|
| 194992 |
+
"eval_runtime": 40.6126,
|
| 194993 |
+
"eval_samples_per_second": 20.708,
|
| 194994 |
+
"eval_steps_per_second": 0.665,
|
| 194995 |
+
"eval_wer": 0.2013764427557531,
|
| 194996 |
+
"step": 75162
|
| 194997 |
}
|
| 194998 |
],
|
| 194999 |
+
"max_steps": 625000,
|
| 195000 |
"num_train_epochs": 5000,
|
| 195001 |
+
"total_flos": 2.115183129640869e+20,
|
| 195002 |
"trial_name": null,
|
| 195003 |
"trial_params": null
|
| 195004 |
}
|
model-bin/finetune/base/{checkpoint-74539 β checkpoint-75162}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1629879092.38512/events.out.tfevents.1629879092.7e498afd5545.905.83
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:de16129103ae23782997a2f6e16821239d7b72f259380afebb6a9c2bafa80298
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629879755.9543498/events.out.tfevents.1629879755.7e498afd5545.905.85
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:daf01e48c1f5fdaa4932cbf36a0e8f1a360a6c9d8d78d37ea17dbc37f9889fbc
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629880403.5211415/events.out.tfevents.1629880403.7e498afd5545.905.87
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:73f895454c759316114bbc447d0826926107eacd8b1c68db68b79c05418dfdb3
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629881049.788205/events.out.tfevents.1629881049.7e498afd5545.905.89
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5eb08ad02c8458e333fddd2e3bcab37c63f5cf5bd79fc60136fe6ab62c5abfad
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629881697.4184577/events.out.tfevents.1629881697.7e498afd5545.905.91
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a9e7fb9b395276852faf078e20ad5747928b80f779c086d8d8d6e472d95dc267
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1629879092.7e498afd5545.905.82
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1f4d2778fc57e6d6438bb3f914aa609d397f723631fb70490c2bf77e9ff71107
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629879755.7e498afd5545.905.84
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5e987a946247bf31d37e50ff0bec8308355b873abadef1ee8a0ea045987b7ecf
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629880403.7e498afd5545.905.86
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0fca7d1a3a06e4fe4d2adbfbe329d1bff6334e8c455a0a4ef00b8cbd407248b0
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629881049.7e498afd5545.905.88
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:36541710ccff020bfe48255466aa1183422f6b931e5ec67e8803f8e2442a82a2
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629881697.7e498afd5545.905.90
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:cee8924f68cd5f67852530719b67535c0badcd4e01529498c6b85fcc616fd300
|
| 3 |
+
size 8622
|