"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/trainer_state.json +793 -4
- model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630246352.8300264/events.out.tfevents.1630246352.cc93b136ebf5.1086.189 +3 -0
- model-bin/finetune/base/log/1630246785.4372547/events.out.tfevents.1630246785.cc93b136ebf5.1086.191 +3 -0
- model-bin/finetune/base/log/1630247219.004471/events.out.tfevents.1630247219.cc93b136ebf5.1086.193 +3 -0
- model-bin/finetune/base/log/1630247655.9469938/events.out.tfevents.1630247655.cc93b136ebf5.1086.195 +3 -0
- model-bin/finetune/base/log/1630248088.0553622/events.out.tfevents.1630248088.cc93b136ebf5.1086.197 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630246352.cc93b136ebf5.1086.188 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630246785.cc93b136ebf5.1086.190 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630247218.cc93b136ebf5.1086.192 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630247655.cc93b136ebf5.1086.194 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630248088.cc93b136ebf5.1086.196 +3 -0
model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1e9eede5251c3f12c63b46d66a179f1dbd2b6d9f12ab83956ee5715976561602
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a6bd337b8f0e2f3b46fc1c4ecb5f3bafbfc576e1e6bde2071ce0f670336e2879
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7a1d6790720c3684b5fab1b80fa056d808968462ad941960785ea5dec6ae7387
|
| 3 |
size 14503
|
model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:884cb1f5ec43d0bdfcfad8769f8573a767f703fd8562f4c240659e0fdd8266d6
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:be40c7eaf300a7ee9457562db278d81160ffb7d793ceebd4250c4e629248949f
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1668011442822563,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -291264,11 +291264,800 @@
|
|
| 291264 |
"eval_steps_per_second": 0.671,
|
| 291265 |
"eval_wer": 0.182829413960433,
|
| 291266 |
"step": 150827
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 291267 |
}
|
| 291268 |
],
|
| 291269 |
-
"max_steps":
|
| 291270 |
"num_train_epochs": 5000,
|
| 291271 |
-
"total_flos": 4.
|
| 291272 |
"trial_name": null,
|
| 291273 |
"trial_params": null
|
| 291274 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1668011442822563,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
| 4 |
+
"epoch": 1220.995983935743,
|
| 5 |
+
"global_step": 151448,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 291264 |
"eval_steps_per_second": 0.671,
|
| 291265 |
"eval_wer": 0.182829413960433,
|
| 291266 |
"step": 150827
|
| 291267 |
+
},
|
| 291268 |
+
{
|
| 291269 |
+
"epoch": 1216.02,
|
| 291270 |
+
"learning_rate": 7.580436187399032e-06,
|
| 291271 |
+
"loss": 0.2669,
|
| 291272 |
+
"step": 150830
|
| 291273 |
+
},
|
| 291274 |
+
{
|
| 291275 |
+
"epoch": 1216.06,
|
| 291276 |
+
"learning_rate": 7.580355411954766e-06,
|
| 291277 |
+
"loss": 0.2552,
|
| 291278 |
+
"step": 150835
|
| 291279 |
+
},
|
| 291280 |
+
{
|
| 291281 |
+
"epoch": 1216.1,
|
| 291282 |
+
"learning_rate": 7.580274636510502e-06,
|
| 291283 |
+
"loss": 0.2963,
|
| 291284 |
+
"step": 150840
|
| 291285 |
+
},
|
| 291286 |
+
{
|
| 291287 |
+
"epoch": 1216.14,
|
| 291288 |
+
"learning_rate": 7.580193861066237e-06,
|
| 291289 |
+
"loss": 0.3486,
|
| 291290 |
+
"step": 150845
|
| 291291 |
+
},
|
| 291292 |
+
{
|
| 291293 |
+
"epoch": 1216.18,
|
| 291294 |
+
"learning_rate": 7.580113085621972e-06,
|
| 291295 |
+
"loss": 0.7076,
|
| 291296 |
+
"step": 150850
|
| 291297 |
+
},
|
| 291298 |
+
{
|
| 291299 |
+
"epoch": 1216.22,
|
| 291300 |
+
"learning_rate": 7.580032310177707e-06,
|
| 291301 |
+
"loss": 0.7936,
|
| 291302 |
+
"step": 150855
|
| 291303 |
+
},
|
| 291304 |
+
{
|
| 291305 |
+
"epoch": 1216.27,
|
| 291306 |
+
"learning_rate": 7.579951534733442e-06,
|
| 291307 |
+
"loss": 0.2483,
|
| 291308 |
+
"step": 150860
|
| 291309 |
+
},
|
| 291310 |
+
{
|
| 291311 |
+
"epoch": 1216.31,
|
| 291312 |
+
"learning_rate": 7.579870759289177e-06,
|
| 291313 |
+
"loss": 0.3237,
|
| 291314 |
+
"step": 150865
|
| 291315 |
+
},
|
| 291316 |
+
{
|
| 291317 |
+
"epoch": 1216.35,
|
| 291318 |
+
"learning_rate": 7.579789983844912e-06,
|
| 291319 |
+
"loss": 0.3846,
|
| 291320 |
+
"step": 150870
|
| 291321 |
+
},
|
| 291322 |
+
{
|
| 291323 |
+
"epoch": 1216.39,
|
| 291324 |
+
"learning_rate": 7.5797092084006466e-06,
|
| 291325 |
+
"loss": 0.673,
|
| 291326 |
+
"step": 150875
|
| 291327 |
+
},
|
| 291328 |
+
{
|
| 291329 |
+
"epoch": 1216.43,
|
| 291330 |
+
"learning_rate": 7.5796284329563816e-06,
|
| 291331 |
+
"loss": 0.897,
|
| 291332 |
+
"step": 150880
|
| 291333 |
+
},
|
| 291334 |
+
{
|
| 291335 |
+
"epoch": 1216.47,
|
| 291336 |
+
"learning_rate": 7.579547657512117e-06,
|
| 291337 |
+
"loss": 0.2648,
|
| 291338 |
+
"step": 150885
|
| 291339 |
+
},
|
| 291340 |
+
{
|
| 291341 |
+
"epoch": 1216.51,
|
| 291342 |
+
"learning_rate": 7.5794668820678515e-06,
|
| 291343 |
+
"loss": 0.3112,
|
| 291344 |
+
"step": 150890
|
| 291345 |
+
},
|
| 291346 |
+
{
|
| 291347 |
+
"epoch": 1216.55,
|
| 291348 |
+
"learning_rate": 7.579386106623587e-06,
|
| 291349 |
+
"loss": 0.367,
|
| 291350 |
+
"step": 150895
|
| 291351 |
+
},
|
| 291352 |
+
{
|
| 291353 |
+
"epoch": 1216.59,
|
| 291354 |
+
"learning_rate": 7.5793053311793215e-06,
|
| 291355 |
+
"loss": 0.7596,
|
| 291356 |
+
"step": 150900
|
| 291357 |
+
},
|
| 291358 |
+
{
|
| 291359 |
+
"epoch": 1216.63,
|
| 291360 |
+
"learning_rate": 7.579224555735057e-06,
|
| 291361 |
+
"loss": 0.8846,
|
| 291362 |
+
"step": 150905
|
| 291363 |
+
},
|
| 291364 |
+
{
|
| 291365 |
+
"epoch": 1216.67,
|
| 291366 |
+
"learning_rate": 7.5791437802907915e-06,
|
| 291367 |
+
"loss": 0.2878,
|
| 291368 |
+
"step": 150910
|
| 291369 |
+
},
|
| 291370 |
+
{
|
| 291371 |
+
"epoch": 1216.71,
|
| 291372 |
+
"learning_rate": 7.579063004846527e-06,
|
| 291373 |
+
"loss": 0.2911,
|
| 291374 |
+
"step": 150915
|
| 291375 |
+
},
|
| 291376 |
+
{
|
| 291377 |
+
"epoch": 1216.75,
|
| 291378 |
+
"learning_rate": 7.578982229402262e-06,
|
| 291379 |
+
"loss": 0.4213,
|
| 291380 |
+
"step": 150920
|
| 291381 |
+
},
|
| 291382 |
+
{
|
| 291383 |
+
"epoch": 1216.79,
|
| 291384 |
+
"learning_rate": 7.578901453957997e-06,
|
| 291385 |
+
"loss": 0.6366,
|
| 291386 |
+
"step": 150925
|
| 291387 |
+
},
|
| 291388 |
+
{
|
| 291389 |
+
"epoch": 1216.83,
|
| 291390 |
+
"learning_rate": 7.578820678513732e-06,
|
| 291391 |
+
"loss": 0.886,
|
| 291392 |
+
"step": 150930
|
| 291393 |
+
},
|
| 291394 |
+
{
|
| 291395 |
+
"epoch": 1216.87,
|
| 291396 |
+
"learning_rate": 7.578739903069467e-06,
|
| 291397 |
+
"loss": 0.2832,
|
| 291398 |
+
"step": 150935
|
| 291399 |
+
},
|
| 291400 |
+
{
|
| 291401 |
+
"epoch": 1216.91,
|
| 291402 |
+
"learning_rate": 7.578659127625203e-06,
|
| 291403 |
+
"loss": 0.2781,
|
| 291404 |
+
"step": 150940
|
| 291405 |
+
},
|
| 291406 |
+
{
|
| 291407 |
+
"epoch": 1216.95,
|
| 291408 |
+
"learning_rate": 7.578578352180937e-06,
|
| 291409 |
+
"loss": 0.3466,
|
| 291410 |
+
"step": 150945
|
| 291411 |
+
},
|
| 291412 |
+
{
|
| 291413 |
+
"epoch": 1216.99,
|
| 291414 |
+
"learning_rate": 7.578497576736673e-06,
|
| 291415 |
+
"loss": 0.7638,
|
| 291416 |
+
"step": 150950
|
| 291417 |
+
},
|
| 291418 |
+
{
|
| 291419 |
+
"epoch": 1217.0,
|
| 291420 |
+
"eval_loss": 0.3567521274089813,
|
| 291421 |
+
"eval_runtime": 40.4597,
|
| 291422 |
+
"eval_samples_per_second": 20.687,
|
| 291423 |
+
"eval_steps_per_second": 0.667,
|
| 291424 |
+
"eval_wer": 0.17730812013348166,
|
| 291425 |
+
"step": 150951
|
| 291426 |
+
},
|
| 291427 |
+
{
|
| 291428 |
+
"epoch": 1207.03,
|
| 291429 |
+
"learning_rate": 7.578416801292407e-06,
|
| 291430 |
+
"loss": 0.2737,
|
| 291431 |
+
"step": 150955
|
| 291432 |
+
},
|
| 291433 |
+
{
|
| 291434 |
+
"epoch": 1207.07,
|
| 291435 |
+
"learning_rate": 7.578336025848143e-06,
|
| 291436 |
+
"loss": 0.261,
|
| 291437 |
+
"step": 150960
|
| 291438 |
+
},
|
| 291439 |
+
{
|
| 291440 |
+
"epoch": 1207.11,
|
| 291441 |
+
"learning_rate": 7.578255250403877e-06,
|
| 291442 |
+
"loss": 0.2653,
|
| 291443 |
+
"step": 150965
|
| 291444 |
+
},
|
| 291445 |
+
{
|
| 291446 |
+
"epoch": 1207.15,
|
| 291447 |
+
"learning_rate": 7.578174474959613e-06,
|
| 291448 |
+
"loss": 0.4096,
|
| 291449 |
+
"step": 150970
|
| 291450 |
+
},
|
| 291451 |
+
{
|
| 291452 |
+
"epoch": 1207.19,
|
| 291453 |
+
"learning_rate": 7.578093699515347e-06,
|
| 291454 |
+
"loss": 0.8663,
|
| 291455 |
+
"step": 150975
|
| 291456 |
+
},
|
| 291457 |
+
{
|
| 291458 |
+
"epoch": 1207.23,
|
| 291459 |
+
"learning_rate": 7.578012924071083e-06,
|
| 291460 |
+
"loss": 0.5652,
|
| 291461 |
+
"step": 150980
|
| 291462 |
+
},
|
| 291463 |
+
{
|
| 291464 |
+
"epoch": 1207.27,
|
| 291465 |
+
"learning_rate": 7.577932148626818e-06,
|
| 291466 |
+
"loss": 0.2107,
|
| 291467 |
+
"step": 150985
|
| 291468 |
+
},
|
| 291469 |
+
{
|
| 291470 |
+
"epoch": 1207.31,
|
| 291471 |
+
"learning_rate": 7.577851373182553e-06,
|
| 291472 |
+
"loss": 0.2782,
|
| 291473 |
+
"step": 150990
|
| 291474 |
+
},
|
| 291475 |
+
{
|
| 291476 |
+
"epoch": 1207.35,
|
| 291477 |
+
"learning_rate": 7.577770597738289e-06,
|
| 291478 |
+
"loss": 0.4066,
|
| 291479 |
+
"step": 150995
|
| 291480 |
+
},
|
| 291481 |
+
{
|
| 291482 |
+
"epoch": 1207.39,
|
| 291483 |
+
"learning_rate": 7.577689822294023e-06,
|
| 291484 |
+
"loss": 0.9056,
|
| 291485 |
+
"step": 151000
|
| 291486 |
+
},
|
| 291487 |
+
{
|
| 291488 |
+
"epoch": 1207.43,
|
| 291489 |
+
"learning_rate": 7.577609046849759e-06,
|
| 291490 |
+
"loss": 0.6806,
|
| 291491 |
+
"step": 151005
|
| 291492 |
+
},
|
| 291493 |
+
{
|
| 291494 |
+
"epoch": 1207.47,
|
| 291495 |
+
"learning_rate": 7.577528271405493e-06,
|
| 291496 |
+
"loss": 0.3028,
|
| 291497 |
+
"step": 151010
|
| 291498 |
+
},
|
| 291499 |
+
{
|
| 291500 |
+
"epoch": 1207.51,
|
| 291501 |
+
"learning_rate": 7.577447495961229e-06,
|
| 291502 |
+
"loss": 0.3177,
|
| 291503 |
+
"step": 151015
|
| 291504 |
+
},
|
| 291505 |
+
{
|
| 291506 |
+
"epoch": 1207.55,
|
| 291507 |
+
"learning_rate": 7.577366720516963e-06,
|
| 291508 |
+
"loss": 0.37,
|
| 291509 |
+
"step": 151020
|
| 291510 |
+
},
|
| 291511 |
+
{
|
| 291512 |
+
"epoch": 1207.59,
|
| 291513 |
+
"learning_rate": 7.577285945072699e-06,
|
| 291514 |
+
"loss": 0.8868,
|
| 291515 |
+
"step": 151025
|
| 291516 |
+
},
|
| 291517 |
+
{
|
| 291518 |
+
"epoch": 1207.63,
|
| 291519 |
+
"learning_rate": 7.577205169628433e-06,
|
| 291520 |
+
"loss": 0.588,
|
| 291521 |
+
"step": 151030
|
| 291522 |
+
},
|
| 291523 |
+
{
|
| 291524 |
+
"epoch": 1207.67,
|
| 291525 |
+
"learning_rate": 7.577124394184169e-06,
|
| 291526 |
+
"loss": 0.277,
|
| 291527 |
+
"step": 151035
|
| 291528 |
+
},
|
| 291529 |
+
{
|
| 291530 |
+
"epoch": 1207.71,
|
| 291531 |
+
"learning_rate": 7.577043618739903e-06,
|
| 291532 |
+
"loss": 0.2769,
|
| 291533 |
+
"step": 151040
|
| 291534 |
+
},
|
| 291535 |
+
{
|
| 291536 |
+
"epoch": 1207.75,
|
| 291537 |
+
"learning_rate": 7.576962843295639e-06,
|
| 291538 |
+
"loss": 0.3973,
|
| 291539 |
+
"step": 151045
|
| 291540 |
+
},
|
| 291541 |
+
{
|
| 291542 |
+
"epoch": 1207.79,
|
| 291543 |
+
"learning_rate": 7.576882067851374e-06,
|
| 291544 |
+
"loss": 0.9731,
|
| 291545 |
+
"step": 151050
|
| 291546 |
+
},
|
| 291547 |
+
{
|
| 291548 |
+
"epoch": 1207.83,
|
| 291549 |
+
"learning_rate": 7.576801292407109e-06,
|
| 291550 |
+
"loss": 0.6487,
|
| 291551 |
+
"step": 151055
|
| 291552 |
+
},
|
| 291553 |
+
{
|
| 291554 |
+
"epoch": 1207.87,
|
| 291555 |
+
"learning_rate": 7.5767205169628445e-06,
|
| 291556 |
+
"loss": 0.2917,
|
| 291557 |
+
"step": 151060
|
| 291558 |
+
},
|
| 291559 |
+
{
|
| 291560 |
+
"epoch": 1207.91,
|
| 291561 |
+
"learning_rate": 7.576639741518579e-06,
|
| 291562 |
+
"loss": 0.2741,
|
| 291563 |
+
"step": 151065
|
| 291564 |
+
},
|
| 291565 |
+
{
|
| 291566 |
+
"epoch": 1207.95,
|
| 291567 |
+
"learning_rate": 7.5765589660743145e-06,
|
| 291568 |
+
"loss": 0.3645,
|
| 291569 |
+
"step": 151070
|
| 291570 |
+
},
|
| 291571 |
+
{
|
| 291572 |
+
"epoch": 1207.99,
|
| 291573 |
+
"learning_rate": 7.576478190630049e-06,
|
| 291574 |
+
"loss": 0.825,
|
| 291575 |
+
"step": 151075
|
| 291576 |
+
},
|
| 291577 |
+
{
|
| 291578 |
+
"epoch": 1208.0,
|
| 291579 |
+
"eval_loss": 0.3887929916381836,
|
| 291580 |
+
"eval_runtime": 39.8424,
|
| 291581 |
+
"eval_samples_per_second": 21.008,
|
| 291582 |
+
"eval_steps_per_second": 0.678,
|
| 291583 |
+
"eval_wer": 0.17792285757166443,
|
| 291584 |
+
"step": 151076
|
| 291585 |
+
},
|
| 291586 |
+
{
|
| 291587 |
+
"epoch": 1218.03,
|
| 291588 |
+
"learning_rate": 7.5763974151857844e-06,
|
| 291589 |
+
"loss": 0.347,
|
| 291590 |
+
"step": 151080
|
| 291591 |
+
},
|
| 291592 |
+
{
|
| 291593 |
+
"epoch": 1218.07,
|
| 291594 |
+
"learning_rate": 7.576316639741519e-06,
|
| 291595 |
+
"loss": 0.2568,
|
| 291596 |
+
"step": 151085
|
| 291597 |
+
},
|
| 291598 |
+
{
|
| 291599 |
+
"epoch": 1218.11,
|
| 291600 |
+
"learning_rate": 7.576235864297254e-06,
|
| 291601 |
+
"loss": 0.3279,
|
| 291602 |
+
"step": 151090
|
| 291603 |
+
},
|
| 291604 |
+
{
|
| 291605 |
+
"epoch": 1218.15,
|
| 291606 |
+
"learning_rate": 7.5761550888529886e-06,
|
| 291607 |
+
"loss": 0.3822,
|
| 291608 |
+
"step": 151095
|
| 291609 |
+
},
|
| 291610 |
+
{
|
| 291611 |
+
"epoch": 1218.19,
|
| 291612 |
+
"learning_rate": 7.576074313408724e-06,
|
| 291613 |
+
"loss": 0.8134,
|
| 291614 |
+
"step": 151100
|
| 291615 |
+
},
|
| 291616 |
+
{
|
| 291617 |
+
"epoch": 1218.23,
|
| 291618 |
+
"learning_rate": 7.5759935379644585e-06,
|
| 291619 |
+
"loss": 0.6241,
|
| 291620 |
+
"step": 151105
|
| 291621 |
+
},
|
| 291622 |
+
{
|
| 291623 |
+
"epoch": 1218.27,
|
| 291624 |
+
"learning_rate": 7.575912762520194e-06,
|
| 291625 |
+
"loss": 0.2843,
|
| 291626 |
+
"step": 151110
|
| 291627 |
+
},
|
| 291628 |
+
{
|
| 291629 |
+
"epoch": 1218.31,
|
| 291630 |
+
"learning_rate": 7.57583198707593e-06,
|
| 291631 |
+
"loss": 0.2978,
|
| 291632 |
+
"step": 151115
|
| 291633 |
+
},
|
| 291634 |
+
{
|
| 291635 |
+
"epoch": 1218.35,
|
| 291636 |
+
"learning_rate": 7.575751211631664e-06,
|
| 291637 |
+
"loss": 0.3903,
|
| 291638 |
+
"step": 151120
|
| 291639 |
+
},
|
| 291640 |
+
{
|
| 291641 |
+
"epoch": 1218.39,
|
| 291642 |
+
"learning_rate": 7.5756704361874e-06,
|
| 291643 |
+
"loss": 0.8314,
|
| 291644 |
+
"step": 151125
|
| 291645 |
+
},
|
| 291646 |
+
{
|
| 291647 |
+
"epoch": 1218.43,
|
| 291648 |
+
"learning_rate": 7.575589660743134e-06,
|
| 291649 |
+
"loss": 0.6191,
|
| 291650 |
+
"step": 151130
|
| 291651 |
+
},
|
| 291652 |
+
{
|
| 291653 |
+
"epoch": 1218.47,
|
| 291654 |
+
"learning_rate": 7.57550888529887e-06,
|
| 291655 |
+
"loss": 0.2565,
|
| 291656 |
+
"step": 151135
|
| 291657 |
+
},
|
| 291658 |
+
{
|
| 291659 |
+
"epoch": 1218.51,
|
| 291660 |
+
"learning_rate": 7.575428109854604e-06,
|
| 291661 |
+
"loss": 0.2411,
|
| 291662 |
+
"step": 151140
|
| 291663 |
+
},
|
| 291664 |
+
{
|
| 291665 |
+
"epoch": 1218.55,
|
| 291666 |
+
"learning_rate": 7.57534733441034e-06,
|
| 291667 |
+
"loss": 0.3642,
|
| 291668 |
+
"step": 151145
|
| 291669 |
+
},
|
| 291670 |
+
{
|
| 291671 |
+
"epoch": 1218.59,
|
| 291672 |
+
"learning_rate": 7.575266558966074e-06,
|
| 291673 |
+
"loss": 0.9774,
|
| 291674 |
+
"step": 151150
|
| 291675 |
+
},
|
| 291676 |
+
{
|
| 291677 |
+
"epoch": 1218.63,
|
| 291678 |
+
"learning_rate": 7.57518578352181e-06,
|
| 291679 |
+
"loss": 0.6026,
|
| 291680 |
+
"step": 151155
|
| 291681 |
+
},
|
| 291682 |
+
{
|
| 291683 |
+
"epoch": 1218.67,
|
| 291684 |
+
"learning_rate": 7.575105008077544e-06,
|
| 291685 |
+
"loss": 0.254,
|
| 291686 |
+
"step": 151160
|
| 291687 |
+
},
|
| 291688 |
+
{
|
| 291689 |
+
"epoch": 1218.71,
|
| 291690 |
+
"learning_rate": 7.57502423263328e-06,
|
| 291691 |
+
"loss": 0.3303,
|
| 291692 |
+
"step": 151165
|
| 291693 |
+
},
|
| 291694 |
+
{
|
| 291695 |
+
"epoch": 1218.76,
|
| 291696 |
+
"learning_rate": 7.574943457189016e-06,
|
| 291697 |
+
"loss": 0.3565,
|
| 291698 |
+
"step": 151170
|
| 291699 |
+
},
|
| 291700 |
+
{
|
| 291701 |
+
"epoch": 1218.8,
|
| 291702 |
+
"learning_rate": 7.57486268174475e-06,
|
| 291703 |
+
"loss": 0.7702,
|
| 291704 |
+
"step": 151175
|
| 291705 |
+
},
|
| 291706 |
+
{
|
| 291707 |
+
"epoch": 1218.84,
|
| 291708 |
+
"learning_rate": 7.574781906300486e-06,
|
| 291709 |
+
"loss": 0.5369,
|
| 291710 |
+
"step": 151180
|
| 291711 |
+
},
|
| 291712 |
+
{
|
| 291713 |
+
"epoch": 1218.88,
|
| 291714 |
+
"learning_rate": 7.57470113085622e-06,
|
| 291715 |
+
"loss": 0.2883,
|
| 291716 |
+
"step": 151185
|
| 291717 |
+
},
|
| 291718 |
+
{
|
| 291719 |
+
"epoch": 1218.92,
|
| 291720 |
+
"learning_rate": 7.574620355411956e-06,
|
| 291721 |
+
"loss": 0.2753,
|
| 291722 |
+
"step": 151190
|
| 291723 |
+
},
|
| 291724 |
+
{
|
| 291725 |
+
"epoch": 1218.96,
|
| 291726 |
+
"learning_rate": 7.57453957996769e-06,
|
| 291727 |
+
"loss": 0.4346,
|
| 291728 |
+
"step": 151195
|
| 291729 |
+
},
|
| 291730 |
+
{
|
| 291731 |
+
"epoch": 1219.0,
|
| 291732 |
+
"learning_rate": 7.574458804523426e-06,
|
| 291733 |
+
"loss": 0.9678,
|
| 291734 |
+
"step": 151200
|
| 291735 |
+
},
|
| 291736 |
+
{
|
| 291737 |
+
"epoch": 1219.0,
|
| 291738 |
+
"eval_loss": 0.40293607115745544,
|
| 291739 |
+
"eval_runtime": 40.725,
|
| 291740 |
+
"eval_samples_per_second": 20.528,
|
| 291741 |
+
"eval_steps_per_second": 0.663,
|
| 291742 |
+
"eval_wer": 0.17951517798645994,
|
| 291743 |
+
"step": 151200
|
| 291744 |
+
},
|
| 291745 |
+
{
|
| 291746 |
+
"epoch": 1219.04,
|
| 291747 |
+
"learning_rate": 7.57437802907916e-06,
|
| 291748 |
+
"loss": 0.3041,
|
| 291749 |
+
"step": 151205
|
| 291750 |
+
},
|
| 291751 |
+
{
|
| 291752 |
+
"epoch": 1219.08,
|
| 291753 |
+
"learning_rate": 7.574297253634896e-06,
|
| 291754 |
+
"loss": 0.2819,
|
| 291755 |
+
"step": 151210
|
| 291756 |
+
},
|
| 291757 |
+
{
|
| 291758 |
+
"epoch": 1219.12,
|
| 291759 |
+
"learning_rate": 7.57421647819063e-06,
|
| 291760 |
+
"loss": 0.2779,
|
| 291761 |
+
"step": 151215
|
| 291762 |
+
},
|
| 291763 |
+
{
|
| 291764 |
+
"epoch": 1219.16,
|
| 291765 |
+
"learning_rate": 7.574135702746366e-06,
|
| 291766 |
+
"loss": 0.4908,
|
| 291767 |
+
"step": 151220
|
| 291768 |
+
},
|
| 291769 |
+
{
|
| 291770 |
+
"epoch": 1219.2,
|
| 291771 |
+
"learning_rate": 7.5740549273021e-06,
|
| 291772 |
+
"loss": 1.049,
|
| 291773 |
+
"step": 151225
|
| 291774 |
+
},
|
| 291775 |
+
{
|
| 291776 |
+
"epoch": 1219.24,
|
| 291777 |
+
"learning_rate": 7.573974151857836e-06,
|
| 291778 |
+
"loss": 0.3181,
|
| 291779 |
+
"step": 151230
|
| 291780 |
+
},
|
| 291781 |
+
{
|
| 291782 |
+
"epoch": 1219.28,
|
| 291783 |
+
"learning_rate": 7.573893376413572e-06,
|
| 291784 |
+
"loss": 0.2837,
|
| 291785 |
+
"step": 151235
|
| 291786 |
+
},
|
| 291787 |
+
{
|
| 291788 |
+
"epoch": 1219.32,
|
| 291789 |
+
"learning_rate": 7.573812600969306e-06,
|
| 291790 |
+
"loss": 0.3345,
|
| 291791 |
+
"step": 151240
|
| 291792 |
+
},
|
| 291793 |
+
{
|
| 291794 |
+
"epoch": 1219.36,
|
| 291795 |
+
"learning_rate": 7.5737318255250416e-06,
|
| 291796 |
+
"loss": 0.4296,
|
| 291797 |
+
"step": 151245
|
| 291798 |
+
},
|
| 291799 |
+
{
|
| 291800 |
+
"epoch": 1219.4,
|
| 291801 |
+
"learning_rate": 7.573651050080776e-06,
|
| 291802 |
+
"loss": 1.2842,
|
| 291803 |
+
"step": 151250
|
| 291804 |
+
},
|
| 291805 |
+
{
|
| 291806 |
+
"epoch": 1219.44,
|
| 291807 |
+
"learning_rate": 7.5735702746365115e-06,
|
| 291808 |
+
"loss": 0.3158,
|
| 291809 |
+
"step": 151255
|
| 291810 |
+
},
|
| 291811 |
+
{
|
| 291812 |
+
"epoch": 1219.48,
|
| 291813 |
+
"learning_rate": 7.573489499192246e-06,
|
| 291814 |
+
"loss": 0.2343,
|
| 291815 |
+
"step": 151260
|
| 291816 |
+
},
|
| 291817 |
+
{
|
| 291818 |
+
"epoch": 1219.52,
|
| 291819 |
+
"learning_rate": 7.5734087237479815e-06,
|
| 291820 |
+
"loss": 0.2876,
|
| 291821 |
+
"step": 151265
|
| 291822 |
+
},
|
| 291823 |
+
{
|
| 291824 |
+
"epoch": 1219.56,
|
| 291825 |
+
"learning_rate": 7.573327948303716e-06,
|
| 291826 |
+
"loss": 0.4508,
|
| 291827 |
+
"step": 151270
|
| 291828 |
+
},
|
| 291829 |
+
{
|
| 291830 |
+
"epoch": 1219.6,
|
| 291831 |
+
"learning_rate": 7.5732471728594515e-06,
|
| 291832 |
+
"loss": 1.1773,
|
| 291833 |
+
"step": 151275
|
| 291834 |
+
},
|
| 291835 |
+
{
|
| 291836 |
+
"epoch": 1219.64,
|
| 291837 |
+
"learning_rate": 7.573166397415186e-06,
|
| 291838 |
+
"loss": 0.27,
|
| 291839 |
+
"step": 151280
|
| 291840 |
+
},
|
| 291841 |
+
{
|
| 291842 |
+
"epoch": 1219.68,
|
| 291843 |
+
"learning_rate": 7.5730856219709215e-06,
|
| 291844 |
+
"loss": 0.2492,
|
| 291845 |
+
"step": 151285
|
| 291846 |
+
},
|
| 291847 |
+
{
|
| 291848 |
+
"epoch": 1219.72,
|
| 291849 |
+
"learning_rate": 7.573004846526657e-06,
|
| 291850 |
+
"loss": 0.3389,
|
| 291851 |
+
"step": 151290
|
| 291852 |
+
},
|
| 291853 |
+
{
|
| 291854 |
+
"epoch": 1219.76,
|
| 291855 |
+
"learning_rate": 7.5729240710823914e-06,
|
| 291856 |
+
"loss": 0.39,
|
| 291857 |
+
"step": 151295
|
| 291858 |
+
},
|
| 291859 |
+
{
|
| 291860 |
+
"epoch": 1219.8,
|
| 291861 |
+
"learning_rate": 7.572843295638127e-06,
|
| 291862 |
+
"loss": 1.2323,
|
| 291863 |
+
"step": 151300
|
| 291864 |
+
},
|
| 291865 |
+
{
|
| 291866 |
+
"epoch": 1219.84,
|
| 291867 |
+
"learning_rate": 7.572762520193861e-06,
|
| 291868 |
+
"loss": 0.2676,
|
| 291869 |
+
"step": 151305
|
| 291870 |
+
},
|
| 291871 |
+
{
|
| 291872 |
+
"epoch": 1219.88,
|
| 291873 |
+
"learning_rate": 7.572681744749597e-06,
|
| 291874 |
+
"loss": 0.2777,
|
| 291875 |
+
"step": 151310
|
| 291876 |
+
},
|
| 291877 |
+
{
|
| 291878 |
+
"epoch": 1219.92,
|
| 291879 |
+
"learning_rate": 7.572600969305331e-06,
|
| 291880 |
+
"loss": 0.3009,
|
| 291881 |
+
"step": 151315
|
| 291882 |
+
},
|
| 291883 |
+
{
|
| 291884 |
+
"epoch": 1219.96,
|
| 291885 |
+
"learning_rate": 7.572520193861067e-06,
|
| 291886 |
+
"loss": 0.4905,
|
| 291887 |
+
"step": 151320
|
| 291888 |
+
},
|
| 291889 |
+
{
|
| 291890 |
+
"epoch": 1220.0,
|
| 291891 |
+
"eval_loss": 0.40317270159721375,
|
| 291892 |
+
"eval_runtime": 40.9503,
|
| 291893 |
+
"eval_samples_per_second": 20.415,
|
| 291894 |
+
"eval_steps_per_second": 0.659,
|
| 291895 |
+
"eval_wer": 0.18318384015168088,
|
| 291896 |
+
"step": 151324
|
| 291897 |
+
},
|
| 291898 |
+
{
|
| 291899 |
+
"epoch": 1220.01,
|
| 291900 |
+
"learning_rate": 7.572439418416801e-06,
|
| 291901 |
+
"loss": 0.2942,
|
| 291902 |
+
"step": 151325
|
| 291903 |
+
},
|
| 291904 |
+
{
|
| 291905 |
+
"epoch": 1220.05,
|
| 291906 |
+
"learning_rate": 7.572358642972537e-06,
|
| 291907 |
+
"loss": 0.2593,
|
| 291908 |
+
"step": 151330
|
| 291909 |
+
},
|
| 291910 |
+
{
|
| 291911 |
+
"epoch": 1220.09,
|
| 291912 |
+
"learning_rate": 7.572277867528271e-06,
|
| 291913 |
+
"loss": 0.2962,
|
| 291914 |
+
"step": 151335
|
| 291915 |
+
},
|
| 291916 |
+
{
|
| 291917 |
+
"epoch": 1220.13,
|
| 291918 |
+
"learning_rate": 7.572197092084007e-06,
|
| 291919 |
+
"loss": 0.3793,
|
| 291920 |
+
"step": 151340
|
| 291921 |
+
},
|
| 291922 |
+
{
|
| 291923 |
+
"epoch": 1220.17,
|
| 291924 |
+
"learning_rate": 7.572116316639743e-06,
|
| 291925 |
+
"loss": 0.473,
|
| 291926 |
+
"step": 151345
|
| 291927 |
+
},
|
| 291928 |
+
{
|
| 291929 |
+
"epoch": 1220.21,
|
| 291930 |
+
"learning_rate": 7.572035541195477e-06,
|
| 291931 |
+
"loss": 1.116,
|
| 291932 |
+
"step": 151350
|
| 291933 |
+
},
|
| 291934 |
+
{
|
| 291935 |
+
"epoch": 1220.25,
|
| 291936 |
+
"learning_rate": 7.571954765751213e-06,
|
| 291937 |
+
"loss": 0.2653,
|
| 291938 |
+
"step": 151355
|
| 291939 |
+
},
|
| 291940 |
+
{
|
| 291941 |
+
"epoch": 1220.29,
|
| 291942 |
+
"learning_rate": 7.571873990306947e-06,
|
| 291943 |
+
"loss": 0.2706,
|
| 291944 |
+
"step": 151360
|
| 291945 |
+
},
|
| 291946 |
+
{
|
| 291947 |
+
"epoch": 1220.33,
|
| 291948 |
+
"learning_rate": 7.571793214862683e-06,
|
| 291949 |
+
"loss": 0.2761,
|
| 291950 |
+
"step": 151365
|
| 291951 |
+
},
|
| 291952 |
+
{
|
| 291953 |
+
"epoch": 1220.37,
|
| 291954 |
+
"learning_rate": 7.571712439418417e-06,
|
| 291955 |
+
"loss": 0.5144,
|
| 291956 |
+
"step": 151370
|
| 291957 |
+
},
|
| 291958 |
+
{
|
| 291959 |
+
"epoch": 1220.41,
|
| 291960 |
+
"learning_rate": 7.571631663974153e-06,
|
| 291961 |
+
"loss": 1.0997,
|
| 291962 |
+
"step": 151375
|
| 291963 |
+
},
|
| 291964 |
+
{
|
| 291965 |
+
"epoch": 1220.45,
|
| 291966 |
+
"learning_rate": 7.571550888529887e-06,
|
| 291967 |
+
"loss": 0.2801,
|
| 291968 |
+
"step": 151380
|
| 291969 |
+
},
|
| 291970 |
+
{
|
| 291971 |
+
"epoch": 1220.49,
|
| 291972 |
+
"learning_rate": 7.571470113085623e-06,
|
| 291973 |
+
"loss": 0.3037,
|
| 291974 |
+
"step": 151385
|
| 291975 |
+
},
|
| 291976 |
+
{
|
| 291977 |
+
"epoch": 1220.53,
|
| 291978 |
+
"learning_rate": 7.571389337641357e-06,
|
| 291979 |
+
"loss": 0.3857,
|
| 291980 |
+
"step": 151390
|
| 291981 |
+
},
|
| 291982 |
+
{
|
| 291983 |
+
"epoch": 1220.57,
|
| 291984 |
+
"learning_rate": 7.571308562197093e-06,
|
| 291985 |
+
"loss": 0.554,
|
| 291986 |
+
"step": 151395
|
| 291987 |
+
},
|
| 291988 |
+
{
|
| 291989 |
+
"epoch": 1220.61,
|
| 291990 |
+
"learning_rate": 7.571227786752827e-06,
|
| 291991 |
+
"loss": 1.0511,
|
| 291992 |
+
"step": 151400
|
| 291993 |
+
},
|
| 291994 |
+
{
|
| 291995 |
+
"epoch": 1220.65,
|
| 291996 |
+
"learning_rate": 7.571147011308563e-06,
|
| 291997 |
+
"loss": 0.2562,
|
| 291998 |
+
"step": 151405
|
| 291999 |
+
},
|
| 292000 |
+
{
|
| 292001 |
+
"epoch": 1220.69,
|
| 292002 |
+
"learning_rate": 7.571066235864299e-06,
|
| 292003 |
+
"loss": 0.2393,
|
| 292004 |
+
"step": 151410
|
| 292005 |
+
},
|
| 292006 |
+
{
|
| 292007 |
+
"epoch": 1220.73,
|
| 292008 |
+
"learning_rate": 7.570985460420033e-06,
|
| 292009 |
+
"loss": 0.3112,
|
| 292010 |
+
"step": 151415
|
| 292011 |
+
},
|
| 292012 |
+
{
|
| 292013 |
+
"epoch": 1220.77,
|
| 292014 |
+
"learning_rate": 7.570904684975769e-06,
|
| 292015 |
+
"loss": 0.5436,
|
| 292016 |
+
"step": 151420
|
| 292017 |
+
},
|
| 292018 |
+
{
|
| 292019 |
+
"epoch": 1220.81,
|
| 292020 |
+
"learning_rate": 7.570823909531503e-06,
|
| 292021 |
+
"loss": 0.9745,
|
| 292022 |
+
"step": 151425
|
| 292023 |
+
},
|
| 292024 |
+
{
|
| 292025 |
+
"epoch": 1220.85,
|
| 292026 |
+
"learning_rate": 7.570743134087239e-06,
|
| 292027 |
+
"loss": 0.2797,
|
| 292028 |
+
"step": 151430
|
| 292029 |
+
},
|
| 292030 |
+
{
|
| 292031 |
+
"epoch": 1220.89,
|
| 292032 |
+
"learning_rate": 7.570662358642973e-06,
|
| 292033 |
+
"loss": 0.2683,
|
| 292034 |
+
"step": 151435
|
| 292035 |
+
},
|
| 292036 |
+
{
|
| 292037 |
+
"epoch": 1220.93,
|
| 292038 |
+
"learning_rate": 7.570581583198709e-06,
|
| 292039 |
+
"loss": 0.3428,
|
| 292040 |
+
"step": 151440
|
| 292041 |
+
},
|
| 292042 |
+
{
|
| 292043 |
+
"epoch": 1220.97,
|
| 292044 |
+
"learning_rate": 7.570500807754443e-06,
|
| 292045 |
+
"loss": 0.4689,
|
| 292046 |
+
"step": 151445
|
| 292047 |
+
},
|
| 292048 |
+
{
|
| 292049 |
+
"epoch": 1221.0,
|
| 292050 |
+
"eval_loss": 0.3577888607978821,
|
| 292051 |
+
"eval_runtime": 42.2225,
|
| 292052 |
+
"eval_samples_per_second": 19.8,
|
| 292053 |
+
"eval_steps_per_second": 0.639,
|
| 292054 |
+
"eval_wer": 0.17625899280575538,
|
| 292055 |
+
"step": 151448
|
| 292056 |
}
|
| 292057 |
],
|
| 292058 |
+
"max_steps": 620000,
|
| 292059 |
"num_train_epochs": 5000,
|
| 292060 |
+
"total_flos": 4.2618930913164073e+20,
|
| 292061 |
"trial_name": null,
|
| 292062 |
"trial_params": null
|
| 292063 |
}
|
model-bin/finetune/base/{checkpoint-150827 β checkpoint-151448}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1630246352.8300264/events.out.tfevents.1630246352.cc93b136ebf5.1086.189
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1d2e006a76411ca66fef0f0601b536f525ad6d1bf56071220648ca0968c5ce39
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630246785.4372547/events.out.tfevents.1630246785.cc93b136ebf5.1086.191
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:88c677975dcb5799d881c0a06fd4f929de0324135c035871197a803122b7e182
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630247219.004471/events.out.tfevents.1630247219.cc93b136ebf5.1086.193
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ba94dba6d83ae89638f94a50b5ed007dd6fa5fff4a9ef3c1d84e599ce33866a2
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630247655.9469938/events.out.tfevents.1630247655.cc93b136ebf5.1086.195
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8b8bb97898f6233ed672cf04852addd9a7e44ee8bfe0ec7376dc5ec8454e6735
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630248088.0553622/events.out.tfevents.1630248088.cc93b136ebf5.1086.197
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b42f4a6bdb41edd1853e3d18e435253edac9dffde47c25b93a644f5ef02d2e20
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630246352.cc93b136ebf5.1086.188
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d09b4ca59b66bc0c0684dee2fe769e86548e33094e265be58139e3f0a4d37830
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630246785.cc93b136ebf5.1086.190
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0056a189953ac9b863ab9640ac16ddada116751dc2c5670f09d12244155c7710
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630247218.cc93b136ebf5.1086.192
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1b7cc40664c5864ee6af5d3191ef612920626590925ef149aacf562c7b185192
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630247655.cc93b136ebf5.1086.194
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b5e892ecadb74ebdc66ca0fcae41c145b395b336ec9b9b70d9c3394abfcfe60f
|
| 3 |
+
size 8462
|
model-bin/finetune/base/log/events.out.tfevents.1630248088.cc93b136ebf5.1086.196
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:222a21717f9e387720acc83eac9a3017b439135e5baedab94209e89526ba9fa0
|
| 3 |
+
size 8622
|