"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-153934 β checkpoint-154557}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-153934 β checkpoint-154557}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-153934 β checkpoint-154557}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-153934 β checkpoint-154557}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-153934 β checkpoint-154557}/rng_state.pth +2 -2
- model-bin/finetune/base/{checkpoint-153934 β checkpoint-154557}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-153934 β checkpoint-154557}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-153934 β checkpoint-154557}/trainer_state.json +799 -4
- model-bin/finetune/base/{checkpoint-153934 β checkpoint-154557}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630257516.869323/events.out.tfevents.1630257516.cc93b136ebf5.1086.239 +3 -0
- model-bin/finetune/base/log/1630257947.7075908/events.out.tfevents.1630257947.cc93b136ebf5.1086.241 +3 -0
- model-bin/finetune/base/log/1630258385.8405457/events.out.tfevents.1630258385.cc93b136ebf5.1086.243 +3 -0
- model-bin/finetune/base/log/1630258817.5863304/events.out.tfevents.1630258817.cc93b136ebf5.1086.245 +3 -0
- model-bin/finetune/base/log/1630259247.5041316/events.out.tfevents.1630259247.cc93b136ebf5.1086.247 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630257516.cc93b136ebf5.1086.238 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630257947.cc93b136ebf5.1086.240 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630258385.cc93b136ebf5.1086.242 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630258817.cc93b136ebf5.1086.244 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630259247.cc93b136ebf5.1086.246 +3 -0
model-bin/finetune/base/{checkpoint-153934 β checkpoint-154557}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-153934 β checkpoint-154557}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:74ba9b2dfdf6dee25d004cb565b958e98dcf9c4fbd59c57c5467560d1f4d0fa0
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-153934 β checkpoint-154557}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-153934 β checkpoint-154557}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bf7e26fb4cb3b91846741a8ca38d407f10f0f1e4bc6d6d1c322f960abd8cf940
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-153934 β checkpoint-154557}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:71bad233cfa0a1d396b5ebc324d1525576abd564039e669d5d2f749724b873a3
|
| 3 |
+
size 14503
|
model-bin/finetune/base/{checkpoint-153934 β checkpoint-154557}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:53047867f5dbec0185f704d1730a6309e0150cb39a35badc8cf7348be089b5bb
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-153934 β checkpoint-154557}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e6d2ae467a50189b89589e279f888de8f1fcdfcd87203d6852aac1cd1477f181
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-153934 β checkpoint-154557}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1668011442822563,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -295215,11 +295215,806 @@
|
|
| 295215 |
"eval_steps_per_second": 0.642,
|
| 295216 |
"eval_wer": 0.17948529411764705,
|
| 295217 |
"step": 153934
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 295218 |
}
|
| 295219 |
],
|
| 295220 |
-
"max_steps":
|
| 295221 |
"num_train_epochs": 5000,
|
| 295222 |
-
"total_flos": 4.
|
| 295223 |
"trial_name": null,
|
| 295224 |
"trial_params": null
|
| 295225 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1668011442822563,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
| 4 |
+
"epoch": 1246.0,
|
| 5 |
+
"global_step": 154557,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 295215 |
"eval_steps_per_second": 0.642,
|
| 295216 |
"eval_wer": 0.17948529411764705,
|
| 295217 |
"step": 153934
|
| 295218 |
+
},
|
| 295219 |
+
{
|
| 295220 |
+
"epoch": 1231.01,
|
| 295221 |
+
"learning_rate": 7.530306946688208e-06,
|
| 295222 |
+
"loss": 0.3618,
|
| 295223 |
+
"step": 153935
|
| 295224 |
+
},
|
| 295225 |
+
{
|
| 295226 |
+
"epoch": 1231.05,
|
| 295227 |
+
"learning_rate": 7.530226171243943e-06,
|
| 295228 |
+
"loss": 0.2959,
|
| 295229 |
+
"step": 153940
|
| 295230 |
+
},
|
| 295231 |
+
{
|
| 295232 |
+
"epoch": 1231.09,
|
| 295233 |
+
"learning_rate": 7.530145395799678e-06,
|
| 295234 |
+
"loss": 0.3115,
|
| 295235 |
+
"step": 153945
|
| 295236 |
+
},
|
| 295237 |
+
{
|
| 295238 |
+
"epoch": 1231.13,
|
| 295239 |
+
"learning_rate": 7.530064620355413e-06,
|
| 295240 |
+
"loss": 0.3225,
|
| 295241 |
+
"step": 153950
|
| 295242 |
+
},
|
| 295243 |
+
{
|
| 295244 |
+
"epoch": 1231.17,
|
| 295245 |
+
"learning_rate": 7.529983844911148e-06,
|
| 295246 |
+
"loss": 0.5211,
|
| 295247 |
+
"step": 153955
|
| 295248 |
+
},
|
| 295249 |
+
{
|
| 295250 |
+
"epoch": 1231.21,
|
| 295251 |
+
"learning_rate": 7.5299030694668826e-06,
|
| 295252 |
+
"loss": 1.0981,
|
| 295253 |
+
"step": 153960
|
| 295254 |
+
},
|
| 295255 |
+
{
|
| 295256 |
+
"epoch": 1231.25,
|
| 295257 |
+
"learning_rate": 7.5298222940226176e-06,
|
| 295258 |
+
"loss": 0.2782,
|
| 295259 |
+
"step": 153965
|
| 295260 |
+
},
|
| 295261 |
+
{
|
| 295262 |
+
"epoch": 1231.29,
|
| 295263 |
+
"learning_rate": 7.5297415185783525e-06,
|
| 295264 |
+
"loss": 0.289,
|
| 295265 |
+
"step": 153970
|
| 295266 |
+
},
|
| 295267 |
+
{
|
| 295268 |
+
"epoch": 1231.33,
|
| 295269 |
+
"learning_rate": 7.5296607431340875e-06,
|
| 295270 |
+
"loss": 0.3406,
|
| 295271 |
+
"step": 153975
|
| 295272 |
+
},
|
| 295273 |
+
{
|
| 295274 |
+
"epoch": 1231.37,
|
| 295275 |
+
"learning_rate": 7.5295799676898225e-06,
|
| 295276 |
+
"loss": 0.4844,
|
| 295277 |
+
"step": 153980
|
| 295278 |
+
},
|
| 295279 |
+
{
|
| 295280 |
+
"epoch": 1231.41,
|
| 295281 |
+
"learning_rate": 7.529499192245558e-06,
|
| 295282 |
+
"loss": 1.0779,
|
| 295283 |
+
"step": 153985
|
| 295284 |
+
},
|
| 295285 |
+
{
|
| 295286 |
+
"epoch": 1231.45,
|
| 295287 |
+
"learning_rate": 7.529418416801293e-06,
|
| 295288 |
+
"loss": 0.3267,
|
| 295289 |
+
"step": 153990
|
| 295290 |
+
},
|
| 295291 |
+
{
|
| 295292 |
+
"epoch": 1231.49,
|
| 295293 |
+
"learning_rate": 7.529337641357028e-06,
|
| 295294 |
+
"loss": 0.2785,
|
| 295295 |
+
"step": 153995
|
| 295296 |
+
},
|
| 295297 |
+
{
|
| 295298 |
+
"epoch": 1231.53,
|
| 295299 |
+
"learning_rate": 7.529256865912763e-06,
|
| 295300 |
+
"loss": 0.2886,
|
| 295301 |
+
"step": 154000
|
| 295302 |
+
},
|
| 295303 |
+
{
|
| 295304 |
+
"epoch": 1231.57,
|
| 295305 |
+
"learning_rate": 7.529176090468498e-06,
|
| 295306 |
+
"loss": 0.4818,
|
| 295307 |
+
"step": 154005
|
| 295308 |
+
},
|
| 295309 |
+
{
|
| 295310 |
+
"epoch": 1231.61,
|
| 295311 |
+
"learning_rate": 7.529095315024233e-06,
|
| 295312 |
+
"loss": 1.0733,
|
| 295313 |
+
"step": 154010
|
| 295314 |
+
},
|
| 295315 |
+
{
|
| 295316 |
+
"epoch": 1231.65,
|
| 295317 |
+
"learning_rate": 7.529014539579968e-06,
|
| 295318 |
+
"loss": 0.2831,
|
| 295319 |
+
"step": 154015
|
| 295320 |
+
},
|
| 295321 |
+
{
|
| 295322 |
+
"epoch": 1231.69,
|
| 295323 |
+
"learning_rate": 7.528933764135703e-06,
|
| 295324 |
+
"loss": 0.3198,
|
| 295325 |
+
"step": 154020
|
| 295326 |
+
},
|
| 295327 |
+
{
|
| 295328 |
+
"epoch": 1231.73,
|
| 295329 |
+
"learning_rate": 7.528852988691438e-06,
|
| 295330 |
+
"loss": 0.3014,
|
| 295331 |
+
"step": 154025
|
| 295332 |
+
},
|
| 295333 |
+
{
|
| 295334 |
+
"epoch": 1231.77,
|
| 295335 |
+
"learning_rate": 7.528772213247173e-06,
|
| 295336 |
+
"loss": 0.5665,
|
| 295337 |
+
"step": 154030
|
| 295338 |
+
},
|
| 295339 |
+
{
|
| 295340 |
+
"epoch": 1231.81,
|
| 295341 |
+
"learning_rate": 7.528691437802908e-06,
|
| 295342 |
+
"loss": 1.0155,
|
| 295343 |
+
"step": 154035
|
| 295344 |
+
},
|
| 295345 |
+
{
|
| 295346 |
+
"epoch": 1231.85,
|
| 295347 |
+
"learning_rate": 7.528610662358644e-06,
|
| 295348 |
+
"loss": 0.3132,
|
| 295349 |
+
"step": 154040
|
| 295350 |
+
},
|
| 295351 |
+
{
|
| 295352 |
+
"epoch": 1231.89,
|
| 295353 |
+
"learning_rate": 7.528529886914378e-06,
|
| 295354 |
+
"loss": 0.2465,
|
| 295355 |
+
"step": 154045
|
| 295356 |
+
},
|
| 295357 |
+
{
|
| 295358 |
+
"epoch": 1231.93,
|
| 295359 |
+
"learning_rate": 7.528449111470114e-06,
|
| 295360 |
+
"loss": 0.3236,
|
| 295361 |
+
"step": 154050
|
| 295362 |
+
},
|
| 295363 |
+
{
|
| 295364 |
+
"epoch": 1231.97,
|
| 295365 |
+
"learning_rate": 7.528368336025849e-06,
|
| 295366 |
+
"loss": 0.4988,
|
| 295367 |
+
"step": 154055
|
| 295368 |
+
},
|
| 295369 |
+
{
|
| 295370 |
+
"epoch": 1232.0,
|
| 295371 |
+
"eval_loss": 0.31069421768188477,
|
| 295372 |
+
"eval_runtime": 41.1007,
|
| 295373 |
+
"eval_samples_per_second": 20.559,
|
| 295374 |
+
"eval_steps_per_second": 0.657,
|
| 295375 |
+
"eval_wer": 0.17686326291079812,
|
| 295376 |
+
"step": 154059
|
| 295377 |
+
},
|
| 295378 |
+
{
|
| 295379 |
+
"epoch": 1232.01,
|
| 295380 |
+
"learning_rate": 7.528287560581584e-06,
|
| 295381 |
+
"loss": 0.4309,
|
| 295382 |
+
"step": 154060
|
| 295383 |
+
},
|
| 295384 |
+
{
|
| 295385 |
+
"epoch": 1232.05,
|
| 295386 |
+
"learning_rate": 7.528206785137319e-06,
|
| 295387 |
+
"loss": 0.2709,
|
| 295388 |
+
"step": 154065
|
| 295389 |
+
},
|
| 295390 |
+
{
|
| 295391 |
+
"epoch": 1232.09,
|
| 295392 |
+
"learning_rate": 7.528126009693054e-06,
|
| 295393 |
+
"loss": 0.3108,
|
| 295394 |
+
"step": 154070
|
| 295395 |
+
},
|
| 295396 |
+
{
|
| 295397 |
+
"epoch": 1232.13,
|
| 295398 |
+
"learning_rate": 7.528045234248789e-06,
|
| 295399 |
+
"loss": 0.3349,
|
| 295400 |
+
"step": 154075
|
| 295401 |
+
},
|
| 295402 |
+
{
|
| 295403 |
+
"epoch": 1232.17,
|
| 295404 |
+
"learning_rate": 7.527964458804524e-06,
|
| 295405 |
+
"loss": 0.3979,
|
| 295406 |
+
"step": 154080
|
| 295407 |
+
},
|
| 295408 |
+
{
|
| 295409 |
+
"epoch": 1232.21,
|
| 295410 |
+
"learning_rate": 7.527883683360259e-06,
|
| 295411 |
+
"loss": 1.162,
|
| 295412 |
+
"step": 154085
|
| 295413 |
+
},
|
| 295414 |
+
{
|
| 295415 |
+
"epoch": 1232.25,
|
| 295416 |
+
"learning_rate": 7.527802907915994e-06,
|
| 295417 |
+
"loss": 0.2879,
|
| 295418 |
+
"step": 154090
|
| 295419 |
+
},
|
| 295420 |
+
{
|
| 295421 |
+
"epoch": 1232.29,
|
| 295422 |
+
"learning_rate": 7.527722132471729e-06,
|
| 295423 |
+
"loss": 0.2757,
|
| 295424 |
+
"step": 154095
|
| 295425 |
+
},
|
| 295426 |
+
{
|
| 295427 |
+
"epoch": 1232.33,
|
| 295428 |
+
"learning_rate": 7.527641357027464e-06,
|
| 295429 |
+
"loss": 0.2955,
|
| 295430 |
+
"step": 154100
|
| 295431 |
+
},
|
| 295432 |
+
{
|
| 295433 |
+
"epoch": 1232.37,
|
| 295434 |
+
"learning_rate": 7.5275605815832e-06,
|
| 295435 |
+
"loss": 0.6014,
|
| 295436 |
+
"step": 154105
|
| 295437 |
+
},
|
| 295438 |
+
{
|
| 295439 |
+
"epoch": 1232.41,
|
| 295440 |
+
"learning_rate": 7.527479806138934e-06,
|
| 295441 |
+
"loss": 1.1433,
|
| 295442 |
+
"step": 154110
|
| 295443 |
+
},
|
| 295444 |
+
{
|
| 295445 |
+
"epoch": 1232.45,
|
| 295446 |
+
"learning_rate": 7.52739903069467e-06,
|
| 295447 |
+
"loss": 0.3034,
|
| 295448 |
+
"step": 154115
|
| 295449 |
+
},
|
| 295450 |
+
{
|
| 295451 |
+
"epoch": 1232.49,
|
| 295452 |
+
"learning_rate": 7.527318255250404e-06,
|
| 295453 |
+
"loss": 0.271,
|
| 295454 |
+
"step": 154120
|
| 295455 |
+
},
|
| 295456 |
+
{
|
| 295457 |
+
"epoch": 1232.53,
|
| 295458 |
+
"learning_rate": 7.52723747980614e-06,
|
| 295459 |
+
"loss": 0.2804,
|
| 295460 |
+
"step": 154125
|
| 295461 |
+
},
|
| 295462 |
+
{
|
| 295463 |
+
"epoch": 1232.57,
|
| 295464 |
+
"learning_rate": 7.527156704361875e-06,
|
| 295465 |
+
"loss": 0.4664,
|
| 295466 |
+
"step": 154130
|
| 295467 |
+
},
|
| 295468 |
+
{
|
| 295469 |
+
"epoch": 1232.61,
|
| 295470 |
+
"learning_rate": 7.52707592891761e-06,
|
| 295471 |
+
"loss": 1.0826,
|
| 295472 |
+
"step": 154135
|
| 295473 |
+
},
|
| 295474 |
+
{
|
| 295475 |
+
"epoch": 1232.65,
|
| 295476 |
+
"learning_rate": 7.526995153473345e-06,
|
| 295477 |
+
"loss": 0.2752,
|
| 295478 |
+
"step": 154140
|
| 295479 |
+
},
|
| 295480 |
+
{
|
| 295481 |
+
"epoch": 1232.69,
|
| 295482 |
+
"learning_rate": 7.52691437802908e-06,
|
| 295483 |
+
"loss": 0.2435,
|
| 295484 |
+
"step": 154145
|
| 295485 |
+
},
|
| 295486 |
+
{
|
| 295487 |
+
"epoch": 1232.73,
|
| 295488 |
+
"learning_rate": 7.526833602584815e-06,
|
| 295489 |
+
"loss": 0.3137,
|
| 295490 |
+
"step": 154150
|
| 295491 |
+
},
|
| 295492 |
+
{
|
| 295493 |
+
"epoch": 1232.77,
|
| 295494 |
+
"learning_rate": 7.52675282714055e-06,
|
| 295495 |
+
"loss": 0.4692,
|
| 295496 |
+
"step": 154155
|
| 295497 |
+
},
|
| 295498 |
+
{
|
| 295499 |
+
"epoch": 1232.81,
|
| 295500 |
+
"learning_rate": 7.5266720516962855e-06,
|
| 295501 |
+
"loss": 1.035,
|
| 295502 |
+
"step": 154160
|
| 295503 |
+
},
|
| 295504 |
+
{
|
| 295505 |
+
"epoch": 1232.85,
|
| 295506 |
+
"learning_rate": 7.52659127625202e-06,
|
| 295507 |
+
"loss": 0.3189,
|
| 295508 |
+
"step": 154165
|
| 295509 |
+
},
|
| 295510 |
+
{
|
| 295511 |
+
"epoch": 1232.89,
|
| 295512 |
+
"learning_rate": 7.5265105008077554e-06,
|
| 295513 |
+
"loss": 0.2449,
|
| 295514 |
+
"step": 154170
|
| 295515 |
+
},
|
| 295516 |
+
{
|
| 295517 |
+
"epoch": 1232.93,
|
| 295518 |
+
"learning_rate": 7.5264297253634896e-06,
|
| 295519 |
+
"loss": 0.3198,
|
| 295520 |
+
"step": 154175
|
| 295521 |
+
},
|
| 295522 |
+
{
|
| 295523 |
+
"epoch": 1232.97,
|
| 295524 |
+
"learning_rate": 7.526348949919225e-06,
|
| 295525 |
+
"loss": 0.5268,
|
| 295526 |
+
"step": 154180
|
| 295527 |
+
},
|
| 295528 |
+
{
|
| 295529 |
+
"epoch": 1233.0,
|
| 295530 |
+
"eval_loss": 0.5121481418609619,
|
| 295531 |
+
"eval_runtime": 41.2486,
|
| 295532 |
+
"eval_samples_per_second": 20.486,
|
| 295533 |
+
"eval_steps_per_second": 0.655,
|
| 295534 |
+
"eval_wer": 0.18070791282726342,
|
| 295535 |
+
"step": 154184
|
| 295536 |
+
},
|
| 295537 |
+
{
|
| 295538 |
+
"epoch": 1233.01,
|
| 295539 |
+
"learning_rate": 7.5262681744749595e-06,
|
| 295540 |
+
"loss": 0.4611,
|
| 295541 |
+
"step": 154185
|
| 295542 |
+
},
|
| 295543 |
+
{
|
| 295544 |
+
"epoch": 1233.05,
|
| 295545 |
+
"learning_rate": 7.526187399030695e-06,
|
| 295546 |
+
"loss": 0.247,
|
| 295547 |
+
"step": 154190
|
| 295548 |
+
},
|
| 295549 |
+
{
|
| 295550 |
+
"epoch": 1233.09,
|
| 295551 |
+
"learning_rate": 7.52610662358643e-06,
|
| 295552 |
+
"loss": 0.3827,
|
| 295553 |
+
"step": 154195
|
| 295554 |
+
},
|
| 295555 |
+
{
|
| 295556 |
+
"epoch": 1233.13,
|
| 295557 |
+
"learning_rate": 7.526025848142165e-06,
|
| 295558 |
+
"loss": 0.3205,
|
| 295559 |
+
"step": 154200
|
| 295560 |
+
},
|
| 295561 |
+
{
|
| 295562 |
+
"epoch": 1233.17,
|
| 295563 |
+
"learning_rate": 7.5259450726979e-06,
|
| 295564 |
+
"loss": 0.5641,
|
| 295565 |
+
"step": 154205
|
| 295566 |
+
},
|
| 295567 |
+
{
|
| 295568 |
+
"epoch": 1233.21,
|
| 295569 |
+
"learning_rate": 7.525864297253635e-06,
|
| 295570 |
+
"loss": 0.9883,
|
| 295571 |
+
"step": 154210
|
| 295572 |
+
},
|
| 295573 |
+
{
|
| 295574 |
+
"epoch": 1233.25,
|
| 295575 |
+
"learning_rate": 7.525783521809371e-06,
|
| 295576 |
+
"loss": 0.2971,
|
| 295577 |
+
"step": 154215
|
| 295578 |
+
},
|
| 295579 |
+
{
|
| 295580 |
+
"epoch": 1233.29,
|
| 295581 |
+
"learning_rate": 7.525702746365105e-06,
|
| 295582 |
+
"loss": 0.2764,
|
| 295583 |
+
"step": 154220
|
| 295584 |
+
},
|
| 295585 |
+
{
|
| 295586 |
+
"epoch": 1233.33,
|
| 295587 |
+
"learning_rate": 7.525621970920841e-06,
|
| 295588 |
+
"loss": 0.2942,
|
| 295589 |
+
"step": 154225
|
| 295590 |
+
},
|
| 295591 |
+
{
|
| 295592 |
+
"epoch": 1233.37,
|
| 295593 |
+
"learning_rate": 7.525541195476575e-06,
|
| 295594 |
+
"loss": 0.6157,
|
| 295595 |
+
"step": 154230
|
| 295596 |
+
},
|
| 295597 |
+
{
|
| 295598 |
+
"epoch": 1233.41,
|
| 295599 |
+
"learning_rate": 7.525460420032311e-06,
|
| 295600 |
+
"loss": 1.1009,
|
| 295601 |
+
"step": 154235
|
| 295602 |
+
},
|
| 295603 |
+
{
|
| 295604 |
+
"epoch": 1233.45,
|
| 295605 |
+
"learning_rate": 7.525379644588045e-06,
|
| 295606 |
+
"loss": 0.2929,
|
| 295607 |
+
"step": 154240
|
| 295608 |
+
},
|
| 295609 |
+
{
|
| 295610 |
+
"epoch": 1233.49,
|
| 295611 |
+
"learning_rate": 7.525298869143781e-06,
|
| 295612 |
+
"loss": 0.36,
|
| 295613 |
+
"step": 154245
|
| 295614 |
+
},
|
| 295615 |
+
{
|
| 295616 |
+
"epoch": 1233.53,
|
| 295617 |
+
"learning_rate": 7.525218093699515e-06,
|
| 295618 |
+
"loss": 0.631,
|
| 295619 |
+
"step": 154250
|
| 295620 |
+
},
|
| 295621 |
+
{
|
| 295622 |
+
"epoch": 1233.57,
|
| 295623 |
+
"learning_rate": 7.525137318255251e-06,
|
| 295624 |
+
"loss": 0.4253,
|
| 295625 |
+
"step": 154255
|
| 295626 |
+
},
|
| 295627 |
+
{
|
| 295628 |
+
"epoch": 1233.61,
|
| 295629 |
+
"learning_rate": 7.525056542810986e-06,
|
| 295630 |
+
"loss": 0.9945,
|
| 295631 |
+
"step": 154260
|
| 295632 |
+
},
|
| 295633 |
+
{
|
| 295634 |
+
"epoch": 1233.65,
|
| 295635 |
+
"learning_rate": 7.524975767366721e-06,
|
| 295636 |
+
"loss": 0.2619,
|
| 295637 |
+
"step": 154265
|
| 295638 |
+
},
|
| 295639 |
+
{
|
| 295640 |
+
"epoch": 1233.69,
|
| 295641 |
+
"learning_rate": 7.524894991922456e-06,
|
| 295642 |
+
"loss": 0.3313,
|
| 295643 |
+
"step": 154270
|
| 295644 |
+
},
|
| 295645 |
+
{
|
| 295646 |
+
"epoch": 1233.73,
|
| 295647 |
+
"learning_rate": 7.524814216478191e-06,
|
| 295648 |
+
"loss": 0.3285,
|
| 295649 |
+
"step": 154275
|
| 295650 |
+
},
|
| 295651 |
+
{
|
| 295652 |
+
"epoch": 1233.76,
|
| 295653 |
+
"learning_rate": 7.524733441033927e-06,
|
| 295654 |
+
"loss": 0.544,
|
| 295655 |
+
"step": 154280
|
| 295656 |
+
},
|
| 295657 |
+
{
|
| 295658 |
+
"epoch": 1233.8,
|
| 295659 |
+
"learning_rate": 7.524652665589661e-06,
|
| 295660 |
+
"loss": 0.8979,
|
| 295661 |
+
"step": 154285
|
| 295662 |
+
},
|
| 295663 |
+
{
|
| 295664 |
+
"epoch": 1233.84,
|
| 295665 |
+
"learning_rate": 7.524571890145397e-06,
|
| 295666 |
+
"loss": 0.2899,
|
| 295667 |
+
"step": 154290
|
| 295668 |
+
},
|
| 295669 |
+
{
|
| 295670 |
+
"epoch": 1233.88,
|
| 295671 |
+
"learning_rate": 7.524491114701131e-06,
|
| 295672 |
+
"loss": 0.3065,
|
| 295673 |
+
"step": 154295
|
| 295674 |
+
},
|
| 295675 |
+
{
|
| 295676 |
+
"epoch": 1233.92,
|
| 295677 |
+
"learning_rate": 7.524410339256867e-06,
|
| 295678 |
+
"loss": 0.3261,
|
| 295679 |
+
"step": 154300
|
| 295680 |
+
},
|
| 295681 |
+
{
|
| 295682 |
+
"epoch": 1233.96,
|
| 295683 |
+
"learning_rate": 7.524329563812601e-06,
|
| 295684 |
+
"loss": 0.5531,
|
| 295685 |
+
"step": 154305
|
| 295686 |
+
},
|
| 295687 |
+
{
|
| 295688 |
+
"epoch": 1234.0,
|
| 295689 |
+
"eval_loss": 0.3117029666900635,
|
| 295690 |
+
"eval_runtime": 41.4162,
|
| 295691 |
+
"eval_samples_per_second": 20.403,
|
| 295692 |
+
"eval_steps_per_second": 0.652,
|
| 295693 |
+
"eval_wer": 0.17106142761643023,
|
| 295694 |
+
"step": 154309
|
| 295695 |
+
},
|
| 295696 |
+
{
|
| 295697 |
+
"epoch": 1244.01,
|
| 295698 |
+
"learning_rate": 7.524248788368337e-06,
|
| 295699 |
+
"loss": 0.6507,
|
| 295700 |
+
"step": 154310
|
| 295701 |
+
},
|
| 295702 |
+
{
|
| 295703 |
+
"epoch": 1244.05,
|
| 295704 |
+
"learning_rate": 7.524168012924071e-06,
|
| 295705 |
+
"loss": 0.2557,
|
| 295706 |
+
"step": 154315
|
| 295707 |
+
},
|
| 295708 |
+
{
|
| 295709 |
+
"epoch": 1244.09,
|
| 295710 |
+
"learning_rate": 7.524087237479807e-06,
|
| 295711 |
+
"loss": 0.2584,
|
| 295712 |
+
"step": 154320
|
| 295713 |
+
},
|
| 295714 |
+
{
|
| 295715 |
+
"epoch": 1244.13,
|
| 295716 |
+
"learning_rate": 7.524006462035541e-06,
|
| 295717 |
+
"loss": 0.3013,
|
| 295718 |
+
"step": 154325
|
| 295719 |
+
},
|
| 295720 |
+
{
|
| 295721 |
+
"epoch": 1244.17,
|
| 295722 |
+
"learning_rate": 7.523925686591277e-06,
|
| 295723 |
+
"loss": 0.4878,
|
| 295724 |
+
"step": 154330
|
| 295725 |
+
},
|
| 295726 |
+
{
|
| 295727 |
+
"epoch": 1244.21,
|
| 295728 |
+
"learning_rate": 7.5238449111470126e-06,
|
| 295729 |
+
"loss": 1.1685,
|
| 295730 |
+
"step": 154335
|
| 295731 |
+
},
|
| 295732 |
+
{
|
| 295733 |
+
"epoch": 1244.25,
|
| 295734 |
+
"learning_rate": 7.523764135702747e-06,
|
| 295735 |
+
"loss": 0.2893,
|
| 295736 |
+
"step": 154340
|
| 295737 |
+
},
|
| 295738 |
+
{
|
| 295739 |
+
"epoch": 1244.29,
|
| 295740 |
+
"learning_rate": 7.5236833602584825e-06,
|
| 295741 |
+
"loss": 0.2858,
|
| 295742 |
+
"step": 154345
|
| 295743 |
+
},
|
| 295744 |
+
{
|
| 295745 |
+
"epoch": 1244.33,
|
| 295746 |
+
"learning_rate": 7.523602584814217e-06,
|
| 295747 |
+
"loss": 0.36,
|
| 295748 |
+
"step": 154350
|
| 295749 |
+
},
|
| 295750 |
+
{
|
| 295751 |
+
"epoch": 1244.37,
|
| 295752 |
+
"learning_rate": 7.5235218093699525e-06,
|
| 295753 |
+
"loss": 0.492,
|
| 295754 |
+
"step": 154355
|
| 295755 |
+
},
|
| 295756 |
+
{
|
| 295757 |
+
"epoch": 1244.41,
|
| 295758 |
+
"learning_rate": 7.523441033925687e-06,
|
| 295759 |
+
"loss": 1.1067,
|
| 295760 |
+
"step": 154360
|
| 295761 |
+
},
|
| 295762 |
+
{
|
| 295763 |
+
"epoch": 1244.45,
|
| 295764 |
+
"learning_rate": 7.5233602584814225e-06,
|
| 295765 |
+
"loss": 0.2341,
|
| 295766 |
+
"step": 154365
|
| 295767 |
+
},
|
| 295768 |
+
{
|
| 295769 |
+
"epoch": 1244.49,
|
| 295770 |
+
"learning_rate": 7.523279483037157e-06,
|
| 295771 |
+
"loss": 0.2546,
|
| 295772 |
+
"step": 154370
|
| 295773 |
+
},
|
| 295774 |
+
{
|
| 295775 |
+
"epoch": 1244.53,
|
| 295776 |
+
"learning_rate": 7.5231987075928925e-06,
|
| 295777 |
+
"loss": 0.2621,
|
| 295778 |
+
"step": 154375
|
| 295779 |
+
},
|
| 295780 |
+
{
|
| 295781 |
+
"epoch": 1244.57,
|
| 295782 |
+
"learning_rate": 7.523117932148627e-06,
|
| 295783 |
+
"loss": 0.4435,
|
| 295784 |
+
"step": 154380
|
| 295785 |
+
},
|
| 295786 |
+
{
|
| 295787 |
+
"epoch": 1244.61,
|
| 295788 |
+
"learning_rate": 7.5230371567043624e-06,
|
| 295789 |
+
"loss": 0.966,
|
| 295790 |
+
"step": 154385
|
| 295791 |
+
},
|
| 295792 |
+
{
|
| 295793 |
+
"epoch": 1244.65,
|
| 295794 |
+
"learning_rate": 7.522956381260098e-06,
|
| 295795 |
+
"loss": 0.277,
|
| 295796 |
+
"step": 154390
|
| 295797 |
+
},
|
| 295798 |
+
{
|
| 295799 |
+
"epoch": 1244.69,
|
| 295800 |
+
"learning_rate": 7.522875605815832e-06,
|
| 295801 |
+
"loss": 0.2706,
|
| 295802 |
+
"step": 154395
|
| 295803 |
+
},
|
| 295804 |
+
{
|
| 295805 |
+
"epoch": 1244.73,
|
| 295806 |
+
"learning_rate": 7.522794830371568e-06,
|
| 295807 |
+
"loss": 0.2751,
|
| 295808 |
+
"step": 154400
|
| 295809 |
+
},
|
| 295810 |
+
{
|
| 295811 |
+
"epoch": 1244.77,
|
| 295812 |
+
"learning_rate": 7.522714054927302e-06,
|
| 295813 |
+
"loss": 0.4626,
|
| 295814 |
+
"step": 154405
|
| 295815 |
+
},
|
| 295816 |
+
{
|
| 295817 |
+
"epoch": 1244.81,
|
| 295818 |
+
"learning_rate": 7.522633279483038e-06,
|
| 295819 |
+
"loss": 1.1171,
|
| 295820 |
+
"step": 154410
|
| 295821 |
+
},
|
| 295822 |
+
{
|
| 295823 |
+
"epoch": 1244.85,
|
| 295824 |
+
"learning_rate": 7.522552504038772e-06,
|
| 295825 |
+
"loss": 0.2921,
|
| 295826 |
+
"step": 154415
|
| 295827 |
+
},
|
| 295828 |
+
{
|
| 295829 |
+
"epoch": 1244.89,
|
| 295830 |
+
"learning_rate": 7.522471728594508e-06,
|
| 295831 |
+
"loss": 0.2548,
|
| 295832 |
+
"step": 154420
|
| 295833 |
+
},
|
| 295834 |
+
{
|
| 295835 |
+
"epoch": 1244.93,
|
| 295836 |
+
"learning_rate": 7.522390953150242e-06,
|
| 295837 |
+
"loss": 0.3238,
|
| 295838 |
+
"step": 154425
|
| 295839 |
+
},
|
| 295840 |
+
{
|
| 295841 |
+
"epoch": 1244.97,
|
| 295842 |
+
"learning_rate": 7.522310177705978e-06,
|
| 295843 |
+
"loss": 0.5491,
|
| 295844 |
+
"step": 154430
|
| 295845 |
+
},
|
| 295846 |
+
{
|
| 295847 |
+
"epoch": 1245.0,
|
| 295848 |
+
"eval_loss": 0.3593205511569977,
|
| 295849 |
+
"eval_runtime": 41.3281,
|
| 295850 |
+
"eval_samples_per_second": 20.446,
|
| 295851 |
+
"eval_steps_per_second": 0.653,
|
| 295852 |
+
"eval_wer": 0.17205071269966335,
|
| 295853 |
+
"step": 154433
|
| 295854 |
+
},
|
| 295855 |
+
{
|
| 295856 |
+
"epoch": 1245.02,
|
| 295857 |
+
"learning_rate": 7.522229402261712e-06,
|
| 295858 |
+
"loss": 0.4303,
|
| 295859 |
+
"step": 154435
|
| 295860 |
+
},
|
| 295861 |
+
{
|
| 295862 |
+
"epoch": 1245.06,
|
| 295863 |
+
"learning_rate": 7.522148626817448e-06,
|
| 295864 |
+
"loss": 0.291,
|
| 295865 |
+
"step": 154440
|
| 295866 |
+
},
|
| 295867 |
+
{
|
| 295868 |
+
"epoch": 1245.1,
|
| 295869 |
+
"learning_rate": 7.522067851373182e-06,
|
| 295870 |
+
"loss": 0.234,
|
| 295871 |
+
"step": 154445
|
| 295872 |
+
},
|
| 295873 |
+
{
|
| 295874 |
+
"epoch": 1245.14,
|
| 295875 |
+
"learning_rate": 7.521987075928918e-06,
|
| 295876 |
+
"loss": 0.2705,
|
| 295877 |
+
"step": 154450
|
| 295878 |
+
},
|
| 295879 |
+
{
|
| 295880 |
+
"epoch": 1245.18,
|
| 295881 |
+
"learning_rate": 7.521906300484654e-06,
|
| 295882 |
+
"loss": 0.6283,
|
| 295883 |
+
"step": 154455
|
| 295884 |
+
},
|
| 295885 |
+
{
|
| 295886 |
+
"epoch": 1245.22,
|
| 295887 |
+
"learning_rate": 7.521825525040388e-06,
|
| 295888 |
+
"loss": 1.0106,
|
| 295889 |
+
"step": 154460
|
| 295890 |
+
},
|
| 295891 |
+
{
|
| 295892 |
+
"epoch": 1245.26,
|
| 295893 |
+
"learning_rate": 7.521744749596124e-06,
|
| 295894 |
+
"loss": 0.3233,
|
| 295895 |
+
"step": 154465
|
| 295896 |
+
},
|
| 295897 |
+
{
|
| 295898 |
+
"epoch": 1245.3,
|
| 295899 |
+
"learning_rate": 7.521663974151858e-06,
|
| 295900 |
+
"loss": 0.3337,
|
| 295901 |
+
"step": 154470
|
| 295902 |
+
},
|
| 295903 |
+
{
|
| 295904 |
+
"epoch": 1245.34,
|
| 295905 |
+
"learning_rate": 7.521583198707594e-06,
|
| 295906 |
+
"loss": 0.2771,
|
| 295907 |
+
"step": 154475
|
| 295908 |
+
},
|
| 295909 |
+
{
|
| 295910 |
+
"epoch": 1245.38,
|
| 295911 |
+
"learning_rate": 7.521502423263328e-06,
|
| 295912 |
+
"loss": 0.5839,
|
| 295913 |
+
"step": 154480
|
| 295914 |
+
},
|
| 295915 |
+
{
|
| 295916 |
+
"epoch": 1245.42,
|
| 295917 |
+
"learning_rate": 7.521421647819064e-06,
|
| 295918 |
+
"loss": 1.0925,
|
| 295919 |
+
"step": 154485
|
| 295920 |
+
},
|
| 295921 |
+
{
|
| 295922 |
+
"epoch": 1245.46,
|
| 295923 |
+
"learning_rate": 7.521340872374798e-06,
|
| 295924 |
+
"loss": 0.3022,
|
| 295925 |
+
"step": 154490
|
| 295926 |
+
},
|
| 295927 |
+
{
|
| 295928 |
+
"epoch": 1245.5,
|
| 295929 |
+
"learning_rate": 7.521260096930534e-06,
|
| 295930 |
+
"loss": 0.2368,
|
| 295931 |
+
"step": 154495
|
| 295932 |
+
},
|
| 295933 |
+
{
|
| 295934 |
+
"epoch": 1245.54,
|
| 295935 |
+
"learning_rate": 7.521179321486268e-06,
|
| 295936 |
+
"loss": 0.3563,
|
| 295937 |
+
"step": 154500
|
| 295938 |
+
},
|
| 295939 |
+
{
|
| 295940 |
+
"epoch": 1245.58,
|
| 295941 |
+
"learning_rate": 7.521098546042004e-06,
|
| 295942 |
+
"loss": 0.5765,
|
| 295943 |
+
"step": 154505
|
| 295944 |
+
},
|
| 295945 |
+
{
|
| 295946 |
+
"epoch": 1245.62,
|
| 295947 |
+
"learning_rate": 7.52101777059774e-06,
|
| 295948 |
+
"loss": 0.9627,
|
| 295949 |
+
"step": 154510
|
| 295950 |
+
},
|
| 295951 |
+
{
|
| 295952 |
+
"epoch": 1245.66,
|
| 295953 |
+
"learning_rate": 7.520936995153474e-06,
|
| 295954 |
+
"loss": 0.2524,
|
| 295955 |
+
"step": 154515
|
| 295956 |
+
},
|
| 295957 |
+
{
|
| 295958 |
+
"epoch": 1245.7,
|
| 295959 |
+
"learning_rate": 7.52085621970921e-06,
|
| 295960 |
+
"loss": 0.2469,
|
| 295961 |
+
"step": 154520
|
| 295962 |
+
},
|
| 295963 |
+
{
|
| 295964 |
+
"epoch": 1245.74,
|
| 295965 |
+
"learning_rate": 7.520775444264944e-06,
|
| 295966 |
+
"loss": 0.2765,
|
| 295967 |
+
"step": 154525
|
| 295968 |
+
},
|
| 295969 |
+
{
|
| 295970 |
+
"epoch": 1245.78,
|
| 295971 |
+
"learning_rate": 7.52069466882068e-06,
|
| 295972 |
+
"loss": 0.6153,
|
| 295973 |
+
"step": 154530
|
| 295974 |
+
},
|
| 295975 |
+
{
|
| 295976 |
+
"epoch": 1245.82,
|
| 295977 |
+
"learning_rate": 7.520613893376414e-06,
|
| 295978 |
+
"loss": 1.0515,
|
| 295979 |
+
"step": 154535
|
| 295980 |
+
},
|
| 295981 |
+
{
|
| 295982 |
+
"epoch": 1245.86,
|
| 295983 |
+
"learning_rate": 7.52053311793215e-06,
|
| 295984 |
+
"loss": 0.234,
|
| 295985 |
+
"step": 154540
|
| 295986 |
+
},
|
| 295987 |
+
{
|
| 295988 |
+
"epoch": 1245.9,
|
| 295989 |
+
"learning_rate": 7.520452342487884e-06,
|
| 295990 |
+
"loss": 0.3025,
|
| 295991 |
+
"step": 154545
|
| 295992 |
+
},
|
| 295993 |
+
{
|
| 295994 |
+
"epoch": 1245.94,
|
| 295995 |
+
"learning_rate": 7.5203715670436196e-06,
|
| 295996 |
+
"loss": 0.397,
|
| 295997 |
+
"step": 154550
|
| 295998 |
+
},
|
| 295999 |
+
{
|
| 296000 |
+
"epoch": 1245.98,
|
| 296001 |
+
"learning_rate": 7.520290791599354e-06,
|
| 296002 |
+
"loss": 0.7449,
|
| 296003 |
+
"step": 154555
|
| 296004 |
+
},
|
| 296005 |
+
{
|
| 296006 |
+
"epoch": 1246.0,
|
| 296007 |
+
"eval_loss": 0.3774397671222687,
|
| 296008 |
+
"eval_runtime": 41.2529,
|
| 296009 |
+
"eval_samples_per_second": 20.483,
|
| 296010 |
+
"eval_steps_per_second": 0.654,
|
| 296011 |
+
"eval_wer": 0.17398648648648649,
|
| 296012 |
+
"step": 154557
|
| 296013 |
}
|
| 296014 |
],
|
| 296015 |
+
"max_steps": 620000,
|
| 296016 |
"num_train_epochs": 5000,
|
| 296017 |
+
"total_flos": 4.349352656879201e+20,
|
| 296018 |
"trial_name": null,
|
| 296019 |
"trial_params": null
|
| 296020 |
}
|
model-bin/finetune/base/{checkpoint-153934 β checkpoint-154557}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1630257516.869323/events.out.tfevents.1630257516.cc93b136ebf5.1086.239
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:47564b61dbc795fbff8ffb584987ed8e45290a50c8654ff6d6b721f69de0552d
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630257947.7075908/events.out.tfevents.1630257947.cc93b136ebf5.1086.241
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5eba5a6fdf49c95649c2e42f6cbd9df77ce1f295c630f4b5a7fb8288d8f10edb
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630258385.8405457/events.out.tfevents.1630258385.cc93b136ebf5.1086.243
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f9622dfc357fbd2e16d483e3d604f89b527ef37a6c7bf1532e2e88c25cf3308a
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630258817.5863304/events.out.tfevents.1630258817.cc93b136ebf5.1086.245
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4b5ad18bb6152dbce91ff419fa98b664cad125426c8e4b639dbbd56c574a2b4c
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630259247.5041316/events.out.tfevents.1630259247.cc93b136ebf5.1086.247
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:292a1010920050304adef6551b9b5d503b6ea8e379eff8bebc7bbfca9e339431
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630257516.cc93b136ebf5.1086.238
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ce8a03899b5fa87bd3d8fc1f6589ec7e1e86ce608c3218a446389363fae9ff14
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630257947.cc93b136ebf5.1086.240
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:53aab3424f08d1bb1a16984264879ec0b58616b1ae4a70ae63f52cdbd5587ecf
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630258385.cc93b136ebf5.1086.242
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b123bb5fe7fb966dedde9824ff10389dfb4aae067cbf6de8b20f125753846cc8
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630258817.cc93b136ebf5.1086.244
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6d4a8b2488c4f860f00627704e43251101cf16498afbb29e86722924531d2062
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630259247.cc93b136ebf5.1086.246
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dc711fe15994869da2da61c1357526bf656193c26bd60275c4cbfff1c554852e
|
| 3 |
+
size 8622
|