"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-144607 β checkpoint-145602}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-144607 β checkpoint-145602}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-144607 β checkpoint-145602}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-145228 β checkpoint-145602}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-144607 β checkpoint-145602}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-145228 β checkpoint-145602}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-144607 β checkpoint-145602}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-145228 β checkpoint-145602}/trainer_state.json +483 -6
- model-bin/finetune/base/{checkpoint-144607 β checkpoint-145602}/training_args.bin +0 -0
- model-bin/finetune/base/{checkpoint-145228 β checkpoint-145852}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-145228 β checkpoint-145852}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-145228 β checkpoint-145852}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-144607 β checkpoint-145852}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-145228 β checkpoint-145852}/rng_state.pth +2 -2
- model-bin/finetune/base/{checkpoint-144607 β checkpoint-145852}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-145228 β checkpoint-145852}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-144607 β checkpoint-145852}/trainer_state.json +1590 -6
- model-bin/finetune/base/{checkpoint-145228 β checkpoint-145852}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630226071.8984504/events.out.tfevents.1630226071.cc93b136ebf5.1086.99 +3 -0
- model-bin/finetune/base/log/1630226502.2753584/events.out.tfevents.1630226502.cc93b136ebf5.1086.101 +3 -0
- model-bin/finetune/base/log/1630226942.7046182/events.out.tfevents.1630226942.cc93b136ebf5.1086.103 +3 -0
- model-bin/finetune/base/log/1630227371.5174663/events.out.tfevents.1630227371.cc93b136ebf5.1086.105 +3 -0
- model-bin/finetune/base/log/1630227801.075265/events.out.tfevents.1630227801.cc93b136ebf5.1086.107 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630226071.cc93b136ebf5.1086.98 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630226502.cc93b136ebf5.1086.100 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630226942.cc93b136ebf5.1086.102 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630227371.cc93b136ebf5.1086.104 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630227801.cc93b136ebf5.1086.106 +3 -0
model-bin/finetune/base/{checkpoint-144607 β checkpoint-145602}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-144607 β checkpoint-145602}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:425263039b2e17958ef29df62ab9ab8878277ad6c4ead7176bda8210b2ed3fe6
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-144607 β checkpoint-145602}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-145228 β checkpoint-145602}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b2376bae8ab9408163ddcedcab213b016840a625804da3cbf1526010b8abc141
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-144607 β checkpoint-145602}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:80390ad562fb737219eaef200a69cb1cb5c9e3d1cb1722a1fdddeff6f6402b7d
|
| 3 |
size 14503
|
model-bin/finetune/base/{checkpoint-145228 β checkpoint-145602}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2d760abfe85cbbe513ba97c70e9515e7e485b75c78f8b50035843503942c2409
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-144607 β checkpoint-145602}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:cfefb4c24229027a9d0175af6026b93ceb96966446b698af9ac66780b0583d09
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-145228 β checkpoint-145602}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
-
"best_metric": 0.
|
| 3 |
-
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -284139,11 +284139,488 @@
|
|
| 284139 |
"eval_steps_per_second": 0.668,
|
| 284140 |
"eval_wer": 0.17565191913272782,
|
| 284141 |
"step": 145228
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 284142 |
}
|
| 284143 |
],
|
| 284144 |
-
"max_steps":
|
| 284145 |
"num_train_epochs": 5000,
|
| 284146 |
-
"total_flos": 4.
|
| 284147 |
"trial_name": null,
|
| 284148 |
"trial_params": null
|
| 284149 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"best_metric": 0.16716790141568252,
|
| 3 |
+
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-145602",
|
| 4 |
+
"epoch": 1164.0,
|
| 5 |
+
"global_step": 145602,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 284139 |
"eval_steps_per_second": 0.668,
|
| 284140 |
"eval_wer": 0.17565191913272782,
|
| 284141 |
"step": 145228
|
| 284142 |
+
},
|
| 284143 |
+
{
|
| 284144 |
+
"epoch": 1171.02,
|
| 284145 |
+
"learning_rate": 7.670872374798062e-06,
|
| 284146 |
+
"loss": 0.3162,
|
| 284147 |
+
"step": 145230
|
| 284148 |
+
},
|
| 284149 |
+
{
|
| 284150 |
+
"epoch": 1171.06,
|
| 284151 |
+
"learning_rate": 7.670791599353798e-06,
|
| 284152 |
+
"loss": 0.268,
|
| 284153 |
+
"step": 145235
|
| 284154 |
+
},
|
| 284155 |
+
{
|
| 284156 |
+
"epoch": 1171.1,
|
| 284157 |
+
"learning_rate": 7.670710823909532e-06,
|
| 284158 |
+
"loss": 0.3148,
|
| 284159 |
+
"step": 145240
|
| 284160 |
+
},
|
| 284161 |
+
{
|
| 284162 |
+
"epoch": 1171.14,
|
| 284163 |
+
"learning_rate": 7.670630048465268e-06,
|
| 284164 |
+
"loss": 0.3431,
|
| 284165 |
+
"step": 145245
|
| 284166 |
+
},
|
| 284167 |
+
{
|
| 284168 |
+
"epoch": 1171.18,
|
| 284169 |
+
"learning_rate": 7.670549273021002e-06,
|
| 284170 |
+
"loss": 0.5562,
|
| 284171 |
+
"step": 145250
|
| 284172 |
+
},
|
| 284173 |
+
{
|
| 284174 |
+
"epoch": 1171.22,
|
| 284175 |
+
"learning_rate": 7.670468497576738e-06,
|
| 284176 |
+
"loss": 0.8706,
|
| 284177 |
+
"step": 145255
|
| 284178 |
+
},
|
| 284179 |
+
{
|
| 284180 |
+
"epoch": 1171.26,
|
| 284181 |
+
"learning_rate": 7.670387722132472e-06,
|
| 284182 |
+
"loss": 0.283,
|
| 284183 |
+
"step": 145260
|
| 284184 |
+
},
|
| 284185 |
+
{
|
| 284186 |
+
"epoch": 1171.3,
|
| 284187 |
+
"learning_rate": 7.670306946688208e-06,
|
| 284188 |
+
"loss": 0.3143,
|
| 284189 |
+
"step": 145265
|
| 284190 |
+
},
|
| 284191 |
+
{
|
| 284192 |
+
"epoch": 1171.34,
|
| 284193 |
+
"learning_rate": 7.670226171243944e-06,
|
| 284194 |
+
"loss": 0.2854,
|
| 284195 |
+
"step": 145270
|
| 284196 |
+
},
|
| 284197 |
+
{
|
| 284198 |
+
"epoch": 1171.38,
|
| 284199 |
+
"learning_rate": 7.670145395799678e-06,
|
| 284200 |
+
"loss": 0.5725,
|
| 284201 |
+
"step": 145275
|
| 284202 |
+
},
|
| 284203 |
+
{
|
| 284204 |
+
"epoch": 1171.42,
|
| 284205 |
+
"learning_rate": 7.670064620355414e-06,
|
| 284206 |
+
"loss": 0.9977,
|
| 284207 |
+
"step": 145280
|
| 284208 |
+
},
|
| 284209 |
+
{
|
| 284210 |
+
"epoch": 1171.46,
|
| 284211 |
+
"learning_rate": 7.669983844911148e-06,
|
| 284212 |
+
"loss": 0.2628,
|
| 284213 |
+
"step": 145285
|
| 284214 |
+
},
|
| 284215 |
+
{
|
| 284216 |
+
"epoch": 1171.5,
|
| 284217 |
+
"learning_rate": 7.669903069466884e-06,
|
| 284218 |
+
"loss": 0.2448,
|
| 284219 |
+
"step": 145290
|
| 284220 |
+
},
|
| 284221 |
+
{
|
| 284222 |
+
"epoch": 1171.54,
|
| 284223 |
+
"learning_rate": 7.669822294022618e-06,
|
| 284224 |
+
"loss": 0.3178,
|
| 284225 |
+
"step": 145295
|
| 284226 |
+
},
|
| 284227 |
+
{
|
| 284228 |
+
"epoch": 1171.58,
|
| 284229 |
+
"learning_rate": 7.669741518578354e-06,
|
| 284230 |
+
"loss": 0.6503,
|
| 284231 |
+
"step": 145300
|
| 284232 |
+
},
|
| 284233 |
+
{
|
| 284234 |
+
"epoch": 1171.62,
|
| 284235 |
+
"learning_rate": 7.669660743134088e-06,
|
| 284236 |
+
"loss": 0.9847,
|
| 284237 |
+
"step": 145305
|
| 284238 |
+
},
|
| 284239 |
+
{
|
| 284240 |
+
"epoch": 1171.66,
|
| 284241 |
+
"learning_rate": 7.669579967689824e-06,
|
| 284242 |
+
"loss": 0.2887,
|
| 284243 |
+
"step": 145310
|
| 284244 |
+
},
|
| 284245 |
+
{
|
| 284246 |
+
"epoch": 1171.7,
|
| 284247 |
+
"learning_rate": 7.669499192245558e-06,
|
| 284248 |
+
"loss": 0.2481,
|
| 284249 |
+
"step": 145315
|
| 284250 |
+
},
|
| 284251 |
+
{
|
| 284252 |
+
"epoch": 1171.74,
|
| 284253 |
+
"learning_rate": 7.669418416801294e-06,
|
| 284254 |
+
"loss": 0.2989,
|
| 284255 |
+
"step": 145320
|
| 284256 |
+
},
|
| 284257 |
+
{
|
| 284258 |
+
"epoch": 1171.78,
|
| 284259 |
+
"learning_rate": 7.669337641357028e-06,
|
| 284260 |
+
"loss": 0.6053,
|
| 284261 |
+
"step": 145325
|
| 284262 |
+
},
|
| 284263 |
+
{
|
| 284264 |
+
"epoch": 1171.82,
|
| 284265 |
+
"learning_rate": 7.669256865912764e-06,
|
| 284266 |
+
"loss": 0.9806,
|
| 284267 |
+
"step": 145330
|
| 284268 |
+
},
|
| 284269 |
+
{
|
| 284270 |
+
"epoch": 1171.86,
|
| 284271 |
+
"learning_rate": 7.669176090468498e-06,
|
| 284272 |
+
"loss": 0.3248,
|
| 284273 |
+
"step": 145335
|
| 284274 |
+
},
|
| 284275 |
+
{
|
| 284276 |
+
"epoch": 1171.9,
|
| 284277 |
+
"learning_rate": 7.669095315024234e-06,
|
| 284278 |
+
"loss": 0.3173,
|
| 284279 |
+
"step": 145340
|
| 284280 |
+
},
|
| 284281 |
+
{
|
| 284282 |
+
"epoch": 1171.94,
|
| 284283 |
+
"learning_rate": 7.66901453957997e-06,
|
| 284284 |
+
"loss": 0.3376,
|
| 284285 |
+
"step": 145345
|
| 284286 |
+
},
|
| 284287 |
+
{
|
| 284288 |
+
"epoch": 1171.98,
|
| 284289 |
+
"learning_rate": 7.668933764135704e-06,
|
| 284290 |
+
"loss": 0.6947,
|
| 284291 |
+
"step": 145350
|
| 284292 |
+
},
|
| 284293 |
+
{
|
| 284294 |
+
"epoch": 1172.0,
|
| 284295 |
+
"eval_loss": 0.47769203782081604,
|
| 284296 |
+
"eval_runtime": 39.6454,
|
| 284297 |
+
"eval_samples_per_second": 21.112,
|
| 284298 |
+
"eval_steps_per_second": 0.681,
|
| 284299 |
+
"eval_wer": 0.18728573557907288,
|
| 284300 |
+
"step": 145352
|
| 284301 |
+
},
|
| 284302 |
+
{
|
| 284303 |
+
"epoch": 1162.02,
|
| 284304 |
+
"learning_rate": 7.66885298869144e-06,
|
| 284305 |
+
"loss": 0.3284,
|
| 284306 |
+
"step": 145355
|
| 284307 |
+
},
|
| 284308 |
+
{
|
| 284309 |
+
"epoch": 1162.06,
|
| 284310 |
+
"learning_rate": 7.668772213247174e-06,
|
| 284311 |
+
"loss": 0.2665,
|
| 284312 |
+
"step": 145360
|
| 284313 |
+
},
|
| 284314 |
+
{
|
| 284315 |
+
"epoch": 1162.1,
|
| 284316 |
+
"learning_rate": 7.66869143780291e-06,
|
| 284317 |
+
"loss": 0.266,
|
| 284318 |
+
"step": 145365
|
| 284319 |
+
},
|
| 284320 |
+
{
|
| 284321 |
+
"epoch": 1162.14,
|
| 284322 |
+
"learning_rate": 7.668610662358644e-06,
|
| 284323 |
+
"loss": 0.3346,
|
| 284324 |
+
"step": 145370
|
| 284325 |
+
},
|
| 284326 |
+
{
|
| 284327 |
+
"epoch": 1162.18,
|
| 284328 |
+
"learning_rate": 7.66852988691438e-06,
|
| 284329 |
+
"loss": 0.6964,
|
| 284330 |
+
"step": 145375
|
| 284331 |
+
},
|
| 284332 |
+
{
|
| 284333 |
+
"epoch": 1162.22,
|
| 284334 |
+
"learning_rate": 7.668449111470114e-06,
|
| 284335 |
+
"loss": 0.8126,
|
| 284336 |
+
"step": 145380
|
| 284337 |
+
},
|
| 284338 |
+
{
|
| 284339 |
+
"epoch": 1162.26,
|
| 284340 |
+
"learning_rate": 7.66836833602585e-06,
|
| 284341 |
+
"loss": 0.2484,
|
| 284342 |
+
"step": 145385
|
| 284343 |
+
},
|
| 284344 |
+
{
|
| 284345 |
+
"epoch": 1162.3,
|
| 284346 |
+
"learning_rate": 7.668287560581584e-06,
|
| 284347 |
+
"loss": 0.2748,
|
| 284348 |
+
"step": 145390
|
| 284349 |
+
},
|
| 284350 |
+
{
|
| 284351 |
+
"epoch": 1162.34,
|
| 284352 |
+
"learning_rate": 7.66820678513732e-06,
|
| 284353 |
+
"loss": 0.3347,
|
| 284354 |
+
"step": 145395
|
| 284355 |
+
},
|
| 284356 |
+
{
|
| 284357 |
+
"epoch": 1162.38,
|
| 284358 |
+
"learning_rate": 7.668126009693053e-06,
|
| 284359 |
+
"loss": 0.7334,
|
| 284360 |
+
"step": 145400
|
| 284361 |
+
},
|
| 284362 |
+
{
|
| 284363 |
+
"epoch": 1162.42,
|
| 284364 |
+
"learning_rate": 7.66804523424879e-06,
|
| 284365 |
+
"loss": 0.7622,
|
| 284366 |
+
"step": 145405
|
| 284367 |
+
},
|
| 284368 |
+
{
|
| 284369 |
+
"epoch": 1162.46,
|
| 284370 |
+
"learning_rate": 7.667964458804525e-06,
|
| 284371 |
+
"loss": 0.2724,
|
| 284372 |
+
"step": 145410
|
| 284373 |
+
},
|
| 284374 |
+
{
|
| 284375 |
+
"epoch": 1162.5,
|
| 284376 |
+
"learning_rate": 7.66788368336026e-06,
|
| 284377 |
+
"loss": 0.2608,
|
| 284378 |
+
"step": 145415
|
| 284379 |
+
},
|
| 284380 |
+
{
|
| 284381 |
+
"epoch": 1162.54,
|
| 284382 |
+
"learning_rate": 7.667802907915995e-06,
|
| 284383 |
+
"loss": 0.3223,
|
| 284384 |
+
"step": 145420
|
| 284385 |
+
},
|
| 284386 |
+
{
|
| 284387 |
+
"epoch": 1162.58,
|
| 284388 |
+
"learning_rate": 7.66772213247173e-06,
|
| 284389 |
+
"loss": 0.6718,
|
| 284390 |
+
"step": 145425
|
| 284391 |
+
},
|
| 284392 |
+
{
|
| 284393 |
+
"epoch": 1162.62,
|
| 284394 |
+
"learning_rate": 7.667641357027465e-06,
|
| 284395 |
+
"loss": 0.8369,
|
| 284396 |
+
"step": 145430
|
| 284397 |
+
},
|
| 284398 |
+
{
|
| 284399 |
+
"epoch": 1162.66,
|
| 284400 |
+
"learning_rate": 7.6675605815832e-06,
|
| 284401 |
+
"loss": 0.2615,
|
| 284402 |
+
"step": 145435
|
| 284403 |
+
},
|
| 284404 |
+
{
|
| 284405 |
+
"epoch": 1162.7,
|
| 284406 |
+
"learning_rate": 7.667479806138935e-06,
|
| 284407 |
+
"loss": 0.2988,
|
| 284408 |
+
"step": 145440
|
| 284409 |
+
},
|
| 284410 |
+
{
|
| 284411 |
+
"epoch": 1162.74,
|
| 284412 |
+
"learning_rate": 7.66739903069467e-06,
|
| 284413 |
+
"loss": 0.3295,
|
| 284414 |
+
"step": 145445
|
| 284415 |
+
},
|
| 284416 |
+
{
|
| 284417 |
+
"epoch": 1162.78,
|
| 284418 |
+
"learning_rate": 7.667318255250405e-06,
|
| 284419 |
+
"loss": 0.7126,
|
| 284420 |
+
"step": 145450
|
| 284421 |
+
},
|
| 284422 |
+
{
|
| 284423 |
+
"epoch": 1162.82,
|
| 284424 |
+
"learning_rate": 7.66723747980614e-06,
|
| 284425 |
+
"loss": 0.8154,
|
| 284426 |
+
"step": 145455
|
| 284427 |
+
},
|
| 284428 |
+
{
|
| 284429 |
+
"epoch": 1162.86,
|
| 284430 |
+
"learning_rate": 7.667156704361875e-06,
|
| 284431 |
+
"loss": 0.2407,
|
| 284432 |
+
"step": 145460
|
| 284433 |
+
},
|
| 284434 |
+
{
|
| 284435 |
+
"epoch": 1162.9,
|
| 284436 |
+
"learning_rate": 7.66707592891761e-06,
|
| 284437 |
+
"loss": 0.2759,
|
| 284438 |
+
"step": 145465
|
| 284439 |
+
},
|
| 284440 |
+
{
|
| 284441 |
+
"epoch": 1162.94,
|
| 284442 |
+
"learning_rate": 7.666995153473345e-06,
|
| 284443 |
+
"loss": 0.2949,
|
| 284444 |
+
"step": 145470
|
| 284445 |
+
},
|
| 284446 |
+
{
|
| 284447 |
+
"epoch": 1162.98,
|
| 284448 |
+
"learning_rate": 7.66691437802908e-06,
|
| 284449 |
+
"loss": 0.7172,
|
| 284450 |
+
"step": 145475
|
| 284451 |
+
},
|
| 284452 |
+
{
|
| 284453 |
+
"epoch": 1163.0,
|
| 284454 |
+
"eval_loss": 0.3563687801361084,
|
| 284455 |
+
"eval_runtime": 41.8928,
|
| 284456 |
+
"eval_samples_per_second": 19.98,
|
| 284457 |
+
"eval_steps_per_second": 0.645,
|
| 284458 |
+
"eval_wer": 0.17516339869281045,
|
| 284459 |
+
"step": 145477
|
| 284460 |
+
},
|
| 284461 |
+
{
|
| 284462 |
+
"epoch": 1163.02,
|
| 284463 |
+
"learning_rate": 7.666833602584815e-06,
|
| 284464 |
+
"loss": 0.3148,
|
| 284465 |
+
"step": 145480
|
| 284466 |
+
},
|
| 284467 |
+
{
|
| 284468 |
+
"epoch": 1163.06,
|
| 284469 |
+
"learning_rate": 7.66675282714055e-06,
|
| 284470 |
+
"loss": 0.2606,
|
| 284471 |
+
"step": 145485
|
| 284472 |
+
},
|
| 284473 |
+
{
|
| 284474 |
+
"epoch": 1163.1,
|
| 284475 |
+
"learning_rate": 7.666672051696285e-06,
|
| 284476 |
+
"loss": 0.3123,
|
| 284477 |
+
"step": 145490
|
| 284478 |
+
},
|
| 284479 |
+
{
|
| 284480 |
+
"epoch": 1163.14,
|
| 284481 |
+
"learning_rate": 7.66659127625202e-06,
|
| 284482 |
+
"loss": 0.3515,
|
| 284483 |
+
"step": 145495
|
| 284484 |
+
},
|
| 284485 |
+
{
|
| 284486 |
+
"epoch": 1163.18,
|
| 284487 |
+
"learning_rate": 7.666510500807755e-06,
|
| 284488 |
+
"loss": 0.7378,
|
| 284489 |
+
"step": 145500
|
| 284490 |
+
},
|
| 284491 |
+
{
|
| 284492 |
+
"epoch": 1163.22,
|
| 284493 |
+
"learning_rate": 7.66642972536349e-06,
|
| 284494 |
+
"loss": 0.8376,
|
| 284495 |
+
"step": 145505
|
| 284496 |
+
},
|
| 284497 |
+
{
|
| 284498 |
+
"epoch": 1163.26,
|
| 284499 |
+
"learning_rate": 7.666348949919225e-06,
|
| 284500 |
+
"loss": 0.3011,
|
| 284501 |
+
"step": 145510
|
| 284502 |
+
},
|
| 284503 |
+
{
|
| 284504 |
+
"epoch": 1163.3,
|
| 284505 |
+
"learning_rate": 7.66626817447496e-06,
|
| 284506 |
+
"loss": 0.3181,
|
| 284507 |
+
"step": 145515
|
| 284508 |
+
},
|
| 284509 |
+
{
|
| 284510 |
+
"epoch": 1163.34,
|
| 284511 |
+
"learning_rate": 7.666187399030695e-06,
|
| 284512 |
+
"loss": 0.4141,
|
| 284513 |
+
"step": 145520
|
| 284514 |
+
},
|
| 284515 |
+
{
|
| 284516 |
+
"epoch": 1163.38,
|
| 284517 |
+
"learning_rate": 7.66610662358643e-06,
|
| 284518 |
+
"loss": 0.7503,
|
| 284519 |
+
"step": 145525
|
| 284520 |
+
},
|
| 284521 |
+
{
|
| 284522 |
+
"epoch": 1163.42,
|
| 284523 |
+
"learning_rate": 7.666025848142165e-06,
|
| 284524 |
+
"loss": 0.8573,
|
| 284525 |
+
"step": 145530
|
| 284526 |
+
},
|
| 284527 |
+
{
|
| 284528 |
+
"epoch": 1163.46,
|
| 284529 |
+
"learning_rate": 7.6659450726979e-06,
|
| 284530 |
+
"loss": 0.283,
|
| 284531 |
+
"step": 145535
|
| 284532 |
+
},
|
| 284533 |
+
{
|
| 284534 |
+
"epoch": 1163.5,
|
| 284535 |
+
"learning_rate": 7.665864297253635e-06,
|
| 284536 |
+
"loss": 0.2771,
|
| 284537 |
+
"step": 145540
|
| 284538 |
+
},
|
| 284539 |
+
{
|
| 284540 |
+
"epoch": 1163.54,
|
| 284541 |
+
"learning_rate": 7.66578352180937e-06,
|
| 284542 |
+
"loss": 0.5021,
|
| 284543 |
+
"step": 145545
|
| 284544 |
+
},
|
| 284545 |
+
{
|
| 284546 |
+
"epoch": 1163.58,
|
| 284547 |
+
"learning_rate": 7.665702746365107e-06,
|
| 284548 |
+
"loss": 0.7615,
|
| 284549 |
+
"step": 145550
|
| 284550 |
+
},
|
| 284551 |
+
{
|
| 284552 |
+
"epoch": 1163.62,
|
| 284553 |
+
"learning_rate": 7.66562197092084e-06,
|
| 284554 |
+
"loss": 0.8149,
|
| 284555 |
+
"step": 145555
|
| 284556 |
+
},
|
| 284557 |
+
{
|
| 284558 |
+
"epoch": 1163.66,
|
| 284559 |
+
"learning_rate": 7.665541195476576e-06,
|
| 284560 |
+
"loss": 0.2333,
|
| 284561 |
+
"step": 145560
|
| 284562 |
+
},
|
| 284563 |
+
{
|
| 284564 |
+
"epoch": 1163.7,
|
| 284565 |
+
"learning_rate": 7.66546042003231e-06,
|
| 284566 |
+
"loss": 0.293,
|
| 284567 |
+
"step": 145565
|
| 284568 |
+
},
|
| 284569 |
+
{
|
| 284570 |
+
"epoch": 1163.74,
|
| 284571 |
+
"learning_rate": 7.665379644588046e-06,
|
| 284572 |
+
"loss": 0.3133,
|
| 284573 |
+
"step": 145570
|
| 284574 |
+
},
|
| 284575 |
+
{
|
| 284576 |
+
"epoch": 1163.78,
|
| 284577 |
+
"learning_rate": 7.66529886914378e-06,
|
| 284578 |
+
"loss": 0.693,
|
| 284579 |
+
"step": 145575
|
| 284580 |
+
},
|
| 284581 |
+
{
|
| 284582 |
+
"epoch": 1163.82,
|
| 284583 |
+
"learning_rate": 7.665218093699516e-06,
|
| 284584 |
+
"loss": 0.9545,
|
| 284585 |
+
"step": 145580
|
| 284586 |
+
},
|
| 284587 |
+
{
|
| 284588 |
+
"epoch": 1163.86,
|
| 284589 |
+
"learning_rate": 7.66513731825525e-06,
|
| 284590 |
+
"loss": 0.3008,
|
| 284591 |
+
"step": 145585
|
| 284592 |
+
},
|
| 284593 |
+
{
|
| 284594 |
+
"epoch": 1163.9,
|
| 284595 |
+
"learning_rate": 7.665056542810986e-06,
|
| 284596 |
+
"loss": 0.3013,
|
| 284597 |
+
"step": 145590
|
| 284598 |
+
},
|
| 284599 |
+
{
|
| 284600 |
+
"epoch": 1163.94,
|
| 284601 |
+
"learning_rate": 7.66497576736672e-06,
|
| 284602 |
+
"loss": 0.538,
|
| 284603 |
+
"step": 145595
|
| 284604 |
+
},
|
| 284605 |
+
{
|
| 284606 |
+
"epoch": 1163.98,
|
| 284607 |
+
"learning_rate": 7.664894991922456e-06,
|
| 284608 |
+
"loss": 0.7369,
|
| 284609 |
+
"step": 145600
|
| 284610 |
+
},
|
| 284611 |
+
{
|
| 284612 |
+
"epoch": 1164.0,
|
| 284613 |
+
"eval_loss": 0.3466392159461975,
|
| 284614 |
+
"eval_runtime": 39.3229,
|
| 284615 |
+
"eval_samples_per_second": 21.285,
|
| 284616 |
+
"eval_steps_per_second": 0.687,
|
| 284617 |
+
"eval_wer": 0.16716790141568252,
|
| 284618 |
+
"step": 145602
|
| 284619 |
}
|
| 284620 |
],
|
| 284621 |
+
"max_steps": 625000,
|
| 284622 |
"num_train_epochs": 5000,
|
| 284623 |
+
"total_flos": 4.097272814922444e+20,
|
| 284624 |
"trial_name": null,
|
| 284625 |
"trial_params": null
|
| 284626 |
}
|
model-bin/finetune/base/{checkpoint-144607 β checkpoint-145602}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-145228 β checkpoint-145852}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-145228 β checkpoint-145852}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6c3e8284682b82934af7c2cb7a47d0aad1afbbab74b9a20b0bdf99e1f1b28b90
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-145228 β checkpoint-145852}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-144607 β checkpoint-145852}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7400c5774686da0c05e6115d2e42b7e425eb92bed75c7039ed458db7dc841243
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-145228 β checkpoint-145852}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:791f31e71e6884d8a8929ef220193ccb17e2cfec3b0e299f7e8687cba4af3ce4
|
| 3 |
+
size 14503
|
model-bin/finetune/base/{checkpoint-144607 β checkpoint-145852}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:28a0b76f68ffd1c2d412704c5679e6f1d5de953fb95b5cacc4cb5d077d0d04b0
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-145228 β checkpoint-145852}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7dff24b5bd65b8a22b77667e8a8418682532690368c32e2fbb3b1785624a9976
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-144607 β checkpoint-145852}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
-
"best_metric": 0.
|
| 3 |
-
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -283350,11 +283350,1595 @@
|
|
| 283350 |
"eval_steps_per_second": 0.679,
|
| 283351 |
"eval_wer": 0.16790141568253503,
|
| 283352 |
"step": 144607
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 283353 |
}
|
| 283354 |
],
|
| 283355 |
-
"max_steps":
|
| 283356 |
"num_train_epochs": 5000,
|
| 283357 |
-
"total_flos": 4.
|
| 283358 |
"trial_name": null,
|
| 283359 |
"trial_params": null
|
| 283360 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"best_metric": 0.16716790141568252,
|
| 3 |
+
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-145602",
|
| 4 |
+
"epoch": 1166.0,
|
| 5 |
+
"global_step": 145852,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 283350 |
"eval_steps_per_second": 0.679,
|
| 283351 |
"eval_wer": 0.16790141568253503,
|
| 283352 |
"step": 144607
|
| 283353 |
+
},
|
| 283354 |
+
{
|
| 283355 |
+
"epoch": 1166.02,
|
| 283356 |
+
"learning_rate": 7.680888529886915e-06,
|
| 283357 |
+
"loss": 0.3216,
|
| 283358 |
+
"step": 144610
|
| 283359 |
+
},
|
| 283360 |
+
{
|
| 283361 |
+
"epoch": 1166.06,
|
| 283362 |
+
"learning_rate": 7.68080775444265e-06,
|
| 283363 |
+
"loss": 0.2238,
|
| 283364 |
+
"step": 144615
|
| 283365 |
+
},
|
| 283366 |
+
{
|
| 283367 |
+
"epoch": 1166.1,
|
| 283368 |
+
"learning_rate": 7.680726978998385e-06,
|
| 283369 |
+
"loss": 0.318,
|
| 283370 |
+
"step": 144620
|
| 283371 |
+
},
|
| 283372 |
+
{
|
| 283373 |
+
"epoch": 1166.14,
|
| 283374 |
+
"learning_rate": 7.680646203554121e-06,
|
| 283375 |
+
"loss": 0.357,
|
| 283376 |
+
"step": 144625
|
| 283377 |
+
},
|
| 283378 |
+
{
|
| 283379 |
+
"epoch": 1166.18,
|
| 283380 |
+
"learning_rate": 7.680565428109855e-06,
|
| 283381 |
+
"loss": 0.8028,
|
| 283382 |
+
"step": 144630
|
| 283383 |
+
},
|
| 283384 |
+
{
|
| 283385 |
+
"epoch": 1166.22,
|
| 283386 |
+
"learning_rate": 7.680484652665591e-06,
|
| 283387 |
+
"loss": 0.8776,
|
| 283388 |
+
"step": 144635
|
| 283389 |
+
},
|
| 283390 |
+
{
|
| 283391 |
+
"epoch": 1166.27,
|
| 283392 |
+
"learning_rate": 7.680403877221325e-06,
|
| 283393 |
+
"loss": 0.2889,
|
| 283394 |
+
"step": 144640
|
| 283395 |
+
},
|
| 283396 |
+
{
|
| 283397 |
+
"epoch": 1166.31,
|
| 283398 |
+
"learning_rate": 7.680323101777061e-06,
|
| 283399 |
+
"loss": 0.3034,
|
| 283400 |
+
"step": 144645
|
| 283401 |
+
},
|
| 283402 |
+
{
|
| 283403 |
+
"epoch": 1166.35,
|
| 283404 |
+
"learning_rate": 7.680242326332795e-06,
|
| 283405 |
+
"loss": 0.3402,
|
| 283406 |
+
"step": 144650
|
| 283407 |
+
},
|
| 283408 |
+
{
|
| 283409 |
+
"epoch": 1166.39,
|
| 283410 |
+
"learning_rate": 7.680161550888531e-06,
|
| 283411 |
+
"loss": 0.6687,
|
| 283412 |
+
"step": 144655
|
| 283413 |
+
},
|
| 283414 |
+
{
|
| 283415 |
+
"epoch": 1166.43,
|
| 283416 |
+
"learning_rate": 7.680080775444265e-06,
|
| 283417 |
+
"loss": 0.8415,
|
| 283418 |
+
"step": 144660
|
| 283419 |
+
},
|
| 283420 |
+
{
|
| 283421 |
+
"epoch": 1166.47,
|
| 283422 |
+
"learning_rate": 7.680000000000001e-06,
|
| 283423 |
+
"loss": 0.283,
|
| 283424 |
+
"step": 144665
|
| 283425 |
+
},
|
| 283426 |
+
{
|
| 283427 |
+
"epoch": 1166.51,
|
| 283428 |
+
"learning_rate": 7.679919224555735e-06,
|
| 283429 |
+
"loss": 0.2845,
|
| 283430 |
+
"step": 144670
|
| 283431 |
+
},
|
| 283432 |
+
{
|
| 283433 |
+
"epoch": 1166.55,
|
| 283434 |
+
"learning_rate": 7.679838449111471e-06,
|
| 283435 |
+
"loss": 0.3641,
|
| 283436 |
+
"step": 144675
|
| 283437 |
+
},
|
| 283438 |
+
{
|
| 283439 |
+
"epoch": 1166.59,
|
| 283440 |
+
"learning_rate": 7.679757673667205e-06,
|
| 283441 |
+
"loss": 0.6544,
|
| 283442 |
+
"step": 144680
|
| 283443 |
+
},
|
| 283444 |
+
{
|
| 283445 |
+
"epoch": 1166.63,
|
| 283446 |
+
"learning_rate": 7.679676898222941e-06,
|
| 283447 |
+
"loss": 0.8825,
|
| 283448 |
+
"step": 144685
|
| 283449 |
+
},
|
| 283450 |
+
{
|
| 283451 |
+
"epoch": 1166.67,
|
| 283452 |
+
"learning_rate": 7.679596122778677e-06,
|
| 283453 |
+
"loss": 0.308,
|
| 283454 |
+
"step": 144690
|
| 283455 |
+
},
|
| 283456 |
+
{
|
| 283457 |
+
"epoch": 1166.71,
|
| 283458 |
+
"learning_rate": 7.679515347334411e-06,
|
| 283459 |
+
"loss": 0.2938,
|
| 283460 |
+
"step": 144695
|
| 283461 |
+
},
|
| 283462 |
+
{
|
| 283463 |
+
"epoch": 1166.75,
|
| 283464 |
+
"learning_rate": 7.679434571890147e-06,
|
| 283465 |
+
"loss": 0.3421,
|
| 283466 |
+
"step": 144700
|
| 283467 |
+
},
|
| 283468 |
+
{
|
| 283469 |
+
"epoch": 1166.79,
|
| 283470 |
+
"learning_rate": 7.679353796445881e-06,
|
| 283471 |
+
"loss": 0.7238,
|
| 283472 |
+
"step": 144705
|
| 283473 |
+
},
|
| 283474 |
+
{
|
| 283475 |
+
"epoch": 1166.83,
|
| 283476 |
+
"learning_rate": 7.679273021001617e-06,
|
| 283477 |
+
"loss": 0.8722,
|
| 283478 |
+
"step": 144710
|
| 283479 |
+
},
|
| 283480 |
+
{
|
| 283481 |
+
"epoch": 1166.87,
|
| 283482 |
+
"learning_rate": 7.679192245557351e-06,
|
| 283483 |
+
"loss": 0.2458,
|
| 283484 |
+
"step": 144715
|
| 283485 |
+
},
|
| 283486 |
+
{
|
| 283487 |
+
"epoch": 1166.91,
|
| 283488 |
+
"learning_rate": 7.679111470113087e-06,
|
| 283489 |
+
"loss": 0.2701,
|
| 283490 |
+
"step": 144720
|
| 283491 |
+
},
|
| 283492 |
+
{
|
| 283493 |
+
"epoch": 1166.95,
|
| 283494 |
+
"learning_rate": 7.679030694668821e-06,
|
| 283495 |
+
"loss": 0.3769,
|
| 283496 |
+
"step": 144725
|
| 283497 |
+
},
|
| 283498 |
+
{
|
| 283499 |
+
"epoch": 1166.99,
|
| 283500 |
+
"learning_rate": 7.678949919224557e-06,
|
| 283501 |
+
"loss": 0.8216,
|
| 283502 |
+
"step": 144730
|
| 283503 |
+
},
|
| 283504 |
+
{
|
| 283505 |
+
"epoch": 1167.0,
|
| 283506 |
+
"eval_loss": 0.3796067237854004,
|
| 283507 |
+
"eval_runtime": 41.6728,
|
| 283508 |
+
"eval_samples_per_second": 20.085,
|
| 283509 |
+
"eval_steps_per_second": 0.648,
|
| 283510 |
+
"eval_wer": 0.18190386427898209,
|
| 283511 |
+
"step": 144731
|
| 283512 |
+
},
|
| 283513 |
+
{
|
| 283514 |
+
"epoch": 1167.03,
|
| 283515 |
+
"learning_rate": 7.67886914378029e-06,
|
| 283516 |
+
"loss": 0.2946,
|
| 283517 |
+
"step": 144735
|
| 283518 |
+
},
|
| 283519 |
+
{
|
| 283520 |
+
"epoch": 1167.07,
|
| 283521 |
+
"learning_rate": 7.678788368336027e-06,
|
| 283522 |
+
"loss": 0.2505,
|
| 283523 |
+
"step": 144740
|
| 283524 |
+
},
|
| 283525 |
+
{
|
| 283526 |
+
"epoch": 1167.11,
|
| 283527 |
+
"learning_rate": 7.678707592891763e-06,
|
| 283528 |
+
"loss": 0.2865,
|
| 283529 |
+
"step": 144745
|
| 283530 |
+
},
|
| 283531 |
+
{
|
| 283532 |
+
"epoch": 1167.15,
|
| 283533 |
+
"learning_rate": 7.678626817447497e-06,
|
| 283534 |
+
"loss": 0.3545,
|
| 283535 |
+
"step": 144750
|
| 283536 |
+
},
|
| 283537 |
+
{
|
| 283538 |
+
"epoch": 1167.19,
|
| 283539 |
+
"learning_rate": 7.678546042003232e-06,
|
| 283540 |
+
"loss": 0.9134,
|
| 283541 |
+
"step": 144755
|
| 283542 |
+
},
|
| 283543 |
+
{
|
| 283544 |
+
"epoch": 1167.23,
|
| 283545 |
+
"learning_rate": 7.678465266558967e-06,
|
| 283546 |
+
"loss": 0.655,
|
| 283547 |
+
"step": 144760
|
| 283548 |
+
},
|
| 283549 |
+
{
|
| 283550 |
+
"epoch": 1167.27,
|
| 283551 |
+
"learning_rate": 7.678384491114702e-06,
|
| 283552 |
+
"loss": 0.2705,
|
| 283553 |
+
"step": 144765
|
| 283554 |
+
},
|
| 283555 |
+
{
|
| 283556 |
+
"epoch": 1167.31,
|
| 283557 |
+
"learning_rate": 7.678303715670437e-06,
|
| 283558 |
+
"loss": 0.2562,
|
| 283559 |
+
"step": 144770
|
| 283560 |
+
},
|
| 283561 |
+
{
|
| 283562 |
+
"epoch": 1167.35,
|
| 283563 |
+
"learning_rate": 7.678222940226172e-06,
|
| 283564 |
+
"loss": 0.4007,
|
| 283565 |
+
"step": 144775
|
| 283566 |
+
},
|
| 283567 |
+
{
|
| 283568 |
+
"epoch": 1167.39,
|
| 283569 |
+
"learning_rate": 7.678142164781907e-06,
|
| 283570 |
+
"loss": 0.8377,
|
| 283571 |
+
"step": 144780
|
| 283572 |
+
},
|
| 283573 |
+
{
|
| 283574 |
+
"epoch": 1167.43,
|
| 283575 |
+
"learning_rate": 7.678061389337642e-06,
|
| 283576 |
+
"loss": 0.7035,
|
| 283577 |
+
"step": 144785
|
| 283578 |
+
},
|
| 283579 |
+
{
|
| 283580 |
+
"epoch": 1167.47,
|
| 283581 |
+
"learning_rate": 7.677980613893377e-06,
|
| 283582 |
+
"loss": 0.2795,
|
| 283583 |
+
"step": 144790
|
| 283584 |
+
},
|
| 283585 |
+
{
|
| 283586 |
+
"epoch": 1167.51,
|
| 283587 |
+
"learning_rate": 7.677899838449112e-06,
|
| 283588 |
+
"loss": 0.2829,
|
| 283589 |
+
"step": 144795
|
| 283590 |
+
},
|
| 283591 |
+
{
|
| 283592 |
+
"epoch": 1167.55,
|
| 283593 |
+
"learning_rate": 7.677819063004848e-06,
|
| 283594 |
+
"loss": 0.3805,
|
| 283595 |
+
"step": 144800
|
| 283596 |
+
},
|
| 283597 |
+
{
|
| 283598 |
+
"epoch": 1167.59,
|
| 283599 |
+
"learning_rate": 7.677738287560582e-06,
|
| 283600 |
+
"loss": 0.7594,
|
| 283601 |
+
"step": 144805
|
| 283602 |
+
},
|
| 283603 |
+
{
|
| 283604 |
+
"epoch": 1167.63,
|
| 283605 |
+
"learning_rate": 7.677657512116318e-06,
|
| 283606 |
+
"loss": 0.6312,
|
| 283607 |
+
"step": 144810
|
| 283608 |
+
},
|
| 283609 |
+
{
|
| 283610 |
+
"epoch": 1167.67,
|
| 283611 |
+
"learning_rate": 7.677576736672052e-06,
|
| 283612 |
+
"loss": 0.257,
|
| 283613 |
+
"step": 144815
|
| 283614 |
+
},
|
| 283615 |
+
{
|
| 283616 |
+
"epoch": 1167.71,
|
| 283617 |
+
"learning_rate": 7.677495961227788e-06,
|
| 283618 |
+
"loss": 0.2697,
|
| 283619 |
+
"step": 144820
|
| 283620 |
+
},
|
| 283621 |
+
{
|
| 283622 |
+
"epoch": 1167.76,
|
| 283623 |
+
"learning_rate": 7.677415185783522e-06,
|
| 283624 |
+
"loss": 0.3065,
|
| 283625 |
+
"step": 144825
|
| 283626 |
+
},
|
| 283627 |
+
{
|
| 283628 |
+
"epoch": 1167.8,
|
| 283629 |
+
"learning_rate": 7.677334410339258e-06,
|
| 283630 |
+
"loss": 0.8168,
|
| 283631 |
+
"step": 144830
|
| 283632 |
+
},
|
| 283633 |
+
{
|
| 283634 |
+
"epoch": 1167.84,
|
| 283635 |
+
"learning_rate": 7.677253634894992e-06,
|
| 283636 |
+
"loss": 0.6347,
|
| 283637 |
+
"step": 144835
|
| 283638 |
+
},
|
| 283639 |
+
{
|
| 283640 |
+
"epoch": 1167.88,
|
| 283641 |
+
"learning_rate": 7.677172859450728e-06,
|
| 283642 |
+
"loss": 0.2511,
|
| 283643 |
+
"step": 144840
|
| 283644 |
+
},
|
| 283645 |
+
{
|
| 283646 |
+
"epoch": 1167.92,
|
| 283647 |
+
"learning_rate": 7.677092084006462e-06,
|
| 283648 |
+
"loss": 0.2972,
|
| 283649 |
+
"step": 144845
|
| 283650 |
+
},
|
| 283651 |
+
{
|
| 283652 |
+
"epoch": 1167.96,
|
| 283653 |
+
"learning_rate": 7.677011308562198e-06,
|
| 283654 |
+
"loss": 0.4141,
|
| 283655 |
+
"step": 144850
|
| 283656 |
+
},
|
| 283657 |
+
{
|
| 283658 |
+
"epoch": 1168.0,
|
| 283659 |
+
"learning_rate": 7.676930533117932e-06,
|
| 283660 |
+
"loss": 1.1673,
|
| 283661 |
+
"step": 144855
|
| 283662 |
+
},
|
| 283663 |
+
{
|
| 283664 |
+
"epoch": 1168.0,
|
| 283665 |
+
"eval_loss": 0.34798669815063477,
|
| 283666 |
+
"eval_runtime": 41.3687,
|
| 283667 |
+
"eval_samples_per_second": 20.233,
|
| 283668 |
+
"eval_steps_per_second": 0.653,
|
| 283669 |
+
"eval_wer": 0.17150972252567184,
|
| 283670 |
+
"step": 144855
|
| 283671 |
+
},
|
| 283672 |
+
{
|
| 283673 |
+
"epoch": 1158.04,
|
| 283674 |
+
"learning_rate": 7.676849757673668e-06,
|
| 283675 |
+
"loss": 0.3284,
|
| 283676 |
+
"step": 144860
|
| 283677 |
+
},
|
| 283678 |
+
{
|
| 283679 |
+
"epoch": 1158.08,
|
| 283680 |
+
"learning_rate": 7.676768982229404e-06,
|
| 283681 |
+
"loss": 0.2587,
|
| 283682 |
+
"step": 144865
|
| 283683 |
+
},
|
| 283684 |
+
{
|
| 283685 |
+
"epoch": 1158.12,
|
| 283686 |
+
"learning_rate": 7.676688206785138e-06,
|
| 283687 |
+
"loss": 0.3396,
|
| 283688 |
+
"step": 144870
|
| 283689 |
+
},
|
| 283690 |
+
{
|
| 283691 |
+
"epoch": 1158.16,
|
| 283692 |
+
"learning_rate": 7.676607431340874e-06,
|
| 283693 |
+
"loss": 0.4414,
|
| 283694 |
+
"step": 144875
|
| 283695 |
+
},
|
| 283696 |
+
{
|
| 283697 |
+
"epoch": 1158.2,
|
| 283698 |
+
"learning_rate": 7.676526655896608e-06,
|
| 283699 |
+
"loss": 1.0351,
|
| 283700 |
+
"step": 144880
|
| 283701 |
+
},
|
| 283702 |
+
{
|
| 283703 |
+
"epoch": 1158.24,
|
| 283704 |
+
"learning_rate": 7.676445880452344e-06,
|
| 283705 |
+
"loss": 0.2941,
|
| 283706 |
+
"step": 144885
|
| 283707 |
+
},
|
| 283708 |
+
{
|
| 283709 |
+
"epoch": 1158.28,
|
| 283710 |
+
"learning_rate": 7.676365105008078e-06,
|
| 283711 |
+
"loss": 0.2294,
|
| 283712 |
+
"step": 144890
|
| 283713 |
+
},
|
| 283714 |
+
{
|
| 283715 |
+
"epoch": 1158.32,
|
| 283716 |
+
"learning_rate": 7.676284329563814e-06,
|
| 283717 |
+
"loss": 0.326,
|
| 283718 |
+
"step": 144895
|
| 283719 |
+
},
|
| 283720 |
+
{
|
| 283721 |
+
"epoch": 1158.36,
|
| 283722 |
+
"learning_rate": 7.676203554119548e-06,
|
| 283723 |
+
"loss": 0.4232,
|
| 283724 |
+
"step": 144900
|
| 283725 |
+
},
|
| 283726 |
+
{
|
| 283727 |
+
"epoch": 1158.4,
|
| 283728 |
+
"learning_rate": 7.676122778675284e-06,
|
| 283729 |
+
"loss": 1.1964,
|
| 283730 |
+
"step": 144905
|
| 283731 |
+
},
|
| 283732 |
+
{
|
| 283733 |
+
"epoch": 1158.44,
|
| 283734 |
+
"learning_rate": 7.676042003231018e-06,
|
| 283735 |
+
"loss": 0.2844,
|
| 283736 |
+
"step": 144910
|
| 283737 |
+
},
|
| 283738 |
+
{
|
| 283739 |
+
"epoch": 1158.48,
|
| 283740 |
+
"learning_rate": 7.675961227786754e-06,
|
| 283741 |
+
"loss": 0.2528,
|
| 283742 |
+
"step": 144915
|
| 283743 |
+
},
|
| 283744 |
+
{
|
| 283745 |
+
"epoch": 1158.52,
|
| 283746 |
+
"learning_rate": 7.67588045234249e-06,
|
| 283747 |
+
"loss": 0.313,
|
| 283748 |
+
"step": 144920
|
| 283749 |
+
},
|
| 283750 |
+
{
|
| 283751 |
+
"epoch": 1158.56,
|
| 283752 |
+
"learning_rate": 7.675799676898224e-06,
|
| 283753 |
+
"loss": 0.4891,
|
| 283754 |
+
"step": 144925
|
| 283755 |
+
},
|
| 283756 |
+
{
|
| 283757 |
+
"epoch": 1158.6,
|
| 283758 |
+
"learning_rate": 7.67571890145396e-06,
|
| 283759 |
+
"loss": 1.3052,
|
| 283760 |
+
"step": 144930
|
| 283761 |
+
},
|
| 283762 |
+
{
|
| 283763 |
+
"epoch": 1158.64,
|
| 283764 |
+
"learning_rate": 7.675638126009694e-06,
|
| 283765 |
+
"loss": 0.3103,
|
| 283766 |
+
"step": 144935
|
| 283767 |
+
},
|
| 283768 |
+
{
|
| 283769 |
+
"epoch": 1158.68,
|
| 283770 |
+
"learning_rate": 7.67555735056543e-06,
|
| 283771 |
+
"loss": 0.3308,
|
| 283772 |
+
"step": 144940
|
| 283773 |
+
},
|
| 283774 |
+
{
|
| 283775 |
+
"epoch": 1158.72,
|
| 283776 |
+
"learning_rate": 7.675476575121164e-06,
|
| 283777 |
+
"loss": 0.3203,
|
| 283778 |
+
"step": 144945
|
| 283779 |
+
},
|
| 283780 |
+
{
|
| 283781 |
+
"epoch": 1158.76,
|
| 283782 |
+
"learning_rate": 7.6753957996769e-06,
|
| 283783 |
+
"loss": 0.3745,
|
| 283784 |
+
"step": 144950
|
| 283785 |
+
},
|
| 283786 |
+
{
|
| 283787 |
+
"epoch": 1158.8,
|
| 283788 |
+
"learning_rate": 7.675315024232634e-06,
|
| 283789 |
+
"loss": 1.1487,
|
| 283790 |
+
"step": 144955
|
| 283791 |
+
},
|
| 283792 |
+
{
|
| 283793 |
+
"epoch": 1158.84,
|
| 283794 |
+
"learning_rate": 7.67523424878837e-06,
|
| 283795 |
+
"loss": 0.4522,
|
| 283796 |
+
"step": 144960
|
| 283797 |
+
},
|
| 283798 |
+
{
|
| 283799 |
+
"epoch": 1158.88,
|
| 283800 |
+
"learning_rate": 7.675153473344104e-06,
|
| 283801 |
+
"loss": 0.3011,
|
| 283802 |
+
"step": 144965
|
| 283803 |
+
},
|
| 283804 |
+
{
|
| 283805 |
+
"epoch": 1158.92,
|
| 283806 |
+
"learning_rate": 7.67507269789984e-06,
|
| 283807 |
+
"loss": 0.2663,
|
| 283808 |
+
"step": 144970
|
| 283809 |
+
},
|
| 283810 |
+
{
|
| 283811 |
+
"epoch": 1158.96,
|
| 283812 |
+
"learning_rate": 7.674991922455575e-06,
|
| 283813 |
+
"loss": 0.4301,
|
| 283814 |
+
"step": 144975
|
| 283815 |
+
},
|
| 283816 |
+
{
|
| 283817 |
+
"epoch": 1159.0,
|
| 283818 |
+
"learning_rate": 7.67491114701131e-06,
|
| 283819 |
+
"loss": 1.3414,
|
| 283820 |
+
"step": 144980
|
| 283821 |
+
},
|
| 283822 |
+
{
|
| 283823 |
+
"epoch": 1159.0,
|
| 283824 |
+
"eval_loss": 0.3729027807712555,
|
| 283825 |
+
"eval_runtime": 40.7509,
|
| 283826 |
+
"eval_samples_per_second": 20.539,
|
| 283827 |
+
"eval_steps_per_second": 0.663,
|
| 283828 |
+
"eval_wer": 0.17167350527549824,
|
| 283829 |
+
"step": 144980
|
| 283830 |
+
},
|
| 283831 |
+
{
|
| 283832 |
+
"epoch": 1169.04,
|
| 283833 |
+
"learning_rate": 7.674830371567045e-06,
|
| 283834 |
+
"loss": 0.3148,
|
| 283835 |
+
"step": 144985
|
| 283836 |
+
},
|
| 283837 |
+
{
|
| 283838 |
+
"epoch": 1169.08,
|
| 283839 |
+
"learning_rate": 7.67474959612278e-06,
|
| 283840 |
+
"loss": 0.2913,
|
| 283841 |
+
"step": 144990
|
| 283842 |
+
},
|
| 283843 |
+
{
|
| 283844 |
+
"epoch": 1169.12,
|
| 283845 |
+
"learning_rate": 7.674668820678515e-06,
|
| 283846 |
+
"loss": 0.2694,
|
| 283847 |
+
"step": 144995
|
| 283848 |
+
},
|
| 283849 |
+
{
|
| 283850 |
+
"epoch": 1169.16,
|
| 283851 |
+
"learning_rate": 7.67458804523425e-06,
|
| 283852 |
+
"loss": 0.478,
|
| 283853 |
+
"step": 145000
|
| 283854 |
+
},
|
| 283855 |
+
{
|
| 283856 |
+
"epoch": 1169.2,
|
| 283857 |
+
"learning_rate": 7.674507269789985e-06,
|
| 283858 |
+
"loss": 1.1648,
|
| 283859 |
+
"step": 145005
|
| 283860 |
+
},
|
| 283861 |
+
{
|
| 283862 |
+
"epoch": 1169.24,
|
| 283863 |
+
"learning_rate": 7.67442649434572e-06,
|
| 283864 |
+
"loss": 0.3068,
|
| 283865 |
+
"step": 145010
|
| 283866 |
+
},
|
| 283867 |
+
{
|
| 283868 |
+
"epoch": 1169.28,
|
| 283869 |
+
"learning_rate": 7.674345718901455e-06,
|
| 283870 |
+
"loss": 0.2923,
|
| 283871 |
+
"step": 145015
|
| 283872 |
+
},
|
| 283873 |
+
{
|
| 283874 |
+
"epoch": 1169.32,
|
| 283875 |
+
"learning_rate": 7.67426494345719e-06,
|
| 283876 |
+
"loss": 0.277,
|
| 283877 |
+
"step": 145020
|
| 283878 |
+
},
|
| 283879 |
+
{
|
| 283880 |
+
"epoch": 1169.36,
|
| 283881 |
+
"learning_rate": 7.674184168012925e-06,
|
| 283882 |
+
"loss": 0.5259,
|
| 283883 |
+
"step": 145025
|
| 283884 |
+
},
|
| 283885 |
+
{
|
| 283886 |
+
"epoch": 1169.4,
|
| 283887 |
+
"learning_rate": 7.67410339256866e-06,
|
| 283888 |
+
"loss": 1.3239,
|
| 283889 |
+
"step": 145030
|
| 283890 |
+
},
|
| 283891 |
+
{
|
| 283892 |
+
"epoch": 1169.44,
|
| 283893 |
+
"learning_rate": 7.674022617124395e-06,
|
| 283894 |
+
"loss": 0.2994,
|
| 283895 |
+
"step": 145035
|
| 283896 |
+
},
|
| 283897 |
+
{
|
| 283898 |
+
"epoch": 1169.48,
|
| 283899 |
+
"learning_rate": 7.673941841680131e-06,
|
| 283900 |
+
"loss": 0.2812,
|
| 283901 |
+
"step": 145040
|
| 283902 |
+
},
|
| 283903 |
+
{
|
| 283904 |
+
"epoch": 1169.52,
|
| 283905 |
+
"learning_rate": 7.673861066235865e-06,
|
| 283906 |
+
"loss": 0.2935,
|
| 283907 |
+
"step": 145045
|
| 283908 |
+
},
|
| 283909 |
+
{
|
| 283910 |
+
"epoch": 1169.56,
|
| 283911 |
+
"learning_rate": 7.673780290791601e-06,
|
| 283912 |
+
"loss": 0.4962,
|
| 283913 |
+
"step": 145050
|
| 283914 |
+
},
|
| 283915 |
+
{
|
| 283916 |
+
"epoch": 1169.6,
|
| 283917 |
+
"learning_rate": 7.673699515347335e-06,
|
| 283918 |
+
"loss": 1.2089,
|
| 283919 |
+
"step": 145055
|
| 283920 |
+
},
|
| 283921 |
+
{
|
| 283922 |
+
"epoch": 1169.64,
|
| 283923 |
+
"learning_rate": 7.673618739903071e-06,
|
| 283924 |
+
"loss": 0.2642,
|
| 283925 |
+
"step": 145060
|
| 283926 |
+
},
|
| 283927 |
+
{
|
| 283928 |
+
"epoch": 1169.68,
|
| 283929 |
+
"learning_rate": 7.673537964458805e-06,
|
| 283930 |
+
"loss": 0.2744,
|
| 283931 |
+
"step": 145065
|
| 283932 |
+
},
|
| 283933 |
+
{
|
| 283934 |
+
"epoch": 1169.72,
|
| 283935 |
+
"learning_rate": 7.673457189014541e-06,
|
| 283936 |
+
"loss": 0.278,
|
| 283937 |
+
"step": 145070
|
| 283938 |
+
},
|
| 283939 |
+
{
|
| 283940 |
+
"epoch": 1169.76,
|
| 283941 |
+
"learning_rate": 7.673376413570275e-06,
|
| 283942 |
+
"loss": 0.4382,
|
| 283943 |
+
"step": 145075
|
| 283944 |
+
},
|
| 283945 |
+
{
|
| 283946 |
+
"epoch": 1169.8,
|
| 283947 |
+
"learning_rate": 7.673295638126011e-06,
|
| 283948 |
+
"loss": 1.0989,
|
| 283949 |
+
"step": 145080
|
| 283950 |
+
},
|
| 283951 |
+
{
|
| 283952 |
+
"epoch": 1169.84,
|
| 283953 |
+
"learning_rate": 7.673214862681745e-06,
|
| 283954 |
+
"loss": 0.3134,
|
| 283955 |
+
"step": 145085
|
| 283956 |
+
},
|
| 283957 |
+
{
|
| 283958 |
+
"epoch": 1169.88,
|
| 283959 |
+
"learning_rate": 7.673134087237481e-06,
|
| 283960 |
+
"loss": 0.3014,
|
| 283961 |
+
"step": 145090
|
| 283962 |
+
},
|
| 283963 |
+
{
|
| 283964 |
+
"epoch": 1169.92,
|
| 283965 |
+
"learning_rate": 7.673053311793217e-06,
|
| 283966 |
+
"loss": 0.2862,
|
| 283967 |
+
"step": 145095
|
| 283968 |
+
},
|
| 283969 |
+
{
|
| 283970 |
+
"epoch": 1169.96,
|
| 283971 |
+
"learning_rate": 7.672972536348951e-06,
|
| 283972 |
+
"loss": 0.5695,
|
| 283973 |
+
"step": 145100
|
| 283974 |
+
},
|
| 283975 |
+
{
|
| 283976 |
+
"epoch": 1170.0,
|
| 283977 |
+
"eval_loss": 0.39969587326049805,
|
| 283978 |
+
"eval_runtime": 40.575,
|
| 283979 |
+
"eval_samples_per_second": 20.628,
|
| 283980 |
+
"eval_steps_per_second": 0.665,
|
| 283981 |
+
"eval_wer": 0.1783857926653191,
|
| 283982 |
+
"step": 145104
|
| 283983 |
+
},
|
| 283984 |
+
{
|
| 283985 |
+
"epoch": 1170.01,
|
| 283986 |
+
"learning_rate": 7.672891760904687e-06,
|
| 283987 |
+
"loss": 0.3477,
|
| 283988 |
+
"step": 145105
|
| 283989 |
+
},
|
| 283990 |
+
{
|
| 283991 |
+
"epoch": 1170.05,
|
| 283992 |
+
"learning_rate": 7.67281098546042e-06,
|
| 283993 |
+
"loss": 0.3213,
|
| 283994 |
+
"step": 145110
|
| 283995 |
+
},
|
| 283996 |
+
{
|
| 283997 |
+
"epoch": 1170.09,
|
| 283998 |
+
"learning_rate": 7.672730210016157e-06,
|
| 283999 |
+
"loss": 0.275,
|
| 284000 |
+
"step": 145115
|
| 284001 |
+
},
|
| 284002 |
+
{
|
| 284003 |
+
"epoch": 1170.13,
|
| 284004 |
+
"learning_rate": 7.67264943457189e-06,
|
| 284005 |
+
"loss": 0.3189,
|
| 284006 |
+
"step": 145120
|
| 284007 |
+
},
|
| 284008 |
+
{
|
| 284009 |
+
"epoch": 1170.17,
|
| 284010 |
+
"learning_rate": 7.672568659127627e-06,
|
| 284011 |
+
"loss": 0.5509,
|
| 284012 |
+
"step": 145125
|
| 284013 |
+
},
|
| 284014 |
+
{
|
| 284015 |
+
"epoch": 1170.21,
|
| 284016 |
+
"learning_rate": 7.67248788368336e-06,
|
| 284017 |
+
"loss": 1.0719,
|
| 284018 |
+
"step": 145130
|
| 284019 |
+
},
|
| 284020 |
+
{
|
| 284021 |
+
"epoch": 1170.25,
|
| 284022 |
+
"learning_rate": 7.672407108239097e-06,
|
| 284023 |
+
"loss": 0.2919,
|
| 284024 |
+
"step": 145135
|
| 284025 |
+
},
|
| 284026 |
+
{
|
| 284027 |
+
"epoch": 1170.29,
|
| 284028 |
+
"learning_rate": 7.67232633279483e-06,
|
| 284029 |
+
"loss": 0.2532,
|
| 284030 |
+
"step": 145140
|
| 284031 |
+
},
|
| 284032 |
+
{
|
| 284033 |
+
"epoch": 1170.33,
|
| 284034 |
+
"learning_rate": 7.672245557350567e-06,
|
| 284035 |
+
"loss": 0.3189,
|
| 284036 |
+
"step": 145145
|
| 284037 |
+
},
|
| 284038 |
+
{
|
| 284039 |
+
"epoch": 1170.37,
|
| 284040 |
+
"learning_rate": 7.672164781906302e-06,
|
| 284041 |
+
"loss": 0.4942,
|
| 284042 |
+
"step": 145150
|
| 284043 |
+
},
|
| 284044 |
+
{
|
| 284045 |
+
"epoch": 1170.41,
|
| 284046 |
+
"learning_rate": 7.672084006462037e-06,
|
| 284047 |
+
"loss": 1.1852,
|
| 284048 |
+
"step": 145155
|
| 284049 |
+
},
|
| 284050 |
+
{
|
| 284051 |
+
"epoch": 1170.45,
|
| 284052 |
+
"learning_rate": 7.672003231017772e-06,
|
| 284053 |
+
"loss": 0.3175,
|
| 284054 |
+
"step": 145160
|
| 284055 |
+
},
|
| 284056 |
+
{
|
| 284057 |
+
"epoch": 1170.49,
|
| 284058 |
+
"learning_rate": 7.671922455573507e-06,
|
| 284059 |
+
"loss": 0.3137,
|
| 284060 |
+
"step": 145165
|
| 284061 |
+
},
|
| 284062 |
+
{
|
| 284063 |
+
"epoch": 1170.53,
|
| 284064 |
+
"learning_rate": 7.671841680129242e-06,
|
| 284065 |
+
"loss": 0.2959,
|
| 284066 |
+
"step": 145170
|
| 284067 |
+
},
|
| 284068 |
+
{
|
| 284069 |
+
"epoch": 1170.57,
|
| 284070 |
+
"learning_rate": 7.671760904684977e-06,
|
| 284071 |
+
"loss": 0.4471,
|
| 284072 |
+
"step": 145175
|
| 284073 |
+
},
|
| 284074 |
+
{
|
| 284075 |
+
"epoch": 1170.61,
|
| 284076 |
+
"learning_rate": 7.671680129240712e-06,
|
| 284077 |
+
"loss": 1.0092,
|
| 284078 |
+
"step": 145180
|
| 284079 |
+
},
|
| 284080 |
+
{
|
| 284081 |
+
"epoch": 1170.65,
|
| 284082 |
+
"learning_rate": 7.671599353796446e-06,
|
| 284083 |
+
"loss": 0.3617,
|
| 284084 |
+
"step": 145185
|
| 284085 |
+
},
|
| 284086 |
+
{
|
| 284087 |
+
"epoch": 1170.69,
|
| 284088 |
+
"learning_rate": 7.671518578352182e-06,
|
| 284089 |
+
"loss": 0.3141,
|
| 284090 |
+
"step": 145190
|
| 284091 |
+
},
|
| 284092 |
+
{
|
| 284093 |
+
"epoch": 1170.73,
|
| 284094 |
+
"learning_rate": 7.671437802907916e-06,
|
| 284095 |
+
"loss": 0.2982,
|
| 284096 |
+
"step": 145195
|
| 284097 |
+
},
|
| 284098 |
+
{
|
| 284099 |
+
"epoch": 1170.77,
|
| 284100 |
+
"learning_rate": 7.671357027463652e-06,
|
| 284101 |
+
"loss": 0.4728,
|
| 284102 |
+
"step": 145200
|
| 284103 |
+
},
|
| 284104 |
+
{
|
| 284105 |
+
"epoch": 1170.81,
|
| 284106 |
+
"learning_rate": 7.671276252019386e-06,
|
| 284107 |
+
"loss": 1.086,
|
| 284108 |
+
"step": 145205
|
| 284109 |
+
},
|
| 284110 |
+
{
|
| 284111 |
+
"epoch": 1170.85,
|
| 284112 |
+
"learning_rate": 7.671195476575122e-06,
|
| 284113 |
+
"loss": 0.3326,
|
| 284114 |
+
"step": 145210
|
| 284115 |
+
},
|
| 284116 |
+
{
|
| 284117 |
+
"epoch": 1170.89,
|
| 284118 |
+
"learning_rate": 7.671114701130858e-06,
|
| 284119 |
+
"loss": 0.2725,
|
| 284120 |
+
"step": 145215
|
| 284121 |
+
},
|
| 284122 |
+
{
|
| 284123 |
+
"epoch": 1170.93,
|
| 284124 |
+
"learning_rate": 7.671033925686592e-06,
|
| 284125 |
+
"loss": 0.3219,
|
| 284126 |
+
"step": 145220
|
| 284127 |
+
},
|
| 284128 |
+
{
|
| 284129 |
+
"epoch": 1170.97,
|
| 284130 |
+
"learning_rate": 7.670953150242328e-06,
|
| 284131 |
+
"loss": 0.5993,
|
| 284132 |
+
"step": 145225
|
| 284133 |
+
},
|
| 284134 |
+
{
|
| 284135 |
+
"epoch": 1171.0,
|
| 284136 |
+
"eval_loss": 0.3494146466255188,
|
| 284137 |
+
"eval_runtime": 40.4075,
|
| 284138 |
+
"eval_samples_per_second": 20.714,
|
| 284139 |
+
"eval_steps_per_second": 0.668,
|
| 284140 |
+
"eval_wer": 0.17565191913272782,
|
| 284141 |
+
"step": 145228
|
| 284142 |
+
},
|
| 284143 |
+
{
|
| 284144 |
+
"epoch": 1171.02,
|
| 284145 |
+
"learning_rate": 7.670872374798062e-06,
|
| 284146 |
+
"loss": 0.3162,
|
| 284147 |
+
"step": 145230
|
| 284148 |
+
},
|
| 284149 |
+
{
|
| 284150 |
+
"epoch": 1171.06,
|
| 284151 |
+
"learning_rate": 7.670791599353798e-06,
|
| 284152 |
+
"loss": 0.268,
|
| 284153 |
+
"step": 145235
|
| 284154 |
+
},
|
| 284155 |
+
{
|
| 284156 |
+
"epoch": 1171.1,
|
| 284157 |
+
"learning_rate": 7.670710823909532e-06,
|
| 284158 |
+
"loss": 0.3148,
|
| 284159 |
+
"step": 145240
|
| 284160 |
+
},
|
| 284161 |
+
{
|
| 284162 |
+
"epoch": 1171.14,
|
| 284163 |
+
"learning_rate": 7.670630048465268e-06,
|
| 284164 |
+
"loss": 0.3431,
|
| 284165 |
+
"step": 145245
|
| 284166 |
+
},
|
| 284167 |
+
{
|
| 284168 |
+
"epoch": 1171.18,
|
| 284169 |
+
"learning_rate": 7.670549273021002e-06,
|
| 284170 |
+
"loss": 0.5562,
|
| 284171 |
+
"step": 145250
|
| 284172 |
+
},
|
| 284173 |
+
{
|
| 284174 |
+
"epoch": 1171.22,
|
| 284175 |
+
"learning_rate": 7.670468497576738e-06,
|
| 284176 |
+
"loss": 0.8706,
|
| 284177 |
+
"step": 145255
|
| 284178 |
+
},
|
| 284179 |
+
{
|
| 284180 |
+
"epoch": 1171.26,
|
| 284181 |
+
"learning_rate": 7.670387722132472e-06,
|
| 284182 |
+
"loss": 0.283,
|
| 284183 |
+
"step": 145260
|
| 284184 |
+
},
|
| 284185 |
+
{
|
| 284186 |
+
"epoch": 1171.3,
|
| 284187 |
+
"learning_rate": 7.670306946688208e-06,
|
| 284188 |
+
"loss": 0.3143,
|
| 284189 |
+
"step": 145265
|
| 284190 |
+
},
|
| 284191 |
+
{
|
| 284192 |
+
"epoch": 1171.34,
|
| 284193 |
+
"learning_rate": 7.670226171243944e-06,
|
| 284194 |
+
"loss": 0.2854,
|
| 284195 |
+
"step": 145270
|
| 284196 |
+
},
|
| 284197 |
+
{
|
| 284198 |
+
"epoch": 1171.38,
|
| 284199 |
+
"learning_rate": 7.670145395799678e-06,
|
| 284200 |
+
"loss": 0.5725,
|
| 284201 |
+
"step": 145275
|
| 284202 |
+
},
|
| 284203 |
+
{
|
| 284204 |
+
"epoch": 1171.42,
|
| 284205 |
+
"learning_rate": 7.670064620355414e-06,
|
| 284206 |
+
"loss": 0.9977,
|
| 284207 |
+
"step": 145280
|
| 284208 |
+
},
|
| 284209 |
+
{
|
| 284210 |
+
"epoch": 1171.46,
|
| 284211 |
+
"learning_rate": 7.669983844911148e-06,
|
| 284212 |
+
"loss": 0.2628,
|
| 284213 |
+
"step": 145285
|
| 284214 |
+
},
|
| 284215 |
+
{
|
| 284216 |
+
"epoch": 1171.5,
|
| 284217 |
+
"learning_rate": 7.669903069466884e-06,
|
| 284218 |
+
"loss": 0.2448,
|
| 284219 |
+
"step": 145290
|
| 284220 |
+
},
|
| 284221 |
+
{
|
| 284222 |
+
"epoch": 1171.54,
|
| 284223 |
+
"learning_rate": 7.669822294022618e-06,
|
| 284224 |
+
"loss": 0.3178,
|
| 284225 |
+
"step": 145295
|
| 284226 |
+
},
|
| 284227 |
+
{
|
| 284228 |
+
"epoch": 1171.58,
|
| 284229 |
+
"learning_rate": 7.669741518578354e-06,
|
| 284230 |
+
"loss": 0.6503,
|
| 284231 |
+
"step": 145300
|
| 284232 |
+
},
|
| 284233 |
+
{
|
| 284234 |
+
"epoch": 1171.62,
|
| 284235 |
+
"learning_rate": 7.669660743134088e-06,
|
| 284236 |
+
"loss": 0.9847,
|
| 284237 |
+
"step": 145305
|
| 284238 |
+
},
|
| 284239 |
+
{
|
| 284240 |
+
"epoch": 1171.66,
|
| 284241 |
+
"learning_rate": 7.669579967689824e-06,
|
| 284242 |
+
"loss": 0.2887,
|
| 284243 |
+
"step": 145310
|
| 284244 |
+
},
|
| 284245 |
+
{
|
| 284246 |
+
"epoch": 1171.7,
|
| 284247 |
+
"learning_rate": 7.669499192245558e-06,
|
| 284248 |
+
"loss": 0.2481,
|
| 284249 |
+
"step": 145315
|
| 284250 |
+
},
|
| 284251 |
+
{
|
| 284252 |
+
"epoch": 1171.74,
|
| 284253 |
+
"learning_rate": 7.669418416801294e-06,
|
| 284254 |
+
"loss": 0.2989,
|
| 284255 |
+
"step": 145320
|
| 284256 |
+
},
|
| 284257 |
+
{
|
| 284258 |
+
"epoch": 1171.78,
|
| 284259 |
+
"learning_rate": 7.669337641357028e-06,
|
| 284260 |
+
"loss": 0.6053,
|
| 284261 |
+
"step": 145325
|
| 284262 |
+
},
|
| 284263 |
+
{
|
| 284264 |
+
"epoch": 1171.82,
|
| 284265 |
+
"learning_rate": 7.669256865912764e-06,
|
| 284266 |
+
"loss": 0.9806,
|
| 284267 |
+
"step": 145330
|
| 284268 |
+
},
|
| 284269 |
+
{
|
| 284270 |
+
"epoch": 1171.86,
|
| 284271 |
+
"learning_rate": 7.669176090468498e-06,
|
| 284272 |
+
"loss": 0.3248,
|
| 284273 |
+
"step": 145335
|
| 284274 |
+
},
|
| 284275 |
+
{
|
| 284276 |
+
"epoch": 1171.9,
|
| 284277 |
+
"learning_rate": 7.669095315024234e-06,
|
| 284278 |
+
"loss": 0.3173,
|
| 284279 |
+
"step": 145340
|
| 284280 |
+
},
|
| 284281 |
+
{
|
| 284282 |
+
"epoch": 1171.94,
|
| 284283 |
+
"learning_rate": 7.66901453957997e-06,
|
| 284284 |
+
"loss": 0.3376,
|
| 284285 |
+
"step": 145345
|
| 284286 |
+
},
|
| 284287 |
+
{
|
| 284288 |
+
"epoch": 1171.98,
|
| 284289 |
+
"learning_rate": 7.668933764135704e-06,
|
| 284290 |
+
"loss": 0.6947,
|
| 284291 |
+
"step": 145350
|
| 284292 |
+
},
|
| 284293 |
+
{
|
| 284294 |
+
"epoch": 1172.0,
|
| 284295 |
+
"eval_loss": 0.47769203782081604,
|
| 284296 |
+
"eval_runtime": 39.6454,
|
| 284297 |
+
"eval_samples_per_second": 21.112,
|
| 284298 |
+
"eval_steps_per_second": 0.681,
|
| 284299 |
+
"eval_wer": 0.18728573557907288,
|
| 284300 |
+
"step": 145352
|
| 284301 |
+
},
|
| 284302 |
+
{
|
| 284303 |
+
"epoch": 1162.02,
|
| 284304 |
+
"learning_rate": 7.66885298869144e-06,
|
| 284305 |
+
"loss": 0.3284,
|
| 284306 |
+
"step": 145355
|
| 284307 |
+
},
|
| 284308 |
+
{
|
| 284309 |
+
"epoch": 1162.06,
|
| 284310 |
+
"learning_rate": 7.668772213247174e-06,
|
| 284311 |
+
"loss": 0.2665,
|
| 284312 |
+
"step": 145360
|
| 284313 |
+
},
|
| 284314 |
+
{
|
| 284315 |
+
"epoch": 1162.1,
|
| 284316 |
+
"learning_rate": 7.66869143780291e-06,
|
| 284317 |
+
"loss": 0.266,
|
| 284318 |
+
"step": 145365
|
| 284319 |
+
},
|
| 284320 |
+
{
|
| 284321 |
+
"epoch": 1162.14,
|
| 284322 |
+
"learning_rate": 7.668610662358644e-06,
|
| 284323 |
+
"loss": 0.3346,
|
| 284324 |
+
"step": 145370
|
| 284325 |
+
},
|
| 284326 |
+
{
|
| 284327 |
+
"epoch": 1162.18,
|
| 284328 |
+
"learning_rate": 7.66852988691438e-06,
|
| 284329 |
+
"loss": 0.6964,
|
| 284330 |
+
"step": 145375
|
| 284331 |
+
},
|
| 284332 |
+
{
|
| 284333 |
+
"epoch": 1162.22,
|
| 284334 |
+
"learning_rate": 7.668449111470114e-06,
|
| 284335 |
+
"loss": 0.8126,
|
| 284336 |
+
"step": 145380
|
| 284337 |
+
},
|
| 284338 |
+
{
|
| 284339 |
+
"epoch": 1162.26,
|
| 284340 |
+
"learning_rate": 7.66836833602585e-06,
|
| 284341 |
+
"loss": 0.2484,
|
| 284342 |
+
"step": 145385
|
| 284343 |
+
},
|
| 284344 |
+
{
|
| 284345 |
+
"epoch": 1162.3,
|
| 284346 |
+
"learning_rate": 7.668287560581584e-06,
|
| 284347 |
+
"loss": 0.2748,
|
| 284348 |
+
"step": 145390
|
| 284349 |
+
},
|
| 284350 |
+
{
|
| 284351 |
+
"epoch": 1162.34,
|
| 284352 |
+
"learning_rate": 7.66820678513732e-06,
|
| 284353 |
+
"loss": 0.3347,
|
| 284354 |
+
"step": 145395
|
| 284355 |
+
},
|
| 284356 |
+
{
|
| 284357 |
+
"epoch": 1162.38,
|
| 284358 |
+
"learning_rate": 7.668126009693053e-06,
|
| 284359 |
+
"loss": 0.7334,
|
| 284360 |
+
"step": 145400
|
| 284361 |
+
},
|
| 284362 |
+
{
|
| 284363 |
+
"epoch": 1162.42,
|
| 284364 |
+
"learning_rate": 7.66804523424879e-06,
|
| 284365 |
+
"loss": 0.7622,
|
| 284366 |
+
"step": 145405
|
| 284367 |
+
},
|
| 284368 |
+
{
|
| 284369 |
+
"epoch": 1162.46,
|
| 284370 |
+
"learning_rate": 7.667964458804525e-06,
|
| 284371 |
+
"loss": 0.2724,
|
| 284372 |
+
"step": 145410
|
| 284373 |
+
},
|
| 284374 |
+
{
|
| 284375 |
+
"epoch": 1162.5,
|
| 284376 |
+
"learning_rate": 7.66788368336026e-06,
|
| 284377 |
+
"loss": 0.2608,
|
| 284378 |
+
"step": 145415
|
| 284379 |
+
},
|
| 284380 |
+
{
|
| 284381 |
+
"epoch": 1162.54,
|
| 284382 |
+
"learning_rate": 7.667802907915995e-06,
|
| 284383 |
+
"loss": 0.3223,
|
| 284384 |
+
"step": 145420
|
| 284385 |
+
},
|
| 284386 |
+
{
|
| 284387 |
+
"epoch": 1162.58,
|
| 284388 |
+
"learning_rate": 7.66772213247173e-06,
|
| 284389 |
+
"loss": 0.6718,
|
| 284390 |
+
"step": 145425
|
| 284391 |
+
},
|
| 284392 |
+
{
|
| 284393 |
+
"epoch": 1162.62,
|
| 284394 |
+
"learning_rate": 7.667641357027465e-06,
|
| 284395 |
+
"loss": 0.8369,
|
| 284396 |
+
"step": 145430
|
| 284397 |
+
},
|
| 284398 |
+
{
|
| 284399 |
+
"epoch": 1162.66,
|
| 284400 |
+
"learning_rate": 7.6675605815832e-06,
|
| 284401 |
+
"loss": 0.2615,
|
| 284402 |
+
"step": 145435
|
| 284403 |
+
},
|
| 284404 |
+
{
|
| 284405 |
+
"epoch": 1162.7,
|
| 284406 |
+
"learning_rate": 7.667479806138935e-06,
|
| 284407 |
+
"loss": 0.2988,
|
| 284408 |
+
"step": 145440
|
| 284409 |
+
},
|
| 284410 |
+
{
|
| 284411 |
+
"epoch": 1162.74,
|
| 284412 |
+
"learning_rate": 7.66739903069467e-06,
|
| 284413 |
+
"loss": 0.3295,
|
| 284414 |
+
"step": 145445
|
| 284415 |
+
},
|
| 284416 |
+
{
|
| 284417 |
+
"epoch": 1162.78,
|
| 284418 |
+
"learning_rate": 7.667318255250405e-06,
|
| 284419 |
+
"loss": 0.7126,
|
| 284420 |
+
"step": 145450
|
| 284421 |
+
},
|
| 284422 |
+
{
|
| 284423 |
+
"epoch": 1162.82,
|
| 284424 |
+
"learning_rate": 7.66723747980614e-06,
|
| 284425 |
+
"loss": 0.8154,
|
| 284426 |
+
"step": 145455
|
| 284427 |
+
},
|
| 284428 |
+
{
|
| 284429 |
+
"epoch": 1162.86,
|
| 284430 |
+
"learning_rate": 7.667156704361875e-06,
|
| 284431 |
+
"loss": 0.2407,
|
| 284432 |
+
"step": 145460
|
| 284433 |
+
},
|
| 284434 |
+
{
|
| 284435 |
+
"epoch": 1162.9,
|
| 284436 |
+
"learning_rate": 7.66707592891761e-06,
|
| 284437 |
+
"loss": 0.2759,
|
| 284438 |
+
"step": 145465
|
| 284439 |
+
},
|
| 284440 |
+
{
|
| 284441 |
+
"epoch": 1162.94,
|
| 284442 |
+
"learning_rate": 7.666995153473345e-06,
|
| 284443 |
+
"loss": 0.2949,
|
| 284444 |
+
"step": 145470
|
| 284445 |
+
},
|
| 284446 |
+
{
|
| 284447 |
+
"epoch": 1162.98,
|
| 284448 |
+
"learning_rate": 7.66691437802908e-06,
|
| 284449 |
+
"loss": 0.7172,
|
| 284450 |
+
"step": 145475
|
| 284451 |
+
},
|
| 284452 |
+
{
|
| 284453 |
+
"epoch": 1163.0,
|
| 284454 |
+
"eval_loss": 0.3563687801361084,
|
| 284455 |
+
"eval_runtime": 41.8928,
|
| 284456 |
+
"eval_samples_per_second": 19.98,
|
| 284457 |
+
"eval_steps_per_second": 0.645,
|
| 284458 |
+
"eval_wer": 0.17516339869281045,
|
| 284459 |
+
"step": 145477
|
| 284460 |
+
},
|
| 284461 |
+
{
|
| 284462 |
+
"epoch": 1163.02,
|
| 284463 |
+
"learning_rate": 7.666833602584815e-06,
|
| 284464 |
+
"loss": 0.3148,
|
| 284465 |
+
"step": 145480
|
| 284466 |
+
},
|
| 284467 |
+
{
|
| 284468 |
+
"epoch": 1163.06,
|
| 284469 |
+
"learning_rate": 7.66675282714055e-06,
|
| 284470 |
+
"loss": 0.2606,
|
| 284471 |
+
"step": 145485
|
| 284472 |
+
},
|
| 284473 |
+
{
|
| 284474 |
+
"epoch": 1163.1,
|
| 284475 |
+
"learning_rate": 7.666672051696285e-06,
|
| 284476 |
+
"loss": 0.3123,
|
| 284477 |
+
"step": 145490
|
| 284478 |
+
},
|
| 284479 |
+
{
|
| 284480 |
+
"epoch": 1163.14,
|
| 284481 |
+
"learning_rate": 7.66659127625202e-06,
|
| 284482 |
+
"loss": 0.3515,
|
| 284483 |
+
"step": 145495
|
| 284484 |
+
},
|
| 284485 |
+
{
|
| 284486 |
+
"epoch": 1163.18,
|
| 284487 |
+
"learning_rate": 7.666510500807755e-06,
|
| 284488 |
+
"loss": 0.7378,
|
| 284489 |
+
"step": 145500
|
| 284490 |
+
},
|
| 284491 |
+
{
|
| 284492 |
+
"epoch": 1163.22,
|
| 284493 |
+
"learning_rate": 7.66642972536349e-06,
|
| 284494 |
+
"loss": 0.8376,
|
| 284495 |
+
"step": 145505
|
| 284496 |
+
},
|
| 284497 |
+
{
|
| 284498 |
+
"epoch": 1163.26,
|
| 284499 |
+
"learning_rate": 7.666348949919225e-06,
|
| 284500 |
+
"loss": 0.3011,
|
| 284501 |
+
"step": 145510
|
| 284502 |
+
},
|
| 284503 |
+
{
|
| 284504 |
+
"epoch": 1163.3,
|
| 284505 |
+
"learning_rate": 7.66626817447496e-06,
|
| 284506 |
+
"loss": 0.3181,
|
| 284507 |
+
"step": 145515
|
| 284508 |
+
},
|
| 284509 |
+
{
|
| 284510 |
+
"epoch": 1163.34,
|
| 284511 |
+
"learning_rate": 7.666187399030695e-06,
|
| 284512 |
+
"loss": 0.4141,
|
| 284513 |
+
"step": 145520
|
| 284514 |
+
},
|
| 284515 |
+
{
|
| 284516 |
+
"epoch": 1163.38,
|
| 284517 |
+
"learning_rate": 7.66610662358643e-06,
|
| 284518 |
+
"loss": 0.7503,
|
| 284519 |
+
"step": 145525
|
| 284520 |
+
},
|
| 284521 |
+
{
|
| 284522 |
+
"epoch": 1163.42,
|
| 284523 |
+
"learning_rate": 7.666025848142165e-06,
|
| 284524 |
+
"loss": 0.8573,
|
| 284525 |
+
"step": 145530
|
| 284526 |
+
},
|
| 284527 |
+
{
|
| 284528 |
+
"epoch": 1163.46,
|
| 284529 |
+
"learning_rate": 7.6659450726979e-06,
|
| 284530 |
+
"loss": 0.283,
|
| 284531 |
+
"step": 145535
|
| 284532 |
+
},
|
| 284533 |
+
{
|
| 284534 |
+
"epoch": 1163.5,
|
| 284535 |
+
"learning_rate": 7.665864297253635e-06,
|
| 284536 |
+
"loss": 0.2771,
|
| 284537 |
+
"step": 145540
|
| 284538 |
+
},
|
| 284539 |
+
{
|
| 284540 |
+
"epoch": 1163.54,
|
| 284541 |
+
"learning_rate": 7.66578352180937e-06,
|
| 284542 |
+
"loss": 0.5021,
|
| 284543 |
+
"step": 145545
|
| 284544 |
+
},
|
| 284545 |
+
{
|
| 284546 |
+
"epoch": 1163.58,
|
| 284547 |
+
"learning_rate": 7.665702746365107e-06,
|
| 284548 |
+
"loss": 0.7615,
|
| 284549 |
+
"step": 145550
|
| 284550 |
+
},
|
| 284551 |
+
{
|
| 284552 |
+
"epoch": 1163.62,
|
| 284553 |
+
"learning_rate": 7.66562197092084e-06,
|
| 284554 |
+
"loss": 0.8149,
|
| 284555 |
+
"step": 145555
|
| 284556 |
+
},
|
| 284557 |
+
{
|
| 284558 |
+
"epoch": 1163.66,
|
| 284559 |
+
"learning_rate": 7.665541195476576e-06,
|
| 284560 |
+
"loss": 0.2333,
|
| 284561 |
+
"step": 145560
|
| 284562 |
+
},
|
| 284563 |
+
{
|
| 284564 |
+
"epoch": 1163.7,
|
| 284565 |
+
"learning_rate": 7.66546042003231e-06,
|
| 284566 |
+
"loss": 0.293,
|
| 284567 |
+
"step": 145565
|
| 284568 |
+
},
|
| 284569 |
+
{
|
| 284570 |
+
"epoch": 1163.74,
|
| 284571 |
+
"learning_rate": 7.665379644588046e-06,
|
| 284572 |
+
"loss": 0.3133,
|
| 284573 |
+
"step": 145570
|
| 284574 |
+
},
|
| 284575 |
+
{
|
| 284576 |
+
"epoch": 1163.78,
|
| 284577 |
+
"learning_rate": 7.66529886914378e-06,
|
| 284578 |
+
"loss": 0.693,
|
| 284579 |
+
"step": 145575
|
| 284580 |
+
},
|
| 284581 |
+
{
|
| 284582 |
+
"epoch": 1163.82,
|
| 284583 |
+
"learning_rate": 7.665218093699516e-06,
|
| 284584 |
+
"loss": 0.9545,
|
| 284585 |
+
"step": 145580
|
| 284586 |
+
},
|
| 284587 |
+
{
|
| 284588 |
+
"epoch": 1163.86,
|
| 284589 |
+
"learning_rate": 7.66513731825525e-06,
|
| 284590 |
+
"loss": 0.3008,
|
| 284591 |
+
"step": 145585
|
| 284592 |
+
},
|
| 284593 |
+
{
|
| 284594 |
+
"epoch": 1163.9,
|
| 284595 |
+
"learning_rate": 7.665056542810986e-06,
|
| 284596 |
+
"loss": 0.3013,
|
| 284597 |
+
"step": 145590
|
| 284598 |
+
},
|
| 284599 |
+
{
|
| 284600 |
+
"epoch": 1163.94,
|
| 284601 |
+
"learning_rate": 7.66497576736672e-06,
|
| 284602 |
+
"loss": 0.538,
|
| 284603 |
+
"step": 145595
|
| 284604 |
+
},
|
| 284605 |
+
{
|
| 284606 |
+
"epoch": 1163.98,
|
| 284607 |
+
"learning_rate": 7.664894991922456e-06,
|
| 284608 |
+
"loss": 0.7369,
|
| 284609 |
+
"step": 145600
|
| 284610 |
+
},
|
| 284611 |
+
{
|
| 284612 |
+
"epoch": 1164.0,
|
| 284613 |
+
"eval_loss": 0.3466392159461975,
|
| 284614 |
+
"eval_runtime": 39.3229,
|
| 284615 |
+
"eval_samples_per_second": 21.285,
|
| 284616 |
+
"eval_steps_per_second": 0.687,
|
| 284617 |
+
"eval_wer": 0.16716790141568252,
|
| 284618 |
+
"step": 145602
|
| 284619 |
+
},
|
| 284620 |
+
{
|
| 284621 |
+
"epoch": 1164.02,
|
| 284622 |
+
"learning_rate": 7.66481421647819e-06,
|
| 284623 |
+
"loss": 0.3273,
|
| 284624 |
+
"step": 145605
|
| 284625 |
+
},
|
| 284626 |
+
{
|
| 284627 |
+
"epoch": 1164.06,
|
| 284628 |
+
"learning_rate": 7.664733441033926e-06,
|
| 284629 |
+
"loss": 0.2867,
|
| 284630 |
+
"step": 145610
|
| 284631 |
+
},
|
| 284632 |
+
{
|
| 284633 |
+
"epoch": 1164.1,
|
| 284634 |
+
"learning_rate": 7.664652665589662e-06,
|
| 284635 |
+
"loss": 0.3066,
|
| 284636 |
+
"step": 145615
|
| 284637 |
+
},
|
| 284638 |
+
{
|
| 284639 |
+
"epoch": 1164.14,
|
| 284640 |
+
"learning_rate": 7.664571890145396e-06,
|
| 284641 |
+
"loss": 0.3405,
|
| 284642 |
+
"step": 145620
|
| 284643 |
+
},
|
| 284644 |
+
{
|
| 284645 |
+
"epoch": 1164.18,
|
| 284646 |
+
"learning_rate": 7.664491114701132e-06,
|
| 284647 |
+
"loss": 0.7475,
|
| 284648 |
+
"step": 145625
|
| 284649 |
+
},
|
| 284650 |
+
{
|
| 284651 |
+
"epoch": 1164.22,
|
| 284652 |
+
"learning_rate": 7.664410339256866e-06,
|
| 284653 |
+
"loss": 0.9464,
|
| 284654 |
+
"step": 145630
|
| 284655 |
+
},
|
| 284656 |
+
{
|
| 284657 |
+
"epoch": 1164.26,
|
| 284658 |
+
"learning_rate": 7.664329563812602e-06,
|
| 284659 |
+
"loss": 0.3219,
|
| 284660 |
+
"step": 145635
|
| 284661 |
+
},
|
| 284662 |
+
{
|
| 284663 |
+
"epoch": 1164.3,
|
| 284664 |
+
"learning_rate": 7.664248788368336e-06,
|
| 284665 |
+
"loss": 0.3095,
|
| 284666 |
+
"step": 145640
|
| 284667 |
+
},
|
| 284668 |
+
{
|
| 284669 |
+
"epoch": 1164.34,
|
| 284670 |
+
"learning_rate": 7.664168012924072e-06,
|
| 284671 |
+
"loss": 0.3189,
|
| 284672 |
+
"step": 145645
|
| 284673 |
+
},
|
| 284674 |
+
{
|
| 284675 |
+
"epoch": 1164.38,
|
| 284676 |
+
"learning_rate": 7.664087237479806e-06,
|
| 284677 |
+
"loss": 0.7207,
|
| 284678 |
+
"step": 145650
|
| 284679 |
+
},
|
| 284680 |
+
{
|
| 284681 |
+
"epoch": 1164.42,
|
| 284682 |
+
"learning_rate": 7.664006462035542e-06,
|
| 284683 |
+
"loss": 0.8142,
|
| 284684 |
+
"step": 145655
|
| 284685 |
+
},
|
| 284686 |
+
{
|
| 284687 |
+
"epoch": 1164.46,
|
| 284688 |
+
"learning_rate": 7.663925686591276e-06,
|
| 284689 |
+
"loss": 0.2481,
|
| 284690 |
+
"step": 145660
|
| 284691 |
+
},
|
| 284692 |
+
{
|
| 284693 |
+
"epoch": 1164.5,
|
| 284694 |
+
"learning_rate": 7.663844911147012e-06,
|
| 284695 |
+
"loss": 0.2636,
|
| 284696 |
+
"step": 145665
|
| 284697 |
+
},
|
| 284698 |
+
{
|
| 284699 |
+
"epoch": 1164.54,
|
| 284700 |
+
"learning_rate": 7.663764135702746e-06,
|
| 284701 |
+
"loss": 0.3821,
|
| 284702 |
+
"step": 145670
|
| 284703 |
+
},
|
| 284704 |
+
{
|
| 284705 |
+
"epoch": 1164.58,
|
| 284706 |
+
"learning_rate": 7.663683360258482e-06,
|
| 284707 |
+
"loss": 0.7099,
|
| 284708 |
+
"step": 145675
|
| 284709 |
+
},
|
| 284710 |
+
{
|
| 284711 |
+
"epoch": 1164.62,
|
| 284712 |
+
"learning_rate": 7.663602584814218e-06,
|
| 284713 |
+
"loss": 0.7693,
|
| 284714 |
+
"step": 145680
|
| 284715 |
+
},
|
| 284716 |
+
{
|
| 284717 |
+
"epoch": 1164.66,
|
| 284718 |
+
"learning_rate": 7.663521809369952e-06,
|
| 284719 |
+
"loss": 0.3184,
|
| 284720 |
+
"step": 145685
|
| 284721 |
+
},
|
| 284722 |
+
{
|
| 284723 |
+
"epoch": 1164.7,
|
| 284724 |
+
"learning_rate": 7.663441033925688e-06,
|
| 284725 |
+
"loss": 0.2638,
|
| 284726 |
+
"step": 145690
|
| 284727 |
+
},
|
| 284728 |
+
{
|
| 284729 |
+
"epoch": 1164.74,
|
| 284730 |
+
"learning_rate": 7.663360258481422e-06,
|
| 284731 |
+
"loss": 0.4138,
|
| 284732 |
+
"step": 145695
|
| 284733 |
+
},
|
| 284734 |
+
{
|
| 284735 |
+
"epoch": 1164.78,
|
| 284736 |
+
"learning_rate": 7.663279483037158e-06,
|
| 284737 |
+
"loss": 0.7006,
|
| 284738 |
+
"step": 145700
|
| 284739 |
+
},
|
| 284740 |
+
{
|
| 284741 |
+
"epoch": 1164.82,
|
| 284742 |
+
"learning_rate": 7.663198707592892e-06,
|
| 284743 |
+
"loss": 0.9005,
|
| 284744 |
+
"step": 145705
|
| 284745 |
+
},
|
| 284746 |
+
{
|
| 284747 |
+
"epoch": 1164.86,
|
| 284748 |
+
"learning_rate": 7.663117932148628e-06,
|
| 284749 |
+
"loss": 0.2489,
|
| 284750 |
+
"step": 145710
|
| 284751 |
+
},
|
| 284752 |
+
{
|
| 284753 |
+
"epoch": 1164.9,
|
| 284754 |
+
"learning_rate": 7.663037156704362e-06,
|
| 284755 |
+
"loss": 0.2708,
|
| 284756 |
+
"step": 145715
|
| 284757 |
+
},
|
| 284758 |
+
{
|
| 284759 |
+
"epoch": 1164.94,
|
| 284760 |
+
"learning_rate": 7.662956381260098e-06,
|
| 284761 |
+
"loss": 0.3742,
|
| 284762 |
+
"step": 145720
|
| 284763 |
+
},
|
| 284764 |
+
{
|
| 284765 |
+
"epoch": 1164.98,
|
| 284766 |
+
"learning_rate": 7.662875605815832e-06,
|
| 284767 |
+
"loss": 0.7858,
|
| 284768 |
+
"step": 145725
|
| 284769 |
+
},
|
| 284770 |
+
{
|
| 284771 |
+
"epoch": 1165.0,
|
| 284772 |
+
"eval_loss": 0.38368138670921326,
|
| 284773 |
+
"eval_runtime": 40.6993,
|
| 284774 |
+
"eval_samples_per_second": 20.565,
|
| 284775 |
+
"eval_steps_per_second": 0.663,
|
| 284776 |
+
"eval_wer": 0.1798013485101138,
|
| 284777 |
+
"step": 145727
|
| 284778 |
+
},
|
| 284779 |
+
{
|
| 284780 |
+
"epoch": 1165.02,
|
| 284781 |
+
"learning_rate": 7.662794830371568e-06,
|
| 284782 |
+
"loss": 0.4956,
|
| 284783 |
+
"step": 145730
|
| 284784 |
+
},
|
| 284785 |
+
{
|
| 284786 |
+
"epoch": 1165.06,
|
| 284787 |
+
"learning_rate": 7.662714054927302e-06,
|
| 284788 |
+
"loss": 0.2444,
|
| 284789 |
+
"step": 145735
|
| 284790 |
+
},
|
| 284791 |
+
{
|
| 284792 |
+
"epoch": 1165.1,
|
| 284793 |
+
"learning_rate": 7.662633279483038e-06,
|
| 284794 |
+
"loss": 0.2835,
|
| 284795 |
+
"step": 145740
|
| 284796 |
+
},
|
| 284797 |
+
{
|
| 284798 |
+
"epoch": 1165.14,
|
| 284799 |
+
"learning_rate": 7.662552504038772e-06,
|
| 284800 |
+
"loss": 0.4172,
|
| 284801 |
+
"step": 145745
|
| 284802 |
+
},
|
| 284803 |
+
{
|
| 284804 |
+
"epoch": 1165.18,
|
| 284805 |
+
"learning_rate": 7.662471728594508e-06,
|
| 284806 |
+
"loss": 0.656,
|
| 284807 |
+
"step": 145750
|
| 284808 |
+
},
|
| 284809 |
+
{
|
| 284810 |
+
"epoch": 1165.22,
|
| 284811 |
+
"learning_rate": 7.662390953150244e-06,
|
| 284812 |
+
"loss": 0.8101,
|
| 284813 |
+
"step": 145755
|
| 284814 |
+
},
|
| 284815 |
+
{
|
| 284816 |
+
"epoch": 1165.26,
|
| 284817 |
+
"learning_rate": 7.662310177705978e-06,
|
| 284818 |
+
"loss": 0.2991,
|
| 284819 |
+
"step": 145760
|
| 284820 |
+
},
|
| 284821 |
+
{
|
| 284822 |
+
"epoch": 1165.3,
|
| 284823 |
+
"learning_rate": 7.662229402261714e-06,
|
| 284824 |
+
"loss": 0.2851,
|
| 284825 |
+
"step": 145765
|
| 284826 |
+
},
|
| 284827 |
+
{
|
| 284828 |
+
"epoch": 1165.34,
|
| 284829 |
+
"learning_rate": 7.662148626817448e-06,
|
| 284830 |
+
"loss": 0.3243,
|
| 284831 |
+
"step": 145770
|
| 284832 |
+
},
|
| 284833 |
+
{
|
| 284834 |
+
"epoch": 1165.38,
|
| 284835 |
+
"learning_rate": 7.662067851373183e-06,
|
| 284836 |
+
"loss": 0.7754,
|
| 284837 |
+
"step": 145775
|
| 284838 |
+
},
|
| 284839 |
+
{
|
| 284840 |
+
"epoch": 1165.42,
|
| 284841 |
+
"learning_rate": 7.661987075928918e-06,
|
| 284842 |
+
"loss": 0.9141,
|
| 284843 |
+
"step": 145780
|
| 284844 |
+
},
|
| 284845 |
+
{
|
| 284846 |
+
"epoch": 1165.46,
|
| 284847 |
+
"learning_rate": 7.661906300484653e-06,
|
| 284848 |
+
"loss": 0.2886,
|
| 284849 |
+
"step": 145785
|
| 284850 |
+
},
|
| 284851 |
+
{
|
| 284852 |
+
"epoch": 1165.5,
|
| 284853 |
+
"learning_rate": 7.661825525040388e-06,
|
| 284854 |
+
"loss": 0.305,
|
| 284855 |
+
"step": 145790
|
| 284856 |
+
},
|
| 284857 |
+
{
|
| 284858 |
+
"epoch": 1165.54,
|
| 284859 |
+
"learning_rate": 7.661744749596123e-06,
|
| 284860 |
+
"loss": 0.5382,
|
| 284861 |
+
"step": 145795
|
| 284862 |
+
},
|
| 284863 |
+
{
|
| 284864 |
+
"epoch": 1165.58,
|
| 284865 |
+
"learning_rate": 7.661663974151858e-06,
|
| 284866 |
+
"loss": 0.7651,
|
| 284867 |
+
"step": 145800
|
| 284868 |
+
},
|
| 284869 |
+
{
|
| 284870 |
+
"epoch": 1165.62,
|
| 284871 |
+
"learning_rate": 7.661583198707593e-06,
|
| 284872 |
+
"loss": 0.8388,
|
| 284873 |
+
"step": 145805
|
| 284874 |
+
},
|
| 284875 |
+
{
|
| 284876 |
+
"epoch": 1165.66,
|
| 284877 |
+
"learning_rate": 7.661502423263328e-06,
|
| 284878 |
+
"loss": 0.2597,
|
| 284879 |
+
"step": 145810
|
| 284880 |
+
},
|
| 284881 |
+
{
|
| 284882 |
+
"epoch": 1165.7,
|
| 284883 |
+
"learning_rate": 7.661421647819063e-06,
|
| 284884 |
+
"loss": 0.3229,
|
| 284885 |
+
"step": 145815
|
| 284886 |
+
},
|
| 284887 |
+
{
|
| 284888 |
+
"epoch": 1165.74,
|
| 284889 |
+
"learning_rate": 7.6613408723748e-06,
|
| 284890 |
+
"loss": 0.376,
|
| 284891 |
+
"step": 145820
|
| 284892 |
+
},
|
| 284893 |
+
{
|
| 284894 |
+
"epoch": 1165.78,
|
| 284895 |
+
"learning_rate": 7.661260096930533e-06,
|
| 284896 |
+
"loss": 0.6114,
|
| 284897 |
+
"step": 145825
|
| 284898 |
+
},
|
| 284899 |
+
{
|
| 284900 |
+
"epoch": 1165.82,
|
| 284901 |
+
"learning_rate": 7.66117932148627e-06,
|
| 284902 |
+
"loss": 0.8276,
|
| 284903 |
+
"step": 145830
|
| 284904 |
+
},
|
| 284905 |
+
{
|
| 284906 |
+
"epoch": 1165.86,
|
| 284907 |
+
"learning_rate": 7.661098546042003e-06,
|
| 284908 |
+
"loss": 0.2959,
|
| 284909 |
+
"step": 145835
|
| 284910 |
+
},
|
| 284911 |
+
{
|
| 284912 |
+
"epoch": 1165.9,
|
| 284913 |
+
"learning_rate": 7.66101777059774e-06,
|
| 284914 |
+
"loss": 0.2697,
|
| 284915 |
+
"step": 145840
|
| 284916 |
+
},
|
| 284917 |
+
{
|
| 284918 |
+
"epoch": 1165.94,
|
| 284919 |
+
"learning_rate": 7.660936995153473e-06,
|
| 284920 |
+
"loss": 0.3293,
|
| 284921 |
+
"step": 145845
|
| 284922 |
+
},
|
| 284923 |
+
{
|
| 284924 |
+
"epoch": 1165.98,
|
| 284925 |
+
"learning_rate": 7.660856219709209e-06,
|
| 284926 |
+
"loss": 0.7433,
|
| 284927 |
+
"step": 145850
|
| 284928 |
+
},
|
| 284929 |
+
{
|
| 284930 |
+
"epoch": 1166.0,
|
| 284931 |
+
"eval_loss": 0.33688074350357056,
|
| 284932 |
+
"eval_runtime": 41.2402,
|
| 284933 |
+
"eval_samples_per_second": 20.296,
|
| 284934 |
+
"eval_steps_per_second": 0.655,
|
| 284935 |
+
"eval_wer": 0.17012599228024178,
|
| 284936 |
+
"step": 145852
|
| 284937 |
}
|
| 284938 |
],
|
| 284939 |
+
"max_steps": 625000,
|
| 284940 |
"num_train_epochs": 5000,
|
| 284941 |
+
"total_flos": 4.104335476145737e+20,
|
| 284942 |
"trial_name": null,
|
| 284943 |
"trial_params": null
|
| 284944 |
}
|
model-bin/finetune/base/{checkpoint-145228 β checkpoint-145852}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1630226071.8984504/events.out.tfevents.1630226071.cc93b136ebf5.1086.99
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:950fda5fb6891bedbe19f04d898e4337085688846519f0cbf7e32d1f53a30bd7
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630226502.2753584/events.out.tfevents.1630226502.cc93b136ebf5.1086.101
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:81e8a348434d43107b4bd23687b9036740d2656d44a85f95e0d48e8300c8960c
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630226942.7046182/events.out.tfevents.1630226942.cc93b136ebf5.1086.103
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:65f84bc91cf2166792868abf376269402376f96ba7f36b17d1de5093ece4e578
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630227371.5174663/events.out.tfevents.1630227371.cc93b136ebf5.1086.105
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8138ddf57ae87fdc1b2496a5e815e82551c7bcac688069fc777167f16722bdba
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630227801.075265/events.out.tfevents.1630227801.cc93b136ebf5.1086.107
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b0c2ab0f952d96d77872c519ec99fddd0963e1be7c80ec37907c5e6271933548
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630226071.cc93b136ebf5.1086.98
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:831aa23a57f57c612b30af2c93fce20093b1b6a90cdf699d02c0d9b4538e0398
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630226502.cc93b136ebf5.1086.100
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5adf3892020b461c62a7cb3d13d02a2fe11fd02bb7189c650d98eefd2a55cf6e
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630226942.cc93b136ebf5.1086.102
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:04f87e0868a5a3a52b05bfbd849b69ed1c47f4407e5bc528eef400e93aac58c5
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630227371.cc93b136ebf5.1086.104
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1fd1780dc42c525f366f6d19f3c687da0f1f65bce9aa18df4267de2cfeec52a0
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630227801.cc93b136ebf5.1086.106
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:14440456e8452b5d272fd86d57e8bd92f9707975bda22c4536cf19dadc451de5
|
| 3 |
+
size 8622
|