"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/trainer_state.json +639 -3
- model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630080962.0011778/events.out.tfevents.1630080962.dea8124df033.1100.1 +3 -0
- model-bin/finetune/base/log/1630081546.0007536/events.out.tfevents.1630081546.dea8124df033.1391.1 +3 -0
- model-bin/finetune/base/log/1630081975.5740092/events.out.tfevents.1630081975.dea8124df033.1391.3 +3 -0
- model-bin/finetune/base/log/1630082400.0108204/events.out.tfevents.1630082400.dea8124df033.1391.5 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630080961.dea8124df033.1100.0 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630081545.dea8124df033.1391.0 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630081975.dea8124df033.1391.2 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630082400.dea8124df033.1391.4 +3 -0
model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8f3ce461a41ebb478a3033a35cb4191908dec25dc3b09f575609fc6aa146dd60
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ed4c4bc6da00d916aba5b2523194212a682bc5e8e07fee75fc5d3fcb170a87ce
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e8135104c63b5165310558e47d8ba777edf4c3b459df9a523c789c28dcc1e549
|
| 3 |
size 14503
|
model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:531a84460dc36ad58f317625a8b2df924c497e322bd18a71440cc0d51bac941b
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:57fccb27f6be770c60514fd2f134fe1b6e0a8965992c964c9bd1421740a9ae29
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1743826049391605,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -242184,11 +242184,647 @@
|
|
| 242184 |
"eval_steps_per_second": 0.643,
|
| 242185 |
"eval_wer": 0.18449336594766882,
|
| 242186 |
"step": 112251
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 242187 |
}
|
| 242188 |
],
|
| 242189 |
"max_steps": 625000,
|
| 242190 |
"num_train_epochs": 5000,
|
| 242191 |
-
"total_flos": 3.
|
| 242192 |
"trial_name": null,
|
| 242193 |
"trial_params": null
|
| 242194 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1743826049391605,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-101551",
|
| 4 |
+
"epoch": 902.0,
|
| 5 |
+
"global_step": 112750,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 242184 |
"eval_steps_per_second": 0.643,
|
| 242185 |
"eval_wer": 0.18449336594766882,
|
| 242186 |
"step": 112251
|
| 242187 |
+
},
|
| 242188 |
+
{
|
| 242189 |
+
"epoch": 898.03,
|
| 242190 |
+
"learning_rate": 8.217772435897437e-06,
|
| 242191 |
+
"loss": 0.302,
|
| 242192 |
+
"step": 112255
|
| 242193 |
+
},
|
| 242194 |
+
{
|
| 242195 |
+
"epoch": 898.07,
|
| 242196 |
+
"learning_rate": 8.217692307692308e-06,
|
| 242197 |
+
"loss": 0.2726,
|
| 242198 |
+
"step": 112260
|
| 242199 |
+
},
|
| 242200 |
+
{
|
| 242201 |
+
"epoch": 898.11,
|
| 242202 |
+
"learning_rate": 8.21761217948718e-06,
|
| 242203 |
+
"loss": 0.3286,
|
| 242204 |
+
"step": 112265
|
| 242205 |
+
},
|
| 242206 |
+
{
|
| 242207 |
+
"epoch": 898.15,
|
| 242208 |
+
"learning_rate": 8.217532051282053e-06,
|
| 242209 |
+
"loss": 0.4197,
|
| 242210 |
+
"step": 112270
|
| 242211 |
+
},
|
| 242212 |
+
{
|
| 242213 |
+
"epoch": 898.19,
|
| 242214 |
+
"learning_rate": 8.217451923076924e-06,
|
| 242215 |
+
"loss": 0.8389,
|
| 242216 |
+
"step": 112275
|
| 242217 |
+
},
|
| 242218 |
+
{
|
| 242219 |
+
"epoch": 898.23,
|
| 242220 |
+
"learning_rate": 8.217371794871795e-06,
|
| 242221 |
+
"loss": 0.6037,
|
| 242222 |
+
"step": 112280
|
| 242223 |
+
},
|
| 242224 |
+
{
|
| 242225 |
+
"epoch": 898.27,
|
| 242226 |
+
"learning_rate": 8.217291666666667e-06,
|
| 242227 |
+
"loss": 0.2566,
|
| 242228 |
+
"step": 112285
|
| 242229 |
+
},
|
| 242230 |
+
{
|
| 242231 |
+
"epoch": 898.31,
|
| 242232 |
+
"learning_rate": 8.21721153846154e-06,
|
| 242233 |
+
"loss": 0.3006,
|
| 242234 |
+
"step": 112290
|
| 242235 |
+
},
|
| 242236 |
+
{
|
| 242237 |
+
"epoch": 898.35,
|
| 242238 |
+
"learning_rate": 8.217131410256411e-06,
|
| 242239 |
+
"loss": 0.3725,
|
| 242240 |
+
"step": 112295
|
| 242241 |
+
},
|
| 242242 |
+
{
|
| 242243 |
+
"epoch": 898.39,
|
| 242244 |
+
"learning_rate": 8.217051282051282e-06,
|
| 242245 |
+
"loss": 0.8954,
|
| 242246 |
+
"step": 112300
|
| 242247 |
+
},
|
| 242248 |
+
{
|
| 242249 |
+
"epoch": 898.43,
|
| 242250 |
+
"learning_rate": 8.216971153846156e-06,
|
| 242251 |
+
"loss": 0.7274,
|
| 242252 |
+
"step": 112305
|
| 242253 |
+
},
|
| 242254 |
+
{
|
| 242255 |
+
"epoch": 898.47,
|
| 242256 |
+
"learning_rate": 8.216891025641027e-06,
|
| 242257 |
+
"loss": 0.293,
|
| 242258 |
+
"step": 112310
|
| 242259 |
+
},
|
| 242260 |
+
{
|
| 242261 |
+
"epoch": 898.51,
|
| 242262 |
+
"learning_rate": 8.216810897435898e-06,
|
| 242263 |
+
"loss": 0.2938,
|
| 242264 |
+
"step": 112315
|
| 242265 |
+
},
|
| 242266 |
+
{
|
| 242267 |
+
"epoch": 898.55,
|
| 242268 |
+
"learning_rate": 8.21673076923077e-06,
|
| 242269 |
+
"loss": 0.3959,
|
| 242270 |
+
"step": 112320
|
| 242271 |
+
},
|
| 242272 |
+
{
|
| 242273 |
+
"epoch": 898.59,
|
| 242274 |
+
"learning_rate": 8.216650641025643e-06,
|
| 242275 |
+
"loss": 0.8866,
|
| 242276 |
+
"step": 112325
|
| 242277 |
+
},
|
| 242278 |
+
{
|
| 242279 |
+
"epoch": 898.63,
|
| 242280 |
+
"learning_rate": 8.216570512820512e-06,
|
| 242281 |
+
"loss": 0.8329,
|
| 242282 |
+
"step": 112330
|
| 242283 |
+
},
|
| 242284 |
+
{
|
| 242285 |
+
"epoch": 898.67,
|
| 242286 |
+
"learning_rate": 8.216490384615385e-06,
|
| 242287 |
+
"loss": 0.3571,
|
| 242288 |
+
"step": 112335
|
| 242289 |
+
},
|
| 242290 |
+
{
|
| 242291 |
+
"epoch": 898.71,
|
| 242292 |
+
"learning_rate": 8.216410256410257e-06,
|
| 242293 |
+
"loss": 0.3254,
|
| 242294 |
+
"step": 112340
|
| 242295 |
+
},
|
| 242296 |
+
{
|
| 242297 |
+
"epoch": 898.75,
|
| 242298 |
+
"learning_rate": 8.216330128205128e-06,
|
| 242299 |
+
"loss": 0.3964,
|
| 242300 |
+
"step": 112345
|
| 242301 |
+
},
|
| 242302 |
+
{
|
| 242303 |
+
"epoch": 898.79,
|
| 242304 |
+
"learning_rate": 8.216250000000001e-06,
|
| 242305 |
+
"loss": 1.0306,
|
| 242306 |
+
"step": 112350
|
| 242307 |
+
},
|
| 242308 |
+
{
|
| 242309 |
+
"epoch": 898.83,
|
| 242310 |
+
"learning_rate": 8.216169871794873e-06,
|
| 242311 |
+
"loss": 0.6421,
|
| 242312 |
+
"step": 112355
|
| 242313 |
+
},
|
| 242314 |
+
{
|
| 242315 |
+
"epoch": 898.87,
|
| 242316 |
+
"learning_rate": 8.216089743589744e-06,
|
| 242317 |
+
"loss": 0.3203,
|
| 242318 |
+
"step": 112360
|
| 242319 |
+
},
|
| 242320 |
+
{
|
| 242321 |
+
"epoch": 898.91,
|
| 242322 |
+
"learning_rate": 8.216009615384615e-06,
|
| 242323 |
+
"loss": 0.3004,
|
| 242324 |
+
"step": 112365
|
| 242325 |
+
},
|
| 242326 |
+
{
|
| 242327 |
+
"epoch": 898.95,
|
| 242328 |
+
"learning_rate": 8.215929487179488e-06,
|
| 242329 |
+
"loss": 0.3823,
|
| 242330 |
+
"step": 112370
|
| 242331 |
+
},
|
| 242332 |
+
{
|
| 242333 |
+
"epoch": 898.99,
|
| 242334 |
+
"learning_rate": 8.21584935897436e-06,
|
| 242335 |
+
"loss": 0.8118,
|
| 242336 |
+
"step": 112375
|
| 242337 |
+
},
|
| 242338 |
+
{
|
| 242339 |
+
"epoch": 899.0,
|
| 242340 |
+
"eval_loss": 0.40806370973587036,
|
| 242341 |
+
"eval_runtime": 41.149,
|
| 242342 |
+
"eval_samples_per_second": 20.341,
|
| 242343 |
+
"eval_steps_per_second": 0.656,
|
| 242344 |
+
"eval_wer": 0.18508635386553998,
|
| 242345 |
+
"step": 112376
|
| 242346 |
+
},
|
| 242347 |
+
{
|
| 242348 |
+
"epoch": 906.03,
|
| 242349 |
+
"learning_rate": 8.201357027463652e-06,
|
| 242350 |
+
"loss": 0.3272,
|
| 242351 |
+
"step": 112380
|
| 242352 |
+
},
|
| 242353 |
+
{
|
| 242354 |
+
"epoch": 906.07,
|
| 242355 |
+
"learning_rate": 8.201276252019386e-06,
|
| 242356 |
+
"loss": 0.2657,
|
| 242357 |
+
"step": 112385
|
| 242358 |
+
},
|
| 242359 |
+
{
|
| 242360 |
+
"epoch": 906.11,
|
| 242361 |
+
"learning_rate": 8.201195476575122e-06,
|
| 242362 |
+
"loss": 0.304,
|
| 242363 |
+
"step": 112390
|
| 242364 |
+
},
|
| 242365 |
+
{
|
| 242366 |
+
"epoch": 906.15,
|
| 242367 |
+
"learning_rate": 8.201114701130856e-06,
|
| 242368 |
+
"loss": 0.4485,
|
| 242369 |
+
"step": 112395
|
| 242370 |
+
},
|
| 242371 |
+
{
|
| 242372 |
+
"epoch": 906.19,
|
| 242373 |
+
"learning_rate": 8.201033925686592e-06,
|
| 242374 |
+
"loss": 0.8494,
|
| 242375 |
+
"step": 112400
|
| 242376 |
+
},
|
| 242377 |
+
{
|
| 242378 |
+
"epoch": 906.23,
|
| 242379 |
+
"learning_rate": 8.200953150242326e-06,
|
| 242380 |
+
"loss": 0.5495,
|
| 242381 |
+
"step": 112405
|
| 242382 |
+
},
|
| 242383 |
+
{
|
| 242384 |
+
"epoch": 906.27,
|
| 242385 |
+
"learning_rate": 8.200872374798062e-06,
|
| 242386 |
+
"loss": 0.2863,
|
| 242387 |
+
"step": 112410
|
| 242388 |
+
},
|
| 242389 |
+
{
|
| 242390 |
+
"epoch": 906.31,
|
| 242391 |
+
"learning_rate": 8.200791599353796e-06,
|
| 242392 |
+
"loss": 0.306,
|
| 242393 |
+
"step": 112415
|
| 242394 |
+
},
|
| 242395 |
+
{
|
| 242396 |
+
"epoch": 906.35,
|
| 242397 |
+
"learning_rate": 8.200710823909532e-06,
|
| 242398 |
+
"loss": 0.369,
|
| 242399 |
+
"step": 112420
|
| 242400 |
+
},
|
| 242401 |
+
{
|
| 242402 |
+
"epoch": 906.4,
|
| 242403 |
+
"learning_rate": 8.200630048465266e-06,
|
| 242404 |
+
"loss": 0.912,
|
| 242405 |
+
"step": 112425
|
| 242406 |
+
},
|
| 242407 |
+
{
|
| 242408 |
+
"epoch": 906.44,
|
| 242409 |
+
"learning_rate": 8.200549273021002e-06,
|
| 242410 |
+
"loss": 0.7257,
|
| 242411 |
+
"step": 112430
|
| 242412 |
+
},
|
| 242413 |
+
{
|
| 242414 |
+
"epoch": 906.48,
|
| 242415 |
+
"learning_rate": 8.200468497576736e-06,
|
| 242416 |
+
"loss": 0.2877,
|
| 242417 |
+
"step": 112435
|
| 242418 |
+
},
|
| 242419 |
+
{
|
| 242420 |
+
"epoch": 906.52,
|
| 242421 |
+
"learning_rate": 8.200387722132472e-06,
|
| 242422 |
+
"loss": 0.3152,
|
| 242423 |
+
"step": 112440
|
| 242424 |
+
},
|
| 242425 |
+
{
|
| 242426 |
+
"epoch": 906.56,
|
| 242427 |
+
"learning_rate": 8.200306946688208e-06,
|
| 242428 |
+
"loss": 0.3952,
|
| 242429 |
+
"step": 112445
|
| 242430 |
+
},
|
| 242431 |
+
{
|
| 242432 |
+
"epoch": 906.6,
|
| 242433 |
+
"learning_rate": 8.200226171243942e-06,
|
| 242434 |
+
"loss": 0.8972,
|
| 242435 |
+
"step": 112450
|
| 242436 |
+
},
|
| 242437 |
+
{
|
| 242438 |
+
"epoch": 906.64,
|
| 242439 |
+
"learning_rate": 8.200145395799678e-06,
|
| 242440 |
+
"loss": 0.8771,
|
| 242441 |
+
"step": 112455
|
| 242442 |
+
},
|
| 242443 |
+
{
|
| 242444 |
+
"epoch": 906.68,
|
| 242445 |
+
"learning_rate": 8.200064620355412e-06,
|
| 242446 |
+
"loss": 0.3608,
|
| 242447 |
+
"step": 112460
|
| 242448 |
+
},
|
| 242449 |
+
{
|
| 242450 |
+
"epoch": 906.72,
|
| 242451 |
+
"learning_rate": 8.199983844911148e-06,
|
| 242452 |
+
"loss": 0.3328,
|
| 242453 |
+
"step": 112465
|
| 242454 |
+
},
|
| 242455 |
+
{
|
| 242456 |
+
"epoch": 906.76,
|
| 242457 |
+
"learning_rate": 8.199903069466882e-06,
|
| 242458 |
+
"loss": 0.3953,
|
| 242459 |
+
"step": 112470
|
| 242460 |
+
},
|
| 242461 |
+
{
|
| 242462 |
+
"epoch": 906.8,
|
| 242463 |
+
"learning_rate": 8.199822294022618e-06,
|
| 242464 |
+
"loss": 0.9336,
|
| 242465 |
+
"step": 112475
|
| 242466 |
+
},
|
| 242467 |
+
{
|
| 242468 |
+
"epoch": 906.84,
|
| 242469 |
+
"learning_rate": 8.199741518578352e-06,
|
| 242470 |
+
"loss": 0.7011,
|
| 242471 |
+
"step": 112480
|
| 242472 |
+
},
|
| 242473 |
+
{
|
| 242474 |
+
"epoch": 906.88,
|
| 242475 |
+
"learning_rate": 8.199660743134088e-06,
|
| 242476 |
+
"loss": 0.3199,
|
| 242477 |
+
"step": 112485
|
| 242478 |
+
},
|
| 242479 |
+
{
|
| 242480 |
+
"epoch": 906.92,
|
| 242481 |
+
"learning_rate": 8.199579967689822e-06,
|
| 242482 |
+
"loss": 0.2874,
|
| 242483 |
+
"step": 112490
|
| 242484 |
+
},
|
| 242485 |
+
{
|
| 242486 |
+
"epoch": 906.96,
|
| 242487 |
+
"learning_rate": 8.199499192245558e-06,
|
| 242488 |
+
"loss": 0.378,
|
| 242489 |
+
"step": 112495
|
| 242490 |
+
},
|
| 242491 |
+
{
|
| 242492 |
+
"epoch": 907.0,
|
| 242493 |
+
"learning_rate": 8.199418416801294e-06,
|
| 242494 |
+
"loss": 1.1028,
|
| 242495 |
+
"step": 112500
|
| 242496 |
+
},
|
| 242497 |
+
{
|
| 242498 |
+
"epoch": 907.0,
|
| 242499 |
+
"eval_loss": 0.3665727376937866,
|
| 242500 |
+
"eval_runtime": 40.8442,
|
| 242501 |
+
"eval_samples_per_second": 20.492,
|
| 242502 |
+
"eval_steps_per_second": 0.661,
|
| 242503 |
+
"eval_wer": 0.1838938822847202,
|
| 242504 |
+
"step": 112500
|
| 242505 |
+
},
|
| 242506 |
+
{
|
| 242507 |
+
"epoch": 900.04,
|
| 242508 |
+
"learning_rate": 8.199337641357028e-06,
|
| 242509 |
+
"loss": 0.3578,
|
| 242510 |
+
"step": 112505
|
| 242511 |
+
},
|
| 242512 |
+
{
|
| 242513 |
+
"epoch": 900.08,
|
| 242514 |
+
"learning_rate": 8.199256865912764e-06,
|
| 242515 |
+
"loss": 0.2938,
|
| 242516 |
+
"step": 112510
|
| 242517 |
+
},
|
| 242518 |
+
{
|
| 242519 |
+
"epoch": 900.12,
|
| 242520 |
+
"learning_rate": 8.199176090468498e-06,
|
| 242521 |
+
"loss": 0.3098,
|
| 242522 |
+
"step": 112515
|
| 242523 |
+
},
|
| 242524 |
+
{
|
| 242525 |
+
"epoch": 900.16,
|
| 242526 |
+
"learning_rate": 8.199095315024234e-06,
|
| 242527 |
+
"loss": 0.4641,
|
| 242528 |
+
"step": 112520
|
| 242529 |
+
},
|
| 242530 |
+
{
|
| 242531 |
+
"epoch": 900.2,
|
| 242532 |
+
"learning_rate": 8.199014539579968e-06,
|
| 242533 |
+
"loss": 1.0493,
|
| 242534 |
+
"step": 112525
|
| 242535 |
+
},
|
| 242536 |
+
{
|
| 242537 |
+
"epoch": 900.24,
|
| 242538 |
+
"learning_rate": 8.198933764135704e-06,
|
| 242539 |
+
"loss": 0.3391,
|
| 242540 |
+
"step": 112530
|
| 242541 |
+
},
|
| 242542 |
+
{
|
| 242543 |
+
"epoch": 900.28,
|
| 242544 |
+
"learning_rate": 8.198852988691438e-06,
|
| 242545 |
+
"loss": 0.3322,
|
| 242546 |
+
"step": 112535
|
| 242547 |
+
},
|
| 242548 |
+
{
|
| 242549 |
+
"epoch": 900.32,
|
| 242550 |
+
"learning_rate": 8.198772213247174e-06,
|
| 242551 |
+
"loss": 0.3384,
|
| 242552 |
+
"step": 112540
|
| 242553 |
+
},
|
| 242554 |
+
{
|
| 242555 |
+
"epoch": 900.36,
|
| 242556 |
+
"learning_rate": 8.198691437802908e-06,
|
| 242557 |
+
"loss": 0.4261,
|
| 242558 |
+
"step": 112545
|
| 242559 |
+
},
|
| 242560 |
+
{
|
| 242561 |
+
"epoch": 900.4,
|
| 242562 |
+
"learning_rate": 8.198610662358644e-06,
|
| 242563 |
+
"loss": 1.2445,
|
| 242564 |
+
"step": 112550
|
| 242565 |
+
},
|
| 242566 |
+
{
|
| 242567 |
+
"epoch": 900.44,
|
| 242568 |
+
"learning_rate": 8.19852988691438e-06,
|
| 242569 |
+
"loss": 0.2923,
|
| 242570 |
+
"step": 112555
|
| 242571 |
+
},
|
| 242572 |
+
{
|
| 242573 |
+
"epoch": 900.48,
|
| 242574 |
+
"learning_rate": 8.198449111470114e-06,
|
| 242575 |
+
"loss": 0.2442,
|
| 242576 |
+
"step": 112560
|
| 242577 |
+
},
|
| 242578 |
+
{
|
| 242579 |
+
"epoch": 900.52,
|
| 242580 |
+
"learning_rate": 8.19836833602585e-06,
|
| 242581 |
+
"loss": 0.3361,
|
| 242582 |
+
"step": 112565
|
| 242583 |
+
},
|
| 242584 |
+
{
|
| 242585 |
+
"epoch": 900.56,
|
| 242586 |
+
"learning_rate": 8.198287560581584e-06,
|
| 242587 |
+
"loss": 0.4356,
|
| 242588 |
+
"step": 112570
|
| 242589 |
+
},
|
| 242590 |
+
{
|
| 242591 |
+
"epoch": 900.6,
|
| 242592 |
+
"learning_rate": 8.19820678513732e-06,
|
| 242593 |
+
"loss": 1.1667,
|
| 242594 |
+
"step": 112575
|
| 242595 |
+
},
|
| 242596 |
+
{
|
| 242597 |
+
"epoch": 900.64,
|
| 242598 |
+
"learning_rate": 8.198126009693053e-06,
|
| 242599 |
+
"loss": 0.318,
|
| 242600 |
+
"step": 112580
|
| 242601 |
+
},
|
| 242602 |
+
{
|
| 242603 |
+
"epoch": 900.68,
|
| 242604 |
+
"learning_rate": 8.19804523424879e-06,
|
| 242605 |
+
"loss": 0.3111,
|
| 242606 |
+
"step": 112585
|
| 242607 |
+
},
|
| 242608 |
+
{
|
| 242609 |
+
"epoch": 900.72,
|
| 242610 |
+
"learning_rate": 8.197964458804523e-06,
|
| 242611 |
+
"loss": 0.3473,
|
| 242612 |
+
"step": 112590
|
| 242613 |
+
},
|
| 242614 |
+
{
|
| 242615 |
+
"epoch": 900.76,
|
| 242616 |
+
"learning_rate": 8.19788368336026e-06,
|
| 242617 |
+
"loss": 0.4569,
|
| 242618 |
+
"step": 112595
|
| 242619 |
+
},
|
| 242620 |
+
{
|
| 242621 |
+
"epoch": 900.8,
|
| 242622 |
+
"learning_rate": 8.197802907915993e-06,
|
| 242623 |
+
"loss": 1.2133,
|
| 242624 |
+
"step": 112600
|
| 242625 |
+
},
|
| 242626 |
+
{
|
| 242627 |
+
"epoch": 900.84,
|
| 242628 |
+
"learning_rate": 8.19772213247173e-06,
|
| 242629 |
+
"loss": 0.324,
|
| 242630 |
+
"step": 112605
|
| 242631 |
+
},
|
| 242632 |
+
{
|
| 242633 |
+
"epoch": 900.88,
|
| 242634 |
+
"learning_rate": 8.197641357027463e-06,
|
| 242635 |
+
"loss": 0.2444,
|
| 242636 |
+
"step": 112610
|
| 242637 |
+
},
|
| 242638 |
+
{
|
| 242639 |
+
"epoch": 900.92,
|
| 242640 |
+
"learning_rate": 8.1975605815832e-06,
|
| 242641 |
+
"loss": 0.4005,
|
| 242642 |
+
"step": 112615
|
| 242643 |
+
},
|
| 242644 |
+
{
|
| 242645 |
+
"epoch": 900.96,
|
| 242646 |
+
"learning_rate": 8.197479806138935e-06,
|
| 242647 |
+
"loss": 0.528,
|
| 242648 |
+
"step": 112620
|
| 242649 |
+
},
|
| 242650 |
+
{
|
| 242651 |
+
"epoch": 901.0,
|
| 242652 |
+
"learning_rate": 8.19739903069467e-06,
|
| 242653 |
+
"loss": 1.3845,
|
| 242654 |
+
"step": 112625
|
| 242655 |
+
},
|
| 242656 |
+
{
|
| 242657 |
+
"epoch": 901.0,
|
| 242658 |
+
"eval_loss": 0.4503331482410431,
|
| 242659 |
+
"eval_runtime": 40.4509,
|
| 242660 |
+
"eval_samples_per_second": 20.667,
|
| 242661 |
+
"eval_steps_per_second": 0.667,
|
| 242662 |
+
"eval_wer": 0.18393001543777107,
|
| 242663 |
+
"step": 112625
|
| 242664 |
+
},
|
| 242665 |
+
{
|
| 242666 |
+
"epoch": 901.04,
|
| 242667 |
+
"learning_rate": 8.197318255250405e-06,
|
| 242668 |
+
"loss": 0.3084,
|
| 242669 |
+
"step": 112630
|
| 242670 |
+
},
|
| 242671 |
+
{
|
| 242672 |
+
"epoch": 901.08,
|
| 242673 |
+
"learning_rate": 8.19723747980614e-06,
|
| 242674 |
+
"loss": 0.3249,
|
| 242675 |
+
"step": 112635
|
| 242676 |
+
},
|
| 242677 |
+
{
|
| 242678 |
+
"epoch": 901.12,
|
| 242679 |
+
"learning_rate": 8.197156704361875e-06,
|
| 242680 |
+
"loss": 0.3247,
|
| 242681 |
+
"step": 112640
|
| 242682 |
+
},
|
| 242683 |
+
{
|
| 242684 |
+
"epoch": 901.16,
|
| 242685 |
+
"learning_rate": 8.19707592891761e-06,
|
| 242686 |
+
"loss": 0.4088,
|
| 242687 |
+
"step": 112645
|
| 242688 |
+
},
|
| 242689 |
+
{
|
| 242690 |
+
"epoch": 901.2,
|
| 242691 |
+
"learning_rate": 8.196995153473345e-06,
|
| 242692 |
+
"loss": 1.0528,
|
| 242693 |
+
"step": 112650
|
| 242694 |
+
},
|
| 242695 |
+
{
|
| 242696 |
+
"epoch": 901.24,
|
| 242697 |
+
"learning_rate": 8.19691437802908e-06,
|
| 242698 |
+
"loss": 0.3751,
|
| 242699 |
+
"step": 112655
|
| 242700 |
+
},
|
| 242701 |
+
{
|
| 242702 |
+
"epoch": 901.28,
|
| 242703 |
+
"learning_rate": 8.196833602584815e-06,
|
| 242704 |
+
"loss": 0.3149,
|
| 242705 |
+
"step": 112660
|
| 242706 |
+
},
|
| 242707 |
+
{
|
| 242708 |
+
"epoch": 901.32,
|
| 242709 |
+
"learning_rate": 8.196752827140549e-06,
|
| 242710 |
+
"loss": 0.3781,
|
| 242711 |
+
"step": 112665
|
| 242712 |
+
},
|
| 242713 |
+
{
|
| 242714 |
+
"epoch": 901.36,
|
| 242715 |
+
"learning_rate": 8.196672051696285e-06,
|
| 242716 |
+
"loss": 0.4784,
|
| 242717 |
+
"step": 112670
|
| 242718 |
+
},
|
| 242719 |
+
{
|
| 242720 |
+
"epoch": 901.4,
|
| 242721 |
+
"learning_rate": 8.19659127625202e-06,
|
| 242722 |
+
"loss": 1.2451,
|
| 242723 |
+
"step": 112675
|
| 242724 |
+
},
|
| 242725 |
+
{
|
| 242726 |
+
"epoch": 901.44,
|
| 242727 |
+
"learning_rate": 8.196510500807755e-06,
|
| 242728 |
+
"loss": 0.3353,
|
| 242729 |
+
"step": 112680
|
| 242730 |
+
},
|
| 242731 |
+
{
|
| 242732 |
+
"epoch": 901.48,
|
| 242733 |
+
"learning_rate": 8.19642972536349e-06,
|
| 242734 |
+
"loss": 0.2894,
|
| 242735 |
+
"step": 112685
|
| 242736 |
+
},
|
| 242737 |
+
{
|
| 242738 |
+
"epoch": 901.52,
|
| 242739 |
+
"learning_rate": 8.196348949919225e-06,
|
| 242740 |
+
"loss": 0.3487,
|
| 242741 |
+
"step": 112690
|
| 242742 |
+
},
|
| 242743 |
+
{
|
| 242744 |
+
"epoch": 901.56,
|
| 242745 |
+
"learning_rate": 8.19626817447496e-06,
|
| 242746 |
+
"loss": 0.4494,
|
| 242747 |
+
"step": 112695
|
| 242748 |
+
},
|
| 242749 |
+
{
|
| 242750 |
+
"epoch": 901.6,
|
| 242751 |
+
"learning_rate": 8.196187399030695e-06,
|
| 242752 |
+
"loss": 1.2488,
|
| 242753 |
+
"step": 112700
|
| 242754 |
+
},
|
| 242755 |
+
{
|
| 242756 |
+
"epoch": 901.64,
|
| 242757 |
+
"learning_rate": 8.19610662358643e-06,
|
| 242758 |
+
"loss": 0.3036,
|
| 242759 |
+
"step": 112705
|
| 242760 |
+
},
|
| 242761 |
+
{
|
| 242762 |
+
"epoch": 901.68,
|
| 242763 |
+
"learning_rate": 8.196025848142165e-06,
|
| 242764 |
+
"loss": 0.2959,
|
| 242765 |
+
"step": 112710
|
| 242766 |
+
},
|
| 242767 |
+
{
|
| 242768 |
+
"epoch": 901.72,
|
| 242769 |
+
"learning_rate": 8.1959450726979e-06,
|
| 242770 |
+
"loss": 0.2841,
|
| 242771 |
+
"step": 112715
|
| 242772 |
+
},
|
| 242773 |
+
{
|
| 242774 |
+
"epoch": 901.76,
|
| 242775 |
+
"learning_rate": 8.195864297253635e-06,
|
| 242776 |
+
"loss": 0.4297,
|
| 242777 |
+
"step": 112720
|
| 242778 |
+
},
|
| 242779 |
+
{
|
| 242780 |
+
"epoch": 901.8,
|
| 242781 |
+
"learning_rate": 8.19578352180937e-06,
|
| 242782 |
+
"loss": 1.2188,
|
| 242783 |
+
"step": 112725
|
| 242784 |
+
},
|
| 242785 |
+
{
|
| 242786 |
+
"epoch": 901.84,
|
| 242787 |
+
"learning_rate": 8.195702746365107e-06,
|
| 242788 |
+
"loss": 0.2619,
|
| 242789 |
+
"step": 112730
|
| 242790 |
+
},
|
| 242791 |
+
{
|
| 242792 |
+
"epoch": 901.88,
|
| 242793 |
+
"learning_rate": 8.19562197092084e-06,
|
| 242794 |
+
"loss": 0.2855,
|
| 242795 |
+
"step": 112735
|
| 242796 |
+
},
|
| 242797 |
+
{
|
| 242798 |
+
"epoch": 901.92,
|
| 242799 |
+
"learning_rate": 8.195541195476576e-06,
|
| 242800 |
+
"loss": 0.3261,
|
| 242801 |
+
"step": 112740
|
| 242802 |
+
},
|
| 242803 |
+
{
|
| 242804 |
+
"epoch": 901.96,
|
| 242805 |
+
"learning_rate": 8.19546042003231e-06,
|
| 242806 |
+
"loss": 0.4363,
|
| 242807 |
+
"step": 112745
|
| 242808 |
+
},
|
| 242809 |
+
{
|
| 242810 |
+
"epoch": 902.0,
|
| 242811 |
+
"learning_rate": 8.195379644588046e-06,
|
| 242812 |
+
"loss": 1.1429,
|
| 242813 |
+
"step": 112750
|
| 242814 |
+
},
|
| 242815 |
+
{
|
| 242816 |
+
"epoch": 902.0,
|
| 242817 |
+
"eval_loss": 0.5105034708976746,
|
| 242818 |
+
"eval_runtime": 40.0763,
|
| 242819 |
+
"eval_samples_per_second": 20.86,
|
| 242820 |
+
"eval_steps_per_second": 0.674,
|
| 242821 |
+
"eval_wer": 0.19159501927067893,
|
| 242822 |
+
"step": 112750
|
| 242823 |
}
|
| 242824 |
],
|
| 242825 |
"max_steps": 625000,
|
| 242826 |
"num_train_epochs": 5000,
|
| 242827 |
+
"total_flos": 3.1730977775534506e+20,
|
| 242828 |
"trial_name": null,
|
| 242829 |
"trial_params": null
|
| 242830 |
}
|
model-bin/finetune/base/{checkpoint-112251 β checkpoint-112750}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1630080962.0011778/events.out.tfevents.1630080962.dea8124df033.1100.1
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:65769844150994ff1555f45c9fd77bcdc22d6e17d7ec210515f7dcc46d700e00
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630081546.0007536/events.out.tfevents.1630081546.dea8124df033.1391.1
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6ca7f4f4e86a2d00bb90bf95a40178cbfacb9d1728beb7d620e9f81b02c3ec64
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630081975.5740092/events.out.tfevents.1630081975.dea8124df033.1391.3
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a09512d50d1ef26d2f297f63456ce7b2241a6d47e49e31d1709f1a540a19d283
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630082400.0108204/events.out.tfevents.1630082400.dea8124df033.1391.5
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:591d0fec865524e1bd554d8008efd614fcfcc6a63c2fd4f85f2741bdf87ecfe2
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630080961.dea8124df033.1100.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b6e5f1d6e1f7f53e3ece49153d4e77fcb0baf6ef7d1ad168e857905f19d2fb26
|
| 3 |
+
size 8630
|
model-bin/finetune/base/log/events.out.tfevents.1630081545.dea8124df033.1391.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4ac932a20bfe1b4c19ed8e11b0eed7b25a3e961e714f7aba090ff646d0b4daee
|
| 3 |
+
size 8630
|
model-bin/finetune/base/log/events.out.tfevents.1630081975.dea8124df033.1391.2
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3855248414824359cfde883f6e12381b6483dc45b106e824cc4e0814703b94cb
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630082400.dea8124df033.1391.4
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:65473aae46d56c2a52378f6db4915b2bbf7ee5b710e3bb6d6b4cf153cab8aad5
|
| 3 |
+
size 8622
|