"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-139135 β checkpoint-139632}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-139135 β checkpoint-139632}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-139135 β checkpoint-139632}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-139135 β checkpoint-139632}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-139135 β checkpoint-139632}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-139135 β checkpoint-139632}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-139135 β checkpoint-139632}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-139135 β checkpoint-139632}/trainer_state.json +633 -3
- model-bin/finetune/base/{checkpoint-139135 β checkpoint-139632}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630203840.7050338/events.out.tfevents.1630203840.cc93b136ebf5.1086.1 +3 -0
- model-bin/finetune/base/log/1630204292.5703435/events.out.tfevents.1630204292.cc93b136ebf5.1086.3 +3 -0
- model-bin/finetune/base/log/1630204747.1758041/events.out.tfevents.1630204747.cc93b136ebf5.1086.5 +3 -0
- model-bin/finetune/base/log/1630205190.8026295/events.out.tfevents.1630205190.cc93b136ebf5.1086.7 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630203840.cc93b136ebf5.1086.0 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630204292.cc93b136ebf5.1086.2 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630204747.cc93b136ebf5.1086.4 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630205190.cc93b136ebf5.1086.6 +3 -0
model-bin/finetune/base/{checkpoint-139135 β checkpoint-139632}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-139135 β checkpoint-139632}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d52be403964896ebde8dbc015ab0b31dd8eda84858da6f91dc257157cd345705
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-139135 β checkpoint-139632}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-139135 β checkpoint-139632}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d16fa8e963ee700b85a69e9291858ab4a52074a28116fa5897842dcac0729d57
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-139135 β checkpoint-139632}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:df8dc682c31d6a272b5908de4ff759bebd27e6bb8b81ddd6b7a5edc073866bc8
|
| 3 |
size 14503
|
model-bin/finetune/base/{checkpoint-139135 β checkpoint-139632}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:69fe6b1a171b3398b251c4f574c3897019a42807059a5531cd17cb9782eff416
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-139135 β checkpoint-139632}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0417338fc26171bc7c4c31f1905061af395fd857f395bcb53cac4e888fb96a67
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-139135 β checkpoint-139632}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1689111747851003,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-132910",
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -276390,11 +276390,641 @@
|
|
| 276390 |
"eval_steps_per_second": 0.736,
|
| 276391 |
"eval_wer": 0.17295327903986285,
|
| 276392 |
"step": 139135
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 276393 |
}
|
| 276394 |
],
|
| 276395 |
"max_steps": 620000,
|
| 276396 |
"num_train_epochs": 5000,
|
| 276397 |
-
"total_flos": 3.
|
| 276398 |
"trial_name": null,
|
| 276399 |
"trial_params": null
|
| 276400 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.1689111747851003,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-132910",
|
| 4 |
+
"epoch": 1125.995983935743,
|
| 5 |
+
"global_step": 139632,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 276390 |
"eval_steps_per_second": 0.736,
|
| 276391 |
"eval_wer": 0.17295327903986285,
|
| 276392 |
"step": 139135
|
| 276393 |
+
},
|
| 276394 |
+
{
|
| 276395 |
+
"epoch": 1122.04,
|
| 276396 |
+
"learning_rate": 7.769224555735058e-06,
|
| 276397 |
+
"loss": 0.2935,
|
| 276398 |
+
"step": 139140
|
| 276399 |
+
},
|
| 276400 |
+
{
|
| 276401 |
+
"epoch": 1122.08,
|
| 276402 |
+
"learning_rate": 7.769143780290792e-06,
|
| 276403 |
+
"loss": 0.2453,
|
| 276404 |
+
"step": 139145
|
| 276405 |
+
},
|
| 276406 |
+
{
|
| 276407 |
+
"epoch": 1122.12,
|
| 276408 |
+
"learning_rate": 7.769063004846528e-06,
|
| 276409 |
+
"loss": 0.3313,
|
| 276410 |
+
"step": 139150
|
| 276411 |
+
},
|
| 276412 |
+
{
|
| 276413 |
+
"epoch": 1122.16,
|
| 276414 |
+
"learning_rate": 7.768982229402262e-06,
|
| 276415 |
+
"loss": 0.4549,
|
| 276416 |
+
"step": 139155
|
| 276417 |
+
},
|
| 276418 |
+
{
|
| 276419 |
+
"epoch": 1122.2,
|
| 276420 |
+
"learning_rate": 7.768901453957998e-06,
|
| 276421 |
+
"loss": 1.0227,
|
| 276422 |
+
"step": 139160
|
| 276423 |
+
},
|
| 276424 |
+
{
|
| 276425 |
+
"epoch": 1122.24,
|
| 276426 |
+
"learning_rate": 7.768820678513732e-06,
|
| 276427 |
+
"loss": 0.3099,
|
| 276428 |
+
"step": 139165
|
| 276429 |
+
},
|
| 276430 |
+
{
|
| 276431 |
+
"epoch": 1122.28,
|
| 276432 |
+
"learning_rate": 7.768739903069468e-06,
|
| 276433 |
+
"loss": 0.2407,
|
| 276434 |
+
"step": 139170
|
| 276435 |
+
},
|
| 276436 |
+
{
|
| 276437 |
+
"epoch": 1122.32,
|
| 276438 |
+
"learning_rate": 7.768659127625202e-06,
|
| 276439 |
+
"loss": 0.3476,
|
| 276440 |
+
"step": 139175
|
| 276441 |
+
},
|
| 276442 |
+
{
|
| 276443 |
+
"epoch": 1122.36,
|
| 276444 |
+
"learning_rate": 7.768578352180938e-06,
|
| 276445 |
+
"loss": 0.4268,
|
| 276446 |
+
"step": 139180
|
| 276447 |
+
},
|
| 276448 |
+
{
|
| 276449 |
+
"epoch": 1122.4,
|
| 276450 |
+
"learning_rate": 7.768497576736672e-06,
|
| 276451 |
+
"loss": 1.1477,
|
| 276452 |
+
"step": 139185
|
| 276453 |
+
},
|
| 276454 |
+
{
|
| 276455 |
+
"epoch": 1122.44,
|
| 276456 |
+
"learning_rate": 7.768416801292408e-06,
|
| 276457 |
+
"loss": 0.318,
|
| 276458 |
+
"step": 139190
|
| 276459 |
+
},
|
| 276460 |
+
{
|
| 276461 |
+
"epoch": 1122.48,
|
| 276462 |
+
"learning_rate": 7.768336025848144e-06,
|
| 276463 |
+
"loss": 0.2953,
|
| 276464 |
+
"step": 139195
|
| 276465 |
+
},
|
| 276466 |
+
{
|
| 276467 |
+
"epoch": 1122.52,
|
| 276468 |
+
"learning_rate": 7.768255250403878e-06,
|
| 276469 |
+
"loss": 0.2743,
|
| 276470 |
+
"step": 139200
|
| 276471 |
+
},
|
| 276472 |
+
{
|
| 276473 |
+
"epoch": 1122.56,
|
| 276474 |
+
"learning_rate": 7.768174474959614e-06,
|
| 276475 |
+
"loss": 0.4145,
|
| 276476 |
+
"step": 139205
|
| 276477 |
+
},
|
| 276478 |
+
{
|
| 276479 |
+
"epoch": 1122.6,
|
| 276480 |
+
"learning_rate": 7.768093699515348e-06,
|
| 276481 |
+
"loss": 1.4303,
|
| 276482 |
+
"step": 139210
|
| 276483 |
+
},
|
| 276484 |
+
{
|
| 276485 |
+
"epoch": 1122.65,
|
| 276486 |
+
"learning_rate": 7.768012924071084e-06,
|
| 276487 |
+
"loss": 0.3557,
|
| 276488 |
+
"step": 139215
|
| 276489 |
+
},
|
| 276490 |
+
{
|
| 276491 |
+
"epoch": 1122.69,
|
| 276492 |
+
"learning_rate": 7.767932148626818e-06,
|
| 276493 |
+
"loss": 0.3282,
|
| 276494 |
+
"step": 139220
|
| 276495 |
+
},
|
| 276496 |
+
{
|
| 276497 |
+
"epoch": 1122.73,
|
| 276498 |
+
"learning_rate": 7.767851373182554e-06,
|
| 276499 |
+
"loss": 0.3062,
|
| 276500 |
+
"step": 139225
|
| 276501 |
+
},
|
| 276502 |
+
{
|
| 276503 |
+
"epoch": 1122.77,
|
| 276504 |
+
"learning_rate": 7.767770597738288e-06,
|
| 276505 |
+
"loss": 0.4282,
|
| 276506 |
+
"step": 139230
|
| 276507 |
+
},
|
| 276508 |
+
{
|
| 276509 |
+
"epoch": 1122.81,
|
| 276510 |
+
"learning_rate": 7.767689822294024e-06,
|
| 276511 |
+
"loss": 1.1902,
|
| 276512 |
+
"step": 139235
|
| 276513 |
+
},
|
| 276514 |
+
{
|
| 276515 |
+
"epoch": 1122.85,
|
| 276516 |
+
"learning_rate": 7.767609046849758e-06,
|
| 276517 |
+
"loss": 0.2787,
|
| 276518 |
+
"step": 139240
|
| 276519 |
+
},
|
| 276520 |
+
{
|
| 276521 |
+
"epoch": 1122.89,
|
| 276522 |
+
"learning_rate": 7.767528271405494e-06,
|
| 276523 |
+
"loss": 0.3426,
|
| 276524 |
+
"step": 139245
|
| 276525 |
+
},
|
| 276526 |
+
{
|
| 276527 |
+
"epoch": 1122.93,
|
| 276528 |
+
"learning_rate": 7.76744749596123e-06,
|
| 276529 |
+
"loss": 0.2978,
|
| 276530 |
+
"step": 139250
|
| 276531 |
+
},
|
| 276532 |
+
{
|
| 276533 |
+
"epoch": 1122.97,
|
| 276534 |
+
"learning_rate": 7.767366720516964e-06,
|
| 276535 |
+
"loss": 0.4585,
|
| 276536 |
+
"step": 139255
|
| 276537 |
+
},
|
| 276538 |
+
{
|
| 276539 |
+
"epoch": 1123.0,
|
| 276540 |
+
"eval_loss": 0.3832680881023407,
|
| 276541 |
+
"eval_runtime": 41.7996,
|
| 276542 |
+
"eval_samples_per_second": 20.0,
|
| 276543 |
+
"eval_steps_per_second": 0.646,
|
| 276544 |
+
"eval_wer": 0.17813765182186234,
|
| 276545 |
+
"step": 139259
|
| 276546 |
+
},
|
| 276547 |
+
{
|
| 276548 |
+
"epoch": 1114.01,
|
| 276549 |
+
"learning_rate": 7.7672859450727e-06,
|
| 276550 |
+
"loss": 0.4918,
|
| 276551 |
+
"step": 139260
|
| 276552 |
+
},
|
| 276553 |
+
{
|
| 276554 |
+
"epoch": 1114.05,
|
| 276555 |
+
"learning_rate": 7.767205169628434e-06,
|
| 276556 |
+
"loss": 0.2842,
|
| 276557 |
+
"step": 139265
|
| 276558 |
+
},
|
| 276559 |
+
{
|
| 276560 |
+
"epoch": 1114.09,
|
| 276561 |
+
"learning_rate": 7.76712439418417e-06,
|
| 276562 |
+
"loss": 0.2855,
|
| 276563 |
+
"step": 139270
|
| 276564 |
+
},
|
| 276565 |
+
{
|
| 276566 |
+
"epoch": 1114.13,
|
| 276567 |
+
"learning_rate": 7.767043618739904e-06,
|
| 276568 |
+
"loss": 0.2647,
|
| 276569 |
+
"step": 139275
|
| 276570 |
+
},
|
| 276571 |
+
{
|
| 276572 |
+
"epoch": 1114.17,
|
| 276573 |
+
"learning_rate": 7.76696284329564e-06,
|
| 276574 |
+
"loss": 0.5487,
|
| 276575 |
+
"step": 139280
|
| 276576 |
+
},
|
| 276577 |
+
{
|
| 276578 |
+
"epoch": 1114.21,
|
| 276579 |
+
"learning_rate": 7.766882067851374e-06,
|
| 276580 |
+
"loss": 1.0039,
|
| 276581 |
+
"step": 139285
|
| 276582 |
+
},
|
| 276583 |
+
{
|
| 276584 |
+
"epoch": 1114.25,
|
| 276585 |
+
"learning_rate": 7.76680129240711e-06,
|
| 276586 |
+
"loss": 0.3161,
|
| 276587 |
+
"step": 139290
|
| 276588 |
+
},
|
| 276589 |
+
{
|
| 276590 |
+
"epoch": 1114.29,
|
| 276591 |
+
"learning_rate": 7.766720516962844e-06,
|
| 276592 |
+
"loss": 0.3176,
|
| 276593 |
+
"step": 139295
|
| 276594 |
+
},
|
| 276595 |
+
{
|
| 276596 |
+
"epoch": 1114.33,
|
| 276597 |
+
"learning_rate": 7.76663974151858e-06,
|
| 276598 |
+
"loss": 0.3167,
|
| 276599 |
+
"step": 139300
|
| 276600 |
+
},
|
| 276601 |
+
{
|
| 276602 |
+
"epoch": 1114.37,
|
| 276603 |
+
"learning_rate": 7.766558966074315e-06,
|
| 276604 |
+
"loss": 0.5792,
|
| 276605 |
+
"step": 139305
|
| 276606 |
+
},
|
| 276607 |
+
{
|
| 276608 |
+
"epoch": 1114.41,
|
| 276609 |
+
"learning_rate": 7.76647819063005e-06,
|
| 276610 |
+
"loss": 0.9895,
|
| 276611 |
+
"step": 139310
|
| 276612 |
+
},
|
| 276613 |
+
{
|
| 276614 |
+
"epoch": 1114.45,
|
| 276615 |
+
"learning_rate": 7.766397415185785e-06,
|
| 276616 |
+
"loss": 0.275,
|
| 276617 |
+
"step": 139315
|
| 276618 |
+
},
|
| 276619 |
+
{
|
| 276620 |
+
"epoch": 1114.49,
|
| 276621 |
+
"learning_rate": 7.76631663974152e-06,
|
| 276622 |
+
"loss": 0.2364,
|
| 276623 |
+
"step": 139320
|
| 276624 |
+
},
|
| 276625 |
+
{
|
| 276626 |
+
"epoch": 1114.53,
|
| 276627 |
+
"learning_rate": 7.766235864297255e-06,
|
| 276628 |
+
"loss": 0.3155,
|
| 276629 |
+
"step": 139325
|
| 276630 |
+
},
|
| 276631 |
+
{
|
| 276632 |
+
"epoch": 1114.57,
|
| 276633 |
+
"learning_rate": 7.76615508885299e-06,
|
| 276634 |
+
"loss": 0.5086,
|
| 276635 |
+
"step": 139330
|
| 276636 |
+
},
|
| 276637 |
+
{
|
| 276638 |
+
"epoch": 1114.61,
|
| 276639 |
+
"learning_rate": 7.766074313408725e-06,
|
| 276640 |
+
"loss": 1.0574,
|
| 276641 |
+
"step": 139335
|
| 276642 |
+
},
|
| 276643 |
+
{
|
| 276644 |
+
"epoch": 1114.65,
|
| 276645 |
+
"learning_rate": 7.76599353796446e-06,
|
| 276646 |
+
"loss": 0.3179,
|
| 276647 |
+
"step": 139340
|
| 276648 |
+
},
|
| 276649 |
+
{
|
| 276650 |
+
"epoch": 1114.69,
|
| 276651 |
+
"learning_rate": 7.765912762520195e-06,
|
| 276652 |
+
"loss": 0.3073,
|
| 276653 |
+
"step": 139345
|
| 276654 |
+
},
|
| 276655 |
+
{
|
| 276656 |
+
"epoch": 1114.73,
|
| 276657 |
+
"learning_rate": 7.76583198707593e-06,
|
| 276658 |
+
"loss": 0.3061,
|
| 276659 |
+
"step": 139350
|
| 276660 |
+
},
|
| 276661 |
+
{
|
| 276662 |
+
"epoch": 1114.77,
|
| 276663 |
+
"learning_rate": 7.765751211631665e-06,
|
| 276664 |
+
"loss": 0.5684,
|
| 276665 |
+
"step": 139355
|
| 276666 |
+
},
|
| 276667 |
+
{
|
| 276668 |
+
"epoch": 1114.81,
|
| 276669 |
+
"learning_rate": 7.7656704361874e-06,
|
| 276670 |
+
"loss": 1.2499,
|
| 276671 |
+
"step": 139360
|
| 276672 |
+
},
|
| 276673 |
+
{
|
| 276674 |
+
"epoch": 1114.85,
|
| 276675 |
+
"learning_rate": 7.765589660743135e-06,
|
| 276676 |
+
"loss": 0.3199,
|
| 276677 |
+
"step": 139365
|
| 276678 |
+
},
|
| 276679 |
+
{
|
| 276680 |
+
"epoch": 1114.89,
|
| 276681 |
+
"learning_rate": 7.765508885298871e-06,
|
| 276682 |
+
"loss": 0.2328,
|
| 276683 |
+
"step": 139370
|
| 276684 |
+
},
|
| 276685 |
+
{
|
| 276686 |
+
"epoch": 1114.93,
|
| 276687 |
+
"learning_rate": 7.765428109854605e-06,
|
| 276688 |
+
"loss": 0.3404,
|
| 276689 |
+
"step": 139375
|
| 276690 |
+
},
|
| 276691 |
+
{
|
| 276692 |
+
"epoch": 1114.97,
|
| 276693 |
+
"learning_rate": 7.765347334410341e-06,
|
| 276694 |
+
"loss": 0.6416,
|
| 276695 |
+
"step": 139380
|
| 276696 |
+
},
|
| 276697 |
+
{
|
| 276698 |
+
"epoch": 1115.0,
|
| 276699 |
+
"eval_loss": 0.39288097620010376,
|
| 276700 |
+
"eval_runtime": 42.1595,
|
| 276701 |
+
"eval_samples_per_second": 19.829,
|
| 276702 |
+
"eval_steps_per_second": 0.64,
|
| 276703 |
+
"eval_wer": 0.18118466898954705,
|
| 276704 |
+
"step": 139384
|
| 276705 |
+
},
|
| 276706 |
+
{
|
| 276707 |
+
"epoch": 1124.01,
|
| 276708 |
+
"learning_rate": 7.765266558966075e-06,
|
| 276709 |
+
"loss": 0.3201,
|
| 276710 |
+
"step": 139385
|
| 276711 |
+
},
|
| 276712 |
+
{
|
| 276713 |
+
"epoch": 1124.05,
|
| 276714 |
+
"learning_rate": 7.76518578352181e-06,
|
| 276715 |
+
"loss": 0.2739,
|
| 276716 |
+
"step": 139390
|
| 276717 |
+
},
|
| 276718 |
+
{
|
| 276719 |
+
"epoch": 1124.09,
|
| 276720 |
+
"learning_rate": 7.765105008077545e-06,
|
| 276721 |
+
"loss": 0.3218,
|
| 276722 |
+
"step": 139395
|
| 276723 |
+
},
|
| 276724 |
+
{
|
| 276725 |
+
"epoch": 1124.13,
|
| 276726 |
+
"learning_rate": 7.76502423263328e-06,
|
| 276727 |
+
"loss": 0.3122,
|
| 276728 |
+
"step": 139400
|
| 276729 |
+
},
|
| 276730 |
+
{
|
| 276731 |
+
"epoch": 1124.17,
|
| 276732 |
+
"learning_rate": 7.764943457189015e-06,
|
| 276733 |
+
"loss": 0.5611,
|
| 276734 |
+
"step": 139405
|
| 276735 |
+
},
|
| 276736 |
+
{
|
| 276737 |
+
"epoch": 1124.21,
|
| 276738 |
+
"learning_rate": 7.76486268174475e-06,
|
| 276739 |
+
"loss": 0.9412,
|
| 276740 |
+
"step": 139410
|
| 276741 |
+
},
|
| 276742 |
+
{
|
| 276743 |
+
"epoch": 1124.25,
|
| 276744 |
+
"learning_rate": 7.764781906300485e-06,
|
| 276745 |
+
"loss": 0.2755,
|
| 276746 |
+
"step": 139415
|
| 276747 |
+
},
|
| 276748 |
+
{
|
| 276749 |
+
"epoch": 1124.29,
|
| 276750 |
+
"learning_rate": 7.76470113085622e-06,
|
| 276751 |
+
"loss": 0.2831,
|
| 276752 |
+
"step": 139420
|
| 276753 |
+
},
|
| 276754 |
+
{
|
| 276755 |
+
"epoch": 1124.33,
|
| 276756 |
+
"learning_rate": 7.764620355411957e-06,
|
| 276757 |
+
"loss": 0.3349,
|
| 276758 |
+
"step": 139425
|
| 276759 |
+
},
|
| 276760 |
+
{
|
| 276761 |
+
"epoch": 1124.37,
|
| 276762 |
+
"learning_rate": 7.76453957996769e-06,
|
| 276763 |
+
"loss": 0.5646,
|
| 276764 |
+
"step": 139430
|
| 276765 |
+
},
|
| 276766 |
+
{
|
| 276767 |
+
"epoch": 1124.41,
|
| 276768 |
+
"learning_rate": 7.764458804523427e-06,
|
| 276769 |
+
"loss": 1.1037,
|
| 276770 |
+
"step": 139435
|
| 276771 |
+
},
|
| 276772 |
+
{
|
| 276773 |
+
"epoch": 1124.45,
|
| 276774 |
+
"learning_rate": 7.76437802907916e-06,
|
| 276775 |
+
"loss": 0.31,
|
| 276776 |
+
"step": 139440
|
| 276777 |
+
},
|
| 276778 |
+
{
|
| 276779 |
+
"epoch": 1124.49,
|
| 276780 |
+
"learning_rate": 7.764297253634897e-06,
|
| 276781 |
+
"loss": 0.2414,
|
| 276782 |
+
"step": 139445
|
| 276783 |
+
},
|
| 276784 |
+
{
|
| 276785 |
+
"epoch": 1124.53,
|
| 276786 |
+
"learning_rate": 7.76421647819063e-06,
|
| 276787 |
+
"loss": 0.3528,
|
| 276788 |
+
"step": 139450
|
| 276789 |
+
},
|
| 276790 |
+
{
|
| 276791 |
+
"epoch": 1124.57,
|
| 276792 |
+
"learning_rate": 7.764135702746367e-06,
|
| 276793 |
+
"loss": 0.5236,
|
| 276794 |
+
"step": 139455
|
| 276795 |
+
},
|
| 276796 |
+
{
|
| 276797 |
+
"epoch": 1124.61,
|
| 276798 |
+
"learning_rate": 7.7640549273021e-06,
|
| 276799 |
+
"loss": 1.0225,
|
| 276800 |
+
"step": 139460
|
| 276801 |
+
},
|
| 276802 |
+
{
|
| 276803 |
+
"epoch": 1124.65,
|
| 276804 |
+
"learning_rate": 7.763974151857836e-06,
|
| 276805 |
+
"loss": 0.2665,
|
| 276806 |
+
"step": 139465
|
| 276807 |
+
},
|
| 276808 |
+
{
|
| 276809 |
+
"epoch": 1124.69,
|
| 276810 |
+
"learning_rate": 7.76389337641357e-06,
|
| 276811 |
+
"loss": 0.2925,
|
| 276812 |
+
"step": 139470
|
| 276813 |
+
},
|
| 276814 |
+
{
|
| 276815 |
+
"epoch": 1124.73,
|
| 276816 |
+
"learning_rate": 7.763812600969306e-06,
|
| 276817 |
+
"loss": 0.3135,
|
| 276818 |
+
"step": 139475
|
| 276819 |
+
},
|
| 276820 |
+
{
|
| 276821 |
+
"epoch": 1124.77,
|
| 276822 |
+
"learning_rate": 7.763731825525042e-06,
|
| 276823 |
+
"loss": 0.5225,
|
| 276824 |
+
"step": 139480
|
| 276825 |
+
},
|
| 276826 |
+
{
|
| 276827 |
+
"epoch": 1124.81,
|
| 276828 |
+
"learning_rate": 7.763651050080776e-06,
|
| 276829 |
+
"loss": 1.0957,
|
| 276830 |
+
"step": 139485
|
| 276831 |
+
},
|
| 276832 |
+
{
|
| 276833 |
+
"epoch": 1124.85,
|
| 276834 |
+
"learning_rate": 7.763570274636512e-06,
|
| 276835 |
+
"loss": 0.251,
|
| 276836 |
+
"step": 139490
|
| 276837 |
+
},
|
| 276838 |
+
{
|
| 276839 |
+
"epoch": 1124.89,
|
| 276840 |
+
"learning_rate": 7.763489499192246e-06,
|
| 276841 |
+
"loss": 0.3002,
|
| 276842 |
+
"step": 139495
|
| 276843 |
+
},
|
| 276844 |
+
{
|
| 276845 |
+
"epoch": 1124.93,
|
| 276846 |
+
"learning_rate": 7.763408723747982e-06,
|
| 276847 |
+
"loss": 0.3116,
|
| 276848 |
+
"step": 139500
|
| 276849 |
+
},
|
| 276850 |
+
{
|
| 276851 |
+
"epoch": 1124.97,
|
| 276852 |
+
"learning_rate": 7.763327948303716e-06,
|
| 276853 |
+
"loss": 0.5954,
|
| 276854 |
+
"step": 139505
|
| 276855 |
+
},
|
| 276856 |
+
{
|
| 276857 |
+
"epoch": 1125.0,
|
| 276858 |
+
"eval_loss": 0.3883971869945526,
|
| 276859 |
+
"eval_runtime": 41.5038,
|
| 276860 |
+
"eval_samples_per_second": 20.143,
|
| 276861 |
+
"eval_steps_per_second": 0.651,
|
| 276862 |
+
"eval_wer": 0.18588165062308784,
|
| 276863 |
+
"step": 139508
|
| 276864 |
+
},
|
| 276865 |
+
{
|
| 276866 |
+
"epoch": 1125.02,
|
| 276867 |
+
"learning_rate": 7.763247172859452e-06,
|
| 276868 |
+
"loss": 0.3039,
|
| 276869 |
+
"step": 139510
|
| 276870 |
+
},
|
| 276871 |
+
{
|
| 276872 |
+
"epoch": 1125.06,
|
| 276873 |
+
"learning_rate": 7.763166397415186e-06,
|
| 276874 |
+
"loss": 0.288,
|
| 276875 |
+
"step": 139515
|
| 276876 |
+
},
|
| 276877 |
+
{
|
| 276878 |
+
"epoch": 1125.1,
|
| 276879 |
+
"learning_rate": 7.763085621970922e-06,
|
| 276880 |
+
"loss": 0.3032,
|
| 276881 |
+
"step": 139520
|
| 276882 |
+
},
|
| 276883 |
+
{
|
| 276884 |
+
"epoch": 1125.14,
|
| 276885 |
+
"learning_rate": 7.763004846526656e-06,
|
| 276886 |
+
"loss": 0.338,
|
| 276887 |
+
"step": 139525
|
| 276888 |
+
},
|
| 276889 |
+
{
|
| 276890 |
+
"epoch": 1125.18,
|
| 276891 |
+
"learning_rate": 7.762924071082392e-06,
|
| 276892 |
+
"loss": 0.5968,
|
| 276893 |
+
"step": 139530
|
| 276894 |
+
},
|
| 276895 |
+
{
|
| 276896 |
+
"epoch": 1125.22,
|
| 276897 |
+
"learning_rate": 7.762843295638126e-06,
|
| 276898 |
+
"loss": 0.9873,
|
| 276899 |
+
"step": 139535
|
| 276900 |
+
},
|
| 276901 |
+
{
|
| 276902 |
+
"epoch": 1125.26,
|
| 276903 |
+
"learning_rate": 7.762762520193862e-06,
|
| 276904 |
+
"loss": 0.2503,
|
| 276905 |
+
"step": 139540
|
| 276906 |
+
},
|
| 276907 |
+
{
|
| 276908 |
+
"epoch": 1125.3,
|
| 276909 |
+
"learning_rate": 7.762681744749598e-06,
|
| 276910 |
+
"loss": 0.2329,
|
| 276911 |
+
"step": 139545
|
| 276912 |
+
},
|
| 276913 |
+
{
|
| 276914 |
+
"epoch": 1125.34,
|
| 276915 |
+
"learning_rate": 7.762600969305332e-06,
|
| 276916 |
+
"loss": 0.3378,
|
| 276917 |
+
"step": 139550
|
| 276918 |
+
},
|
| 276919 |
+
{
|
| 276920 |
+
"epoch": 1125.38,
|
| 276921 |
+
"learning_rate": 7.762520193861068e-06,
|
| 276922 |
+
"loss": 0.5935,
|
| 276923 |
+
"step": 139555
|
| 276924 |
+
},
|
| 276925 |
+
{
|
| 276926 |
+
"epoch": 1125.42,
|
| 276927 |
+
"learning_rate": 7.762439418416802e-06,
|
| 276928 |
+
"loss": 1.0363,
|
| 276929 |
+
"step": 139560
|
| 276930 |
+
},
|
| 276931 |
+
{
|
| 276932 |
+
"epoch": 1125.46,
|
| 276933 |
+
"learning_rate": 7.762358642972538e-06,
|
| 276934 |
+
"loss": 0.2453,
|
| 276935 |
+
"step": 139565
|
| 276936 |
+
},
|
| 276937 |
+
{
|
| 276938 |
+
"epoch": 1125.5,
|
| 276939 |
+
"learning_rate": 7.762277867528272e-06,
|
| 276940 |
+
"loss": 0.2953,
|
| 276941 |
+
"step": 139570
|
| 276942 |
+
},
|
| 276943 |
+
{
|
| 276944 |
+
"epoch": 1125.54,
|
| 276945 |
+
"learning_rate": 7.762197092084008e-06,
|
| 276946 |
+
"loss": 0.3298,
|
| 276947 |
+
"step": 139575
|
| 276948 |
+
},
|
| 276949 |
+
{
|
| 276950 |
+
"epoch": 1125.58,
|
| 276951 |
+
"learning_rate": 7.762116316639742e-06,
|
| 276952 |
+
"loss": 0.666,
|
| 276953 |
+
"step": 139580
|
| 276954 |
+
},
|
| 276955 |
+
{
|
| 276956 |
+
"epoch": 1125.62,
|
| 276957 |
+
"learning_rate": 7.762035541195478e-06,
|
| 276958 |
+
"loss": 1.066,
|
| 276959 |
+
"step": 139585
|
| 276960 |
+
},
|
| 276961 |
+
{
|
| 276962 |
+
"epoch": 1125.66,
|
| 276963 |
+
"learning_rate": 7.761954765751212e-06,
|
| 276964 |
+
"loss": 0.3161,
|
| 276965 |
+
"step": 139590
|
| 276966 |
+
},
|
| 276967 |
+
{
|
| 276968 |
+
"epoch": 1125.7,
|
| 276969 |
+
"learning_rate": 7.761873990306948e-06,
|
| 276970 |
+
"loss": 0.3122,
|
| 276971 |
+
"step": 139595
|
| 276972 |
+
},
|
| 276973 |
+
{
|
| 276974 |
+
"epoch": 1125.74,
|
| 276975 |
+
"learning_rate": 7.761793214862684e-06,
|
| 276976 |
+
"loss": 0.2955,
|
| 276977 |
+
"step": 139600
|
| 276978 |
+
},
|
| 276979 |
+
{
|
| 276980 |
+
"epoch": 1125.78,
|
| 276981 |
+
"learning_rate": 7.761712439418418e-06,
|
| 276982 |
+
"loss": 0.5424,
|
| 276983 |
+
"step": 139605
|
| 276984 |
+
},
|
| 276985 |
+
{
|
| 276986 |
+
"epoch": 1125.82,
|
| 276987 |
+
"learning_rate": 7.761631663974154e-06,
|
| 276988 |
+
"loss": 0.8868,
|
| 276989 |
+
"step": 139610
|
| 276990 |
+
},
|
| 276991 |
+
{
|
| 276992 |
+
"epoch": 1125.86,
|
| 276993 |
+
"learning_rate": 7.761550888529888e-06,
|
| 276994 |
+
"loss": 0.2635,
|
| 276995 |
+
"step": 139615
|
| 276996 |
+
},
|
| 276997 |
+
{
|
| 276998 |
+
"epoch": 1125.9,
|
| 276999 |
+
"learning_rate": 7.761470113085624e-06,
|
| 277000 |
+
"loss": 0.3167,
|
| 277001 |
+
"step": 139620
|
| 277002 |
+
},
|
| 277003 |
+
{
|
| 277004 |
+
"epoch": 1125.94,
|
| 277005 |
+
"learning_rate": 7.761389337641358e-06,
|
| 277006 |
+
"loss": 0.3052,
|
| 277007 |
+
"step": 139625
|
| 277008 |
+
},
|
| 277009 |
+
{
|
| 277010 |
+
"epoch": 1125.98,
|
| 277011 |
+
"learning_rate": 7.761308562197094e-06,
|
| 277012 |
+
"loss": 0.5751,
|
| 277013 |
+
"step": 139630
|
| 277014 |
+
},
|
| 277015 |
+
{
|
| 277016 |
+
"epoch": 1126.0,
|
| 277017 |
+
"eval_loss": 0.41787150502204895,
|
| 277018 |
+
"eval_runtime": 42.2152,
|
| 277019 |
+
"eval_samples_per_second": 19.803,
|
| 277020 |
+
"eval_steps_per_second": 0.64,
|
| 277021 |
+
"eval_wer": 0.18331400405914758,
|
| 277022 |
+
"step": 139632
|
| 277023 |
}
|
| 277024 |
],
|
| 277025 |
"max_steps": 620000,
|
| 277026 |
"num_train_epochs": 5000,
|
| 277027 |
+
"total_flos": 3.929101689489278e+20,
|
| 277028 |
"trial_name": null,
|
| 277029 |
"trial_params": null
|
| 277030 |
}
|
model-bin/finetune/base/{checkpoint-139135 β checkpoint-139632}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1630203840.7050338/events.out.tfevents.1630203840.cc93b136ebf5.1086.1
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:273ad8879055fa1e8a3be1e72b9b855c122061cae57d1da0f9bea0929dbd9f0a
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630204292.5703435/events.out.tfevents.1630204292.cc93b136ebf5.1086.3
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5ca2e0c0c8f9b92125a834acb99214859eb42d27631539078d418d4d7a59d332
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630204747.1758041/events.out.tfevents.1630204747.cc93b136ebf5.1086.5
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0bb22e6c31b54936b35de42462c29b9ee7cc4aac67112d63ab6877005aedb2c3
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630205190.8026295/events.out.tfevents.1630205190.cc93b136ebf5.1086.7
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:042e5b679dc07a09f4ec878e65f77977578dc02281f0cd0ad44acf9b158a362c
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630203840.cc93b136ebf5.1086.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:eb14bae35002d5cf692a8356c6ec57ad533583f92066c63b16278303cb7a761d
|
| 3 |
+
size 8470
|
model-bin/finetune/base/log/events.out.tfevents.1630204292.cc93b136ebf5.1086.2
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4fed4d49e432c5adef9fc96ece4196590b69cbc2ef8caf390a569df99a8b2f7c
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630204747.cc93b136ebf5.1086.4
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ac35332ab8480fac9861525d27699460c902247ca9e0af69aba5ee2b42768ad0
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630205190.cc93b136ebf5.1086.6
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ac28020d3a642b31669b12aba95f397cf2494c95d49171996def28b6601352ed
|
| 3 |
+
size 8622
|