"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-60729 β checkpoint-67447}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-66949 β checkpoint-67447}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-60729 β checkpoint-67447}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-66949 β checkpoint-67447}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-66949 β checkpoint-67447}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-66949 β checkpoint-67447}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-60729 β checkpoint-67447}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-66949 β checkpoint-67447}/trainer_state.json +641 -5
- model-bin/finetune/base/{checkpoint-60729 β checkpoint-67447}/training_args.bin +0 -0
- model-bin/finetune/base/{checkpoint-66949 β checkpoint-67572}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-60729 β checkpoint-67572}/optimizer.pt +2 -2
- model-bin/finetune/base/{checkpoint-66949 β checkpoint-67572}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-60729 β checkpoint-67572}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-60729 β checkpoint-67572}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-60729 β checkpoint-67572}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-66949 β checkpoint-67572}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-60729 β checkpoint-67572}/trainer_state.json +0 -0
- model-bin/finetune/base/{checkpoint-66949 β checkpoint-67572}/training_args.bin +0 -0
- model-bin/finetune/base/log/1629834135.9336932/events.out.tfevents.1629834135.c435e1c5ee04.920.211 +3 -0
- model-bin/finetune/base/log/1629834756.205052/events.out.tfevents.1629834756.c435e1c5ee04.920.213 +3 -0
- model-bin/finetune/base/log/1629835507.9034107/events.out.tfevents.1629835507.c435e1c5ee04.920.215 +3 -0
- model-bin/finetune/base/log/1629836153.2164783/events.out.tfevents.1629836153.c435e1c5ee04.920.217 +3 -0
- model-bin/finetune/base/log/1629836799.1113734/events.out.tfevents.1629836799.c435e1c5ee04.920.219 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629834135.c435e1c5ee04.920.210 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629834756.c435e1c5ee04.920.212 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629835507.c435e1c5ee04.920.214 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629836153.c435e1c5ee04.920.216 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629836799.c435e1c5ee04.920.218 +3 -0
model-bin/finetune/base/{checkpoint-60729 β checkpoint-67447}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-66949 β checkpoint-67447}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4f29be78508a07fb73240b11ffa7167d4632163ad60416d94009458f2a159529
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-60729 β checkpoint-67447}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-66949 β checkpoint-67447}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:eb6e5067d265b6d0ac7d38a072eaf79353dd3359b87c14179ec86e37964a2578
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-66949 β checkpoint-67447}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:506537a02628941466ae681cb8b9d6d5f1a22f228b010fe75395d0621cb8473b
|
| 3 |
size 14503
|
model-bin/finetune/base/{checkpoint-66949 β checkpoint-67447}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:87e2d1745c681c91159da5acdbfd8bc474c3ecf40e467f65b9d5603d68c91173
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-60729 β checkpoint-67447}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:754bdca3fd4dc1dd416ee07d27446b6b442be519e3b7e30a419105bdc19f4b05
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-66949 β checkpoint-67447}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
-
"best_metric": 0.
|
| 3 |
-
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -184542,11 +184542,647 @@
|
|
| 184542 |
"eval_steps_per_second": 0.679,
|
| 184543 |
"eval_wer": 0.20303119051105578,
|
| 184544 |
"step": 66949
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 184545 |
}
|
| 184546 |
],
|
| 184547 |
"max_steps": 620000,
|
| 184548 |
"num_train_epochs": 5000,
|
| 184549 |
-
"total_flos": 1.
|
| 184550 |
"trial_name": null,
|
| 184551 |
"trial_params": null
|
| 184552 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"best_metric": 0.1855363713557883,
|
| 3 |
+
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-67447",
|
| 4 |
+
"epoch": 542.995983935743,
|
| 5 |
+
"global_step": 67447,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 184542 |
"eval_steps_per_second": 0.679,
|
| 184543 |
"eval_wer": 0.20303119051105578,
|
| 184544 |
"step": 66949
|
| 184545 |
+
},
|
| 184546 |
+
{
|
| 184547 |
+
"epoch": 539.01,
|
| 184548 |
+
"learning_rate": 8.943509615384616e-06,
|
| 184549 |
+
"loss": 0.5425,
|
| 184550 |
+
"step": 66950
|
| 184551 |
+
},
|
| 184552 |
+
{
|
| 184553 |
+
"epoch": 539.05,
|
| 184554 |
+
"learning_rate": 8.943429487179488e-06,
|
| 184555 |
+
"loss": 0.3015,
|
| 184556 |
+
"step": 66955
|
| 184557 |
+
},
|
| 184558 |
+
{
|
| 184559 |
+
"epoch": 539.09,
|
| 184560 |
+
"learning_rate": 8.943349358974359e-06,
|
| 184561 |
+
"loss": 0.2871,
|
| 184562 |
+
"step": 66960
|
| 184563 |
+
},
|
| 184564 |
+
{
|
| 184565 |
+
"epoch": 539.13,
|
| 184566 |
+
"learning_rate": 8.943269230769232e-06,
|
| 184567 |
+
"loss": 0.3098,
|
| 184568 |
+
"step": 66965
|
| 184569 |
+
},
|
| 184570 |
+
{
|
| 184571 |
+
"epoch": 539.17,
|
| 184572 |
+
"learning_rate": 8.943189102564103e-06,
|
| 184573 |
+
"loss": 0.5809,
|
| 184574 |
+
"step": 66970
|
| 184575 |
+
},
|
| 184576 |
+
{
|
| 184577 |
+
"epoch": 539.21,
|
| 184578 |
+
"learning_rate": 8.943108974358975e-06,
|
| 184579 |
+
"loss": 1.1017,
|
| 184580 |
+
"step": 66975
|
| 184581 |
+
},
|
| 184582 |
+
{
|
| 184583 |
+
"epoch": 539.25,
|
| 184584 |
+
"learning_rate": 8.943028846153848e-06,
|
| 184585 |
+
"loss": 0.3202,
|
| 184586 |
+
"step": 66980
|
| 184587 |
+
},
|
| 184588 |
+
{
|
| 184589 |
+
"epoch": 539.29,
|
| 184590 |
+
"learning_rate": 8.94294871794872e-06,
|
| 184591 |
+
"loss": 0.3306,
|
| 184592 |
+
"step": 66985
|
| 184593 |
+
},
|
| 184594 |
+
{
|
| 184595 |
+
"epoch": 539.33,
|
| 184596 |
+
"learning_rate": 8.94286858974359e-06,
|
| 184597 |
+
"loss": 0.3722,
|
| 184598 |
+
"step": 66990
|
| 184599 |
+
},
|
| 184600 |
+
{
|
| 184601 |
+
"epoch": 539.37,
|
| 184602 |
+
"learning_rate": 8.942788461538462e-06,
|
| 184603 |
+
"loss": 0.5289,
|
| 184604 |
+
"step": 66995
|
| 184605 |
+
},
|
| 184606 |
+
{
|
| 184607 |
+
"epoch": 539.41,
|
| 184608 |
+
"learning_rate": 8.942708333333335e-06,
|
| 184609 |
+
"loss": 1.2412,
|
| 184610 |
+
"step": 67000
|
| 184611 |
+
},
|
| 184612 |
+
{
|
| 184613 |
+
"epoch": 539.45,
|
| 184614 |
+
"learning_rate": 8.942628205128205e-06,
|
| 184615 |
+
"loss": 0.3389,
|
| 184616 |
+
"step": 67005
|
| 184617 |
+
},
|
| 184618 |
+
{
|
| 184619 |
+
"epoch": 539.49,
|
| 184620 |
+
"learning_rate": 8.942548076923078e-06,
|
| 184621 |
+
"loss": 0.3951,
|
| 184622 |
+
"step": 67010
|
| 184623 |
+
},
|
| 184624 |
+
{
|
| 184625 |
+
"epoch": 539.53,
|
| 184626 |
+
"learning_rate": 8.94246794871795e-06,
|
| 184627 |
+
"loss": 0.3677,
|
| 184628 |
+
"step": 67015
|
| 184629 |
+
},
|
| 184630 |
+
{
|
| 184631 |
+
"epoch": 539.57,
|
| 184632 |
+
"learning_rate": 8.94238782051282e-06,
|
| 184633 |
+
"loss": 0.4942,
|
| 184634 |
+
"step": 67020
|
| 184635 |
+
},
|
| 184636 |
+
{
|
| 184637 |
+
"epoch": 539.61,
|
| 184638 |
+
"learning_rate": 8.942307692307693e-06,
|
| 184639 |
+
"loss": 1.1607,
|
| 184640 |
+
"step": 67025
|
| 184641 |
+
},
|
| 184642 |
+
{
|
| 184643 |
+
"epoch": 539.65,
|
| 184644 |
+
"learning_rate": 8.942227564102565e-06,
|
| 184645 |
+
"loss": 0.3204,
|
| 184646 |
+
"step": 67030
|
| 184647 |
+
},
|
| 184648 |
+
{
|
| 184649 |
+
"epoch": 539.69,
|
| 184650 |
+
"learning_rate": 8.942147435897436e-06,
|
| 184651 |
+
"loss": 0.2899,
|
| 184652 |
+
"step": 67035
|
| 184653 |
+
},
|
| 184654 |
+
{
|
| 184655 |
+
"epoch": 539.73,
|
| 184656 |
+
"learning_rate": 8.942067307692308e-06,
|
| 184657 |
+
"loss": 0.5907,
|
| 184658 |
+
"step": 67040
|
| 184659 |
+
},
|
| 184660 |
+
{
|
| 184661 |
+
"epoch": 539.77,
|
| 184662 |
+
"learning_rate": 8.94198717948718e-06,
|
| 184663 |
+
"loss": 0.5997,
|
| 184664 |
+
"step": 67045
|
| 184665 |
+
},
|
| 184666 |
+
{
|
| 184667 |
+
"epoch": 539.81,
|
| 184668 |
+
"learning_rate": 8.941907051282052e-06,
|
| 184669 |
+
"loss": 1.2436,
|
| 184670 |
+
"step": 67050
|
| 184671 |
+
},
|
| 184672 |
+
{
|
| 184673 |
+
"epoch": 539.85,
|
| 184674 |
+
"learning_rate": 8.941826923076923e-06,
|
| 184675 |
+
"loss": 0.372,
|
| 184676 |
+
"step": 67055
|
| 184677 |
+
},
|
| 184678 |
+
{
|
| 184679 |
+
"epoch": 539.9,
|
| 184680 |
+
"learning_rate": 8.941746794871795e-06,
|
| 184681 |
+
"loss": 0.3978,
|
| 184682 |
+
"step": 67060
|
| 184683 |
+
},
|
| 184684 |
+
{
|
| 184685 |
+
"epoch": 539.94,
|
| 184686 |
+
"learning_rate": 8.941666666666668e-06,
|
| 184687 |
+
"loss": 0.3908,
|
| 184688 |
+
"step": 67065
|
| 184689 |
+
},
|
| 184690 |
+
{
|
| 184691 |
+
"epoch": 539.98,
|
| 184692 |
+
"learning_rate": 8.941586538461539e-06,
|
| 184693 |
+
"loss": 0.7207,
|
| 184694 |
+
"step": 67070
|
| 184695 |
+
},
|
| 184696 |
+
{
|
| 184697 |
+
"epoch": 540.0,
|
| 184698 |
+
"eval_loss": 0.42999500036239624,
|
| 184699 |
+
"eval_runtime": 39.6376,
|
| 184700 |
+
"eval_samples_per_second": 21.116,
|
| 184701 |
+
"eval_steps_per_second": 0.681,
|
| 184702 |
+
"eval_wer": 0.19431621465666474,
|
| 184703 |
+
"step": 67073
|
| 184704 |
+
},
|
| 184705 |
+
{
|
| 184706 |
+
"epoch": 536.02,
|
| 184707 |
+
"learning_rate": 8.94150641025641e-06,
|
| 184708 |
+
"loss": 0.4822,
|
| 184709 |
+
"step": 67075
|
| 184710 |
+
},
|
| 184711 |
+
{
|
| 184712 |
+
"epoch": 536.06,
|
| 184713 |
+
"learning_rate": 8.941426282051284e-06,
|
| 184714 |
+
"loss": 0.2812,
|
| 184715 |
+
"step": 67080
|
| 184716 |
+
},
|
| 184717 |
+
{
|
| 184718 |
+
"epoch": 536.1,
|
| 184719 |
+
"learning_rate": 8.941346153846155e-06,
|
| 184720 |
+
"loss": 0.3065,
|
| 184721 |
+
"step": 67085
|
| 184722 |
+
},
|
| 184723 |
+
{
|
| 184724 |
+
"epoch": 536.14,
|
| 184725 |
+
"learning_rate": 8.941266025641026e-06,
|
| 184726 |
+
"loss": 0.3642,
|
| 184727 |
+
"step": 67090
|
| 184728 |
+
},
|
| 184729 |
+
{
|
| 184730 |
+
"epoch": 536.18,
|
| 184731 |
+
"learning_rate": 8.941185897435898e-06,
|
| 184732 |
+
"loss": 0.6448,
|
| 184733 |
+
"step": 67095
|
| 184734 |
+
},
|
| 184735 |
+
{
|
| 184736 |
+
"epoch": 536.22,
|
| 184737 |
+
"learning_rate": 8.94110576923077e-06,
|
| 184738 |
+
"loss": 1.0078,
|
| 184739 |
+
"step": 67100
|
| 184740 |
+
},
|
| 184741 |
+
{
|
| 184742 |
+
"epoch": 536.26,
|
| 184743 |
+
"learning_rate": 8.941025641025642e-06,
|
| 184744 |
+
"loss": 0.3443,
|
| 184745 |
+
"step": 67105
|
| 184746 |
+
},
|
| 184747 |
+
{
|
| 184748 |
+
"epoch": 536.3,
|
| 184749 |
+
"learning_rate": 8.940945512820513e-06,
|
| 184750 |
+
"loss": 0.387,
|
| 184751 |
+
"step": 67110
|
| 184752 |
+
},
|
| 184753 |
+
{
|
| 184754 |
+
"epoch": 536.34,
|
| 184755 |
+
"learning_rate": 8.940865384615386e-06,
|
| 184756 |
+
"loss": 0.3822,
|
| 184757 |
+
"step": 67115
|
| 184758 |
+
},
|
| 184759 |
+
{
|
| 184760 |
+
"epoch": 536.38,
|
| 184761 |
+
"learning_rate": 8.940785256410258e-06,
|
| 184762 |
+
"loss": 0.6696,
|
| 184763 |
+
"step": 67120
|
| 184764 |
+
},
|
| 184765 |
+
{
|
| 184766 |
+
"epoch": 536.42,
|
| 184767 |
+
"learning_rate": 8.940705128205129e-06,
|
| 184768 |
+
"loss": 1.0635,
|
| 184769 |
+
"step": 67125
|
| 184770 |
+
},
|
| 184771 |
+
{
|
| 184772 |
+
"epoch": 536.46,
|
| 184773 |
+
"learning_rate": 8.940625e-06,
|
| 184774 |
+
"loss": 0.2825,
|
| 184775 |
+
"step": 67130
|
| 184776 |
+
},
|
| 184777 |
+
{
|
| 184778 |
+
"epoch": 536.5,
|
| 184779 |
+
"learning_rate": 8.940544871794874e-06,
|
| 184780 |
+
"loss": 0.3859,
|
| 184781 |
+
"step": 67135
|
| 184782 |
+
},
|
| 184783 |
+
{
|
| 184784 |
+
"epoch": 536.54,
|
| 184785 |
+
"learning_rate": 8.940464743589743e-06,
|
| 184786 |
+
"loss": 0.4542,
|
| 184787 |
+
"step": 67140
|
| 184788 |
+
},
|
| 184789 |
+
{
|
| 184790 |
+
"epoch": 536.58,
|
| 184791 |
+
"learning_rate": 8.940384615384616e-06,
|
| 184792 |
+
"loss": 0.7055,
|
| 184793 |
+
"step": 67145
|
| 184794 |
+
},
|
| 184795 |
+
{
|
| 184796 |
+
"epoch": 536.62,
|
| 184797 |
+
"learning_rate": 8.940304487179488e-06,
|
| 184798 |
+
"loss": 0.9673,
|
| 184799 |
+
"step": 67150
|
| 184800 |
+
},
|
| 184801 |
+
{
|
| 184802 |
+
"epoch": 536.66,
|
| 184803 |
+
"learning_rate": 8.940224358974359e-06,
|
| 184804 |
+
"loss": 0.3229,
|
| 184805 |
+
"step": 67155
|
| 184806 |
+
},
|
| 184807 |
+
{
|
| 184808 |
+
"epoch": 536.7,
|
| 184809 |
+
"learning_rate": 8.94014423076923e-06,
|
| 184810 |
+
"loss": 0.3374,
|
| 184811 |
+
"step": 67160
|
| 184812 |
+
},
|
| 184813 |
+
{
|
| 184814 |
+
"epoch": 536.74,
|
| 184815 |
+
"learning_rate": 8.940064102564103e-06,
|
| 184816 |
+
"loss": 0.4204,
|
| 184817 |
+
"step": 67165
|
| 184818 |
+
},
|
| 184819 |
+
{
|
| 184820 |
+
"epoch": 536.78,
|
| 184821 |
+
"learning_rate": 8.939983974358975e-06,
|
| 184822 |
+
"loss": 0.7498,
|
| 184823 |
+
"step": 67170
|
| 184824 |
+
},
|
| 184825 |
+
{
|
| 184826 |
+
"epoch": 536.82,
|
| 184827 |
+
"learning_rate": 8.939903846153846e-06,
|
| 184828 |
+
"loss": 1.0866,
|
| 184829 |
+
"step": 67175
|
| 184830 |
+
},
|
| 184831 |
+
{
|
| 184832 |
+
"epoch": 536.86,
|
| 184833 |
+
"learning_rate": 8.939823717948719e-06,
|
| 184834 |
+
"loss": 0.3682,
|
| 184835 |
+
"step": 67180
|
| 184836 |
+
},
|
| 184837 |
+
{
|
| 184838 |
+
"epoch": 536.9,
|
| 184839 |
+
"learning_rate": 8.93974358974359e-06,
|
| 184840 |
+
"loss": 0.3961,
|
| 184841 |
+
"step": 67185
|
| 184842 |
+
},
|
| 184843 |
+
{
|
| 184844 |
+
"epoch": 536.94,
|
| 184845 |
+
"learning_rate": 8.939663461538462e-06,
|
| 184846 |
+
"loss": 0.3728,
|
| 184847 |
+
"step": 67190
|
| 184848 |
+
},
|
| 184849 |
+
{
|
| 184850 |
+
"epoch": 536.98,
|
| 184851 |
+
"learning_rate": 8.939583333333333e-06,
|
| 184852 |
+
"loss": 0.607,
|
| 184853 |
+
"step": 67195
|
| 184854 |
+
},
|
| 184855 |
+
{
|
| 184856 |
+
"epoch": 537.0,
|
| 184857 |
+
"eval_loss": 0.4263405501842499,
|
| 184858 |
+
"eval_runtime": 39.1689,
|
| 184859 |
+
"eval_samples_per_second": 21.369,
|
| 184860 |
+
"eval_steps_per_second": 0.689,
|
| 184861 |
+
"eval_wer": 0.18978423601937472,
|
| 184862 |
+
"step": 67198
|
| 184863 |
+
},
|
| 184864 |
+
{
|
| 184865 |
+
"epoch": 537.02,
|
| 184866 |
+
"learning_rate": 8.939503205128206e-06,
|
| 184867 |
+
"loss": 0.3609,
|
| 184868 |
+
"step": 67200
|
| 184869 |
+
},
|
| 184870 |
+
{
|
| 184871 |
+
"epoch": 537.06,
|
| 184872 |
+
"learning_rate": 8.939423076923078e-06,
|
| 184873 |
+
"loss": 0.3246,
|
| 184874 |
+
"step": 67205
|
| 184875 |
+
},
|
| 184876 |
+
{
|
| 184877 |
+
"epoch": 537.1,
|
| 184878 |
+
"learning_rate": 8.939342948717949e-06,
|
| 184879 |
+
"loss": 0.3176,
|
| 184880 |
+
"step": 67210
|
| 184881 |
+
},
|
| 184882 |
+
{
|
| 184883 |
+
"epoch": 537.14,
|
| 184884 |
+
"learning_rate": 8.939262820512822e-06,
|
| 184885 |
+
"loss": 0.4736,
|
| 184886 |
+
"step": 67215
|
| 184887 |
+
},
|
| 184888 |
+
{
|
| 184889 |
+
"epoch": 537.18,
|
| 184890 |
+
"learning_rate": 8.939182692307693e-06,
|
| 184891 |
+
"loss": 0.6372,
|
| 184892 |
+
"step": 67220
|
| 184893 |
+
},
|
| 184894 |
+
{
|
| 184895 |
+
"epoch": 537.22,
|
| 184896 |
+
"learning_rate": 8.939102564102565e-06,
|
| 184897 |
+
"loss": 1.0112,
|
| 184898 |
+
"step": 67225
|
| 184899 |
+
},
|
| 184900 |
+
{
|
| 184901 |
+
"epoch": 537.26,
|
| 184902 |
+
"learning_rate": 8.939022435897436e-06,
|
| 184903 |
+
"loss": 0.4262,
|
| 184904 |
+
"step": 67230
|
| 184905 |
+
},
|
| 184906 |
+
{
|
| 184907 |
+
"epoch": 537.3,
|
| 184908 |
+
"learning_rate": 8.93894230769231e-06,
|
| 184909 |
+
"loss": 0.3433,
|
| 184910 |
+
"step": 67235
|
| 184911 |
+
},
|
| 184912 |
+
{
|
| 184913 |
+
"epoch": 537.34,
|
| 184914 |
+
"learning_rate": 8.93886217948718e-06,
|
| 184915 |
+
"loss": 0.363,
|
| 184916 |
+
"step": 67240
|
| 184917 |
+
},
|
| 184918 |
+
{
|
| 184919 |
+
"epoch": 537.38,
|
| 184920 |
+
"learning_rate": 8.938782051282052e-06,
|
| 184921 |
+
"loss": 0.6827,
|
| 184922 |
+
"step": 67245
|
| 184923 |
+
},
|
| 184924 |
+
{
|
| 184925 |
+
"epoch": 537.42,
|
| 184926 |
+
"learning_rate": 8.938701923076923e-06,
|
| 184927 |
+
"loss": 0.9992,
|
| 184928 |
+
"step": 67250
|
| 184929 |
+
},
|
| 184930 |
+
{
|
| 184931 |
+
"epoch": 537.46,
|
| 184932 |
+
"learning_rate": 8.938621794871796e-06,
|
| 184933 |
+
"loss": 0.3153,
|
| 184934 |
+
"step": 67255
|
| 184935 |
+
},
|
| 184936 |
+
{
|
| 184937 |
+
"epoch": 537.5,
|
| 184938 |
+
"learning_rate": 8.938541666666668e-06,
|
| 184939 |
+
"loss": 0.2932,
|
| 184940 |
+
"step": 67260
|
| 184941 |
+
},
|
| 184942 |
+
{
|
| 184943 |
+
"epoch": 537.54,
|
| 184944 |
+
"learning_rate": 8.938461538461539e-06,
|
| 184945 |
+
"loss": 0.3589,
|
| 184946 |
+
"step": 67265
|
| 184947 |
+
},
|
| 184948 |
+
{
|
| 184949 |
+
"epoch": 537.58,
|
| 184950 |
+
"learning_rate": 8.938381410256412e-06,
|
| 184951 |
+
"loss": 0.7378,
|
| 184952 |
+
"step": 67270
|
| 184953 |
+
},
|
| 184954 |
+
{
|
| 184955 |
+
"epoch": 537.62,
|
| 184956 |
+
"learning_rate": 8.938301282051283e-06,
|
| 184957 |
+
"loss": 1.2153,
|
| 184958 |
+
"step": 67275
|
| 184959 |
+
},
|
| 184960 |
+
{
|
| 184961 |
+
"epoch": 537.66,
|
| 184962 |
+
"learning_rate": 8.938221153846155e-06,
|
| 184963 |
+
"loss": 0.3399,
|
| 184964 |
+
"step": 67280
|
| 184965 |
+
},
|
| 184966 |
+
{
|
| 184967 |
+
"epoch": 537.7,
|
| 184968 |
+
"learning_rate": 8.938141025641026e-06,
|
| 184969 |
+
"loss": 0.3661,
|
| 184970 |
+
"step": 67285
|
| 184971 |
+
},
|
| 184972 |
+
{
|
| 184973 |
+
"epoch": 537.74,
|
| 184974 |
+
"learning_rate": 8.9380608974359e-06,
|
| 184975 |
+
"loss": 0.3713,
|
| 184976 |
+
"step": 67290
|
| 184977 |
+
},
|
| 184978 |
+
{
|
| 184979 |
+
"epoch": 537.78,
|
| 184980 |
+
"learning_rate": 8.937980769230769e-06,
|
| 184981 |
+
"loss": 0.6155,
|
| 184982 |
+
"step": 67295
|
| 184983 |
+
},
|
| 184984 |
+
{
|
| 184985 |
+
"epoch": 537.82,
|
| 184986 |
+
"learning_rate": 8.937900641025642e-06,
|
| 184987 |
+
"loss": 1.0946,
|
| 184988 |
+
"step": 67300
|
| 184989 |
+
},
|
| 184990 |
+
{
|
| 184991 |
+
"epoch": 537.86,
|
| 184992 |
+
"learning_rate": 8.937820512820513e-06,
|
| 184993 |
+
"loss": 0.3214,
|
| 184994 |
+
"step": 67305
|
| 184995 |
+
},
|
| 184996 |
+
{
|
| 184997 |
+
"epoch": 537.9,
|
| 184998 |
+
"learning_rate": 8.937740384615385e-06,
|
| 184999 |
+
"loss": 0.3359,
|
| 185000 |
+
"step": 67310
|
| 185001 |
+
},
|
| 185002 |
+
{
|
| 185003 |
+
"epoch": 537.94,
|
| 185004 |
+
"learning_rate": 8.937660256410258e-06,
|
| 185005 |
+
"loss": 0.3965,
|
| 185006 |
+
"step": 67315
|
| 185007 |
+
},
|
| 185008 |
+
{
|
| 185009 |
+
"epoch": 537.98,
|
| 185010 |
+
"learning_rate": 8.937580128205129e-06,
|
| 185011 |
+
"loss": 0.7338,
|
| 185012 |
+
"step": 67320
|
| 185013 |
+
},
|
| 185014 |
+
{
|
| 185015 |
+
"epoch": 538.0,
|
| 185016 |
+
"eval_loss": 0.46577340364456177,
|
| 185017 |
+
"eval_runtime": 38.9427,
|
| 185018 |
+
"eval_samples_per_second": 21.544,
|
| 185019 |
+
"eval_steps_per_second": 0.693,
|
| 185020 |
+
"eval_wer": 0.19503469986406238,
|
| 185021 |
+
"step": 67323
|
| 185022 |
+
},
|
| 185023 |
+
{
|
| 185024 |
+
"epoch": 542.02,
|
| 185025 |
+
"learning_rate": 8.9375e-06,
|
| 185026 |
+
"loss": 0.4422,
|
| 185027 |
+
"step": 67325
|
| 185028 |
+
},
|
| 185029 |
+
{
|
| 185030 |
+
"epoch": 542.06,
|
| 185031 |
+
"learning_rate": 8.937419871794872e-06,
|
| 185032 |
+
"loss": 0.3339,
|
| 185033 |
+
"step": 67330
|
| 185034 |
+
},
|
| 185035 |
+
{
|
| 185036 |
+
"epoch": 542.1,
|
| 185037 |
+
"learning_rate": 8.937339743589745e-06,
|
| 185038 |
+
"loss": 0.3232,
|
| 185039 |
+
"step": 67335
|
| 185040 |
+
},
|
| 185041 |
+
{
|
| 185042 |
+
"epoch": 542.14,
|
| 185043 |
+
"learning_rate": 8.937259615384616e-06,
|
| 185044 |
+
"loss": 0.3989,
|
| 185045 |
+
"step": 67340
|
| 185046 |
+
},
|
| 185047 |
+
{
|
| 185048 |
+
"epoch": 542.18,
|
| 185049 |
+
"learning_rate": 8.937179487179488e-06,
|
| 185050 |
+
"loss": 0.7477,
|
| 185051 |
+
"step": 67345
|
| 185052 |
+
},
|
| 185053 |
+
{
|
| 185054 |
+
"epoch": 542.22,
|
| 185055 |
+
"learning_rate": 8.937099358974359e-06,
|
| 185056 |
+
"loss": 1.1284,
|
| 185057 |
+
"step": 67350
|
| 185058 |
+
},
|
| 185059 |
+
{
|
| 185060 |
+
"epoch": 542.26,
|
| 185061 |
+
"learning_rate": 8.937019230769232e-06,
|
| 185062 |
+
"loss": 0.2786,
|
| 185063 |
+
"step": 67355
|
| 185064 |
+
},
|
| 185065 |
+
{
|
| 185066 |
+
"epoch": 542.3,
|
| 185067 |
+
"learning_rate": 8.936939102564103e-06,
|
| 185068 |
+
"loss": 0.3625,
|
| 185069 |
+
"step": 67360
|
| 185070 |
+
},
|
| 185071 |
+
{
|
| 185072 |
+
"epoch": 542.34,
|
| 185073 |
+
"learning_rate": 8.936858974358975e-06,
|
| 185074 |
+
"loss": 0.4097,
|
| 185075 |
+
"step": 67365
|
| 185076 |
+
},
|
| 185077 |
+
{
|
| 185078 |
+
"epoch": 542.38,
|
| 185079 |
+
"learning_rate": 8.936778846153848e-06,
|
| 185080 |
+
"loss": 0.8628,
|
| 185081 |
+
"step": 67370
|
| 185082 |
+
},
|
| 185083 |
+
{
|
| 185084 |
+
"epoch": 542.42,
|
| 185085 |
+
"learning_rate": 8.936698717948719e-06,
|
| 185086 |
+
"loss": 1.0227,
|
| 185087 |
+
"step": 67375
|
| 185088 |
+
},
|
| 185089 |
+
{
|
| 185090 |
+
"epoch": 542.46,
|
| 185091 |
+
"learning_rate": 8.93661858974359e-06,
|
| 185092 |
+
"loss": 0.2886,
|
| 185093 |
+
"step": 67380
|
| 185094 |
+
},
|
| 185095 |
+
{
|
| 185096 |
+
"epoch": 542.5,
|
| 185097 |
+
"learning_rate": 8.936538461538462e-06,
|
| 185098 |
+
"loss": 0.283,
|
| 185099 |
+
"step": 67385
|
| 185100 |
+
},
|
| 185101 |
+
{
|
| 185102 |
+
"epoch": 542.54,
|
| 185103 |
+
"learning_rate": 8.936458333333335e-06,
|
| 185104 |
+
"loss": 0.3514,
|
| 185105 |
+
"step": 67390
|
| 185106 |
+
},
|
| 185107 |
+
{
|
| 185108 |
+
"epoch": 542.58,
|
| 185109 |
+
"learning_rate": 8.936378205128206e-06,
|
| 185110 |
+
"loss": 0.674,
|
| 185111 |
+
"step": 67395
|
| 185112 |
+
},
|
| 185113 |
+
{
|
| 185114 |
+
"epoch": 542.62,
|
| 185115 |
+
"learning_rate": 8.936298076923078e-06,
|
| 185116 |
+
"loss": 1.0492,
|
| 185117 |
+
"step": 67400
|
| 185118 |
+
},
|
| 185119 |
+
{
|
| 185120 |
+
"epoch": 542.66,
|
| 185121 |
+
"learning_rate": 8.936217948717949e-06,
|
| 185122 |
+
"loss": 0.4139,
|
| 185123 |
+
"step": 67405
|
| 185124 |
+
},
|
| 185125 |
+
{
|
| 185126 |
+
"epoch": 542.7,
|
| 185127 |
+
"learning_rate": 8.936137820512822e-06,
|
| 185128 |
+
"loss": 0.3585,
|
| 185129 |
+
"step": 67410
|
| 185130 |
+
},
|
| 185131 |
+
{
|
| 185132 |
+
"epoch": 542.74,
|
| 185133 |
+
"learning_rate": 8.936057692307693e-06,
|
| 185134 |
+
"loss": 0.3681,
|
| 185135 |
+
"step": 67415
|
| 185136 |
+
},
|
| 185137 |
+
{
|
| 185138 |
+
"epoch": 542.78,
|
| 185139 |
+
"learning_rate": 8.935977564102565e-06,
|
| 185140 |
+
"loss": 0.6159,
|
| 185141 |
+
"step": 67420
|
| 185142 |
+
},
|
| 185143 |
+
{
|
| 185144 |
+
"epoch": 542.82,
|
| 185145 |
+
"learning_rate": 8.935897435897438e-06,
|
| 185146 |
+
"loss": 1.0444,
|
| 185147 |
+
"step": 67425
|
| 185148 |
+
},
|
| 185149 |
+
{
|
| 185150 |
+
"epoch": 542.86,
|
| 185151 |
+
"learning_rate": 8.935817307692309e-06,
|
| 185152 |
+
"loss": 0.3311,
|
| 185153 |
+
"step": 67430
|
| 185154 |
+
},
|
| 185155 |
+
{
|
| 185156 |
+
"epoch": 542.9,
|
| 185157 |
+
"learning_rate": 8.93573717948718e-06,
|
| 185158 |
+
"loss": 0.377,
|
| 185159 |
+
"step": 67435
|
| 185160 |
+
},
|
| 185161 |
+
{
|
| 185162 |
+
"epoch": 542.94,
|
| 185163 |
+
"learning_rate": 8.935657051282052e-06,
|
| 185164 |
+
"loss": 0.3752,
|
| 185165 |
+
"step": 67440
|
| 185166 |
+
},
|
| 185167 |
+
{
|
| 185168 |
+
"epoch": 542.98,
|
| 185169 |
+
"learning_rate": 8.935576923076925e-06,
|
| 185170 |
+
"loss": 0.5475,
|
| 185171 |
+
"step": 67445
|
| 185172 |
+
},
|
| 185173 |
+
{
|
| 185174 |
+
"epoch": 543.0,
|
| 185175 |
+
"eval_loss": 0.35275718569755554,
|
| 185176 |
+
"eval_runtime": 40.6597,
|
| 185177 |
+
"eval_samples_per_second": 20.635,
|
| 185178 |
+
"eval_steps_per_second": 0.664,
|
| 185179 |
+
"eval_wer": 0.1855363713557883,
|
| 185180 |
+
"step": 67447
|
| 185181 |
}
|
| 185182 |
],
|
| 185183 |
"max_steps": 620000,
|
| 185184 |
"num_train_epochs": 5000,
|
| 185185 |
+
"total_flos": 1.897958418039559e+20,
|
| 185186 |
"trial_name": null,
|
| 185187 |
"trial_params": null
|
| 185188 |
}
|
model-bin/finetune/base/{checkpoint-60729 β checkpoint-67447}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-66949 β checkpoint-67572}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-60729 β checkpoint-67572}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:cad4194b8552407741480b0d7cfec2bbde02181f458dbb03ec6afcfecdd08026
|
| 3 |
+
size 722165393
|
model-bin/finetune/base/{checkpoint-66949 β checkpoint-67572}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-60729 β checkpoint-67572}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:43eaa79891cd0d56e96b0fc5b84927dae77481a7bceab190190510d141ccdbf1
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-60729 β checkpoint-67572}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:412a8b45bf8449c4c28a263b3f9f3c718a03c421d731fa385f8caba49e11373c
|
| 3 |
size 14503
|
model-bin/finetune/base/{checkpoint-60729 β checkpoint-67572}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b996ed8da366dae97a015b33368c0ecf067f550c8dd285673547842926659440
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-66949 β checkpoint-67572}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1f3db58c27f09014af98cc064e08602c5304eb789e6d0bbd26dfcf9aa36490f9
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-60729 β checkpoint-67572}/trainer_state.json
RENAMED
|
The diff for this file is too large to render.
See raw diff
|
|
|
model-bin/finetune/base/{checkpoint-66949 β checkpoint-67572}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1629834135.9336932/events.out.tfevents.1629834135.c435e1c5ee04.920.211
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6b150f7b0958b1971f10d638538ea0ff8131c4ffb78d9e0d29a30e05ea5c52e8
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629834756.205052/events.out.tfevents.1629834756.c435e1c5ee04.920.213
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b348d2848eadf4eec5cf10700dcda79d2f33f7a21030d5e409d510173556f0cb
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629835507.9034107/events.out.tfevents.1629835507.c435e1c5ee04.920.215
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:90a811cd4ee486b70c250cef428dcc8561cfcdfb386529528d2a36845aa3f4d0
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629836153.2164783/events.out.tfevents.1629836153.c435e1c5ee04.920.217
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4e81be56ed4378a407beed856272baa7fc03446b3b2d110ff221d5a9466f9e29
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629836799.1113734/events.out.tfevents.1629836799.c435e1c5ee04.920.219
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9240c94e3aeed04b7627a11dcbce62cfe96fe7a1ef7303fe2a4719c4010644fb
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1629834135.c435e1c5ee04.920.210
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:630db7e7ef2a4b28bb5ed551f2405e52100e6a95680cfedb7f7e52c850f61cab
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629834756.c435e1c5ee04.920.212
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1482b5d3d281d592a7479c3d20581ff59351aabc483c6c8017660f1e9b7be927
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629835507.c435e1c5ee04.920.214
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:189290b987e9b5fd88e38a209ed29192b4462584bcb030abfe8804f45d1f6220
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629836153.c435e1c5ee04.920.216
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:aabad5cc92ba027d433b2116f6f914a5b311adf64157de040def01c7c40781f8
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629836799.c435e1c5ee04.920.218
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1e66806bcc547d4ad407e6b0a5a655f795f3facbc7f5cebcc52173308532ef49
|
| 3 |
+
size 8622
|