"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-46666 β checkpoint-52515}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-46666 β checkpoint-52515}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-46666 β checkpoint-52515}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-52015 β checkpoint-52515}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-52015 β checkpoint-52515}/rng_state.pth +2 -2
- model-bin/finetune/base/{checkpoint-52015 β checkpoint-52515}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-46666 β checkpoint-52515}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-52015 β checkpoint-52515}/trainer_state.json +642 -6
- model-bin/finetune/base/{checkpoint-46666 β checkpoint-52515}/training_args.bin +0 -0
- model-bin/finetune/base/{checkpoint-52015 β checkpoint-52639}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-52015 β checkpoint-52639}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-52015 β checkpoint-52639}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-46666 β checkpoint-52639}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-46666 β checkpoint-52639}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-46666 β checkpoint-52639}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-52015 β checkpoint-52639}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-46666 β checkpoint-52639}/trainer_state.json +0 -0
- model-bin/finetune/base/{checkpoint-52015 β checkpoint-52639}/training_args.bin +0 -0
- model-bin/finetune/base/log/1629752462.5473433/events.out.tfevents.1629752462.74272264b15c.932.223 +3 -0
- model-bin/finetune/base/log/1629753109.9928777/events.out.tfevents.1629753109.74272264b15c.932.225 +3 -0
- model-bin/finetune/base/log/1629753750.9052882/events.out.tfevents.1629753750.74272264b15c.932.227 +3 -0
- model-bin/finetune/base/log/1629754385.4896328/events.out.tfevents.1629754385.74272264b15c.932.229 +3 -0
- model-bin/finetune/base/log/1629755043.4480946/events.out.tfevents.1629755043.74272264b15c.932.231 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629752462.74272264b15c.932.222 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629753109.74272264b15c.932.224 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629753750.74272264b15c.932.226 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629754385.74272264b15c.932.228 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629755043.74272264b15c.932.230 +3 -0
model-bin/finetune/base/{checkpoint-46666 β checkpoint-52515}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-46666 β checkpoint-52515}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165009
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:07ad5e8e80074832caa0a98b5c08f323f432edcdb3b2a365d123371f20e8beda
|
| 3 |
size 722165009
|
model-bin/finetune/base/{checkpoint-46666 β checkpoint-52515}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-52015 β checkpoint-52515}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4ad7bfa19a307825777f0d973e4d3894ebe0ab9f5e474e9337514096826b79b2
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-52015 β checkpoint-52515}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:df6b787e2064a641feeeec680cf7d039f7d81c7e51db0009bb53c63dd07dabf9
|
| 3 |
+
size 14567
|
model-bin/finetune/base/{checkpoint-52015 β checkpoint-52515}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1a85bd7e9d0b1ebc11f5f5c5604858f3dbf9d33e447f24534a89df3b0a46a7a1
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-46666 β checkpoint-52515}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:675296120f77ac95e3ef252eb671e40c28127966b4388d2ca4ab32f9c73d4f1a
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-52015 β checkpoint-52515}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
-
"best_metric": 0.
|
| 3 |
-
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -165546,11 +165546,647 @@
|
|
| 165546 |
"eval_steps_per_second": 0.666,
|
| 165547 |
"eval_wer": 0.20132061628760087,
|
| 165548 |
"step": 52015
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 165549 |
}
|
| 165550 |
],
|
| 165551 |
-
"max_steps":
|
| 165552 |
"num_train_epochs": 5000,
|
| 165553 |
-
"total_flos": 1.
|
| 165554 |
"trial_name": null,
|
| 165555 |
"trial_params": null
|
| 165556 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"best_metric": 0.18992848189928482,
|
| 3 |
+
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-52515",
|
| 4 |
+
"epoch": 420.0,
|
| 5 |
+
"global_step": 52515,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 165546 |
"eval_steps_per_second": 0.666,
|
| 165547 |
"eval_wer": 0.20132061628760087,
|
| 165548 |
"step": 52015
|
| 165549 |
+
},
|
| 165550 |
+
{
|
| 165551 |
+
"epoch": 416.04,
|
| 165552 |
+
"learning_rate": 9.182676282051282e-06,
|
| 165553 |
+
"loss": 0.3158,
|
| 165554 |
+
"step": 52020
|
| 165555 |
+
},
|
| 165556 |
+
{
|
| 165557 |
+
"epoch": 416.08,
|
| 165558 |
+
"learning_rate": 9.182596153846155e-06,
|
| 165559 |
+
"loss": 0.3036,
|
| 165560 |
+
"step": 52025
|
| 165561 |
+
},
|
| 165562 |
+
{
|
| 165563 |
+
"epoch": 416.12,
|
| 165564 |
+
"learning_rate": 9.182516025641027e-06,
|
| 165565 |
+
"loss": 0.339,
|
| 165566 |
+
"step": 52030
|
| 165567 |
+
},
|
| 165568 |
+
{
|
| 165569 |
+
"epoch": 416.16,
|
| 165570 |
+
"learning_rate": 9.182435897435898e-06,
|
| 165571 |
+
"loss": 0.4548,
|
| 165572 |
+
"step": 52035
|
| 165573 |
+
},
|
| 165574 |
+
{
|
| 165575 |
+
"epoch": 416.2,
|
| 165576 |
+
"learning_rate": 9.18235576923077e-06,
|
| 165577 |
+
"loss": 1.1986,
|
| 165578 |
+
"step": 52040
|
| 165579 |
+
},
|
| 165580 |
+
{
|
| 165581 |
+
"epoch": 416.24,
|
| 165582 |
+
"learning_rate": 9.182275641025642e-06,
|
| 165583 |
+
"loss": 0.4446,
|
| 165584 |
+
"step": 52045
|
| 165585 |
+
},
|
| 165586 |
+
{
|
| 165587 |
+
"epoch": 416.28,
|
| 165588 |
+
"learning_rate": 9.182195512820514e-06,
|
| 165589 |
+
"loss": 0.3353,
|
| 165590 |
+
"step": 52050
|
| 165591 |
+
},
|
| 165592 |
+
{
|
| 165593 |
+
"epoch": 416.32,
|
| 165594 |
+
"learning_rate": 9.182115384615385e-06,
|
| 165595 |
+
"loss": 0.3124,
|
| 165596 |
+
"step": 52055
|
| 165597 |
+
},
|
| 165598 |
+
{
|
| 165599 |
+
"epoch": 416.36,
|
| 165600 |
+
"learning_rate": 9.182035256410257e-06,
|
| 165601 |
+
"loss": 0.5198,
|
| 165602 |
+
"step": 52060
|
| 165603 |
+
},
|
| 165604 |
+
{
|
| 165605 |
+
"epoch": 416.4,
|
| 165606 |
+
"learning_rate": 9.18195512820513e-06,
|
| 165607 |
+
"loss": 1.2901,
|
| 165608 |
+
"step": 52065
|
| 165609 |
+
},
|
| 165610 |
+
{
|
| 165611 |
+
"epoch": 416.44,
|
| 165612 |
+
"learning_rate": 9.181875000000001e-06,
|
| 165613 |
+
"loss": 0.3501,
|
| 165614 |
+
"step": 52070
|
| 165615 |
+
},
|
| 165616 |
+
{
|
| 165617 |
+
"epoch": 416.48,
|
| 165618 |
+
"learning_rate": 9.181794871794872e-06,
|
| 165619 |
+
"loss": 0.3317,
|
| 165620 |
+
"step": 52075
|
| 165621 |
+
},
|
| 165622 |
+
{
|
| 165623 |
+
"epoch": 416.52,
|
| 165624 |
+
"learning_rate": 9.181714743589745e-06,
|
| 165625 |
+
"loss": 0.3237,
|
| 165626 |
+
"step": 52080
|
| 165627 |
+
},
|
| 165628 |
+
{
|
| 165629 |
+
"epoch": 416.56,
|
| 165630 |
+
"learning_rate": 9.181634615384617e-06,
|
| 165631 |
+
"loss": 0.5543,
|
| 165632 |
+
"step": 52085
|
| 165633 |
+
},
|
| 165634 |
+
{
|
| 165635 |
+
"epoch": 416.6,
|
| 165636 |
+
"learning_rate": 9.181554487179488e-06,
|
| 165637 |
+
"loss": 1.2754,
|
| 165638 |
+
"step": 52090
|
| 165639 |
+
},
|
| 165640 |
+
{
|
| 165641 |
+
"epoch": 416.64,
|
| 165642 |
+
"learning_rate": 9.18147435897436e-06,
|
| 165643 |
+
"loss": 0.3846,
|
| 165644 |
+
"step": 52095
|
| 165645 |
+
},
|
| 165646 |
+
{
|
| 165647 |
+
"epoch": 416.68,
|
| 165648 |
+
"learning_rate": 9.181394230769232e-06,
|
| 165649 |
+
"loss": 0.3265,
|
| 165650 |
+
"step": 52100
|
| 165651 |
+
},
|
| 165652 |
+
{
|
| 165653 |
+
"epoch": 416.72,
|
| 165654 |
+
"learning_rate": 9.181314102564102e-06,
|
| 165655 |
+
"loss": 0.3491,
|
| 165656 |
+
"step": 52105
|
| 165657 |
+
},
|
| 165658 |
+
{
|
| 165659 |
+
"epoch": 416.76,
|
| 165660 |
+
"learning_rate": 9.181233974358975e-06,
|
| 165661 |
+
"loss": 0.4898,
|
| 165662 |
+
"step": 52110
|
| 165663 |
+
},
|
| 165664 |
+
{
|
| 165665 |
+
"epoch": 416.8,
|
| 165666 |
+
"learning_rate": 9.181153846153848e-06,
|
| 165667 |
+
"loss": 1.1767,
|
| 165668 |
+
"step": 52115
|
| 165669 |
+
},
|
| 165670 |
+
{
|
| 165671 |
+
"epoch": 416.84,
|
| 165672 |
+
"learning_rate": 9.181073717948718e-06,
|
| 165673 |
+
"loss": 0.3844,
|
| 165674 |
+
"step": 52120
|
| 165675 |
+
},
|
| 165676 |
+
{
|
| 165677 |
+
"epoch": 416.88,
|
| 165678 |
+
"learning_rate": 9.180993589743591e-06,
|
| 165679 |
+
"loss": 0.3551,
|
| 165680 |
+
"step": 52125
|
| 165681 |
+
},
|
| 165682 |
+
{
|
| 165683 |
+
"epoch": 416.92,
|
| 165684 |
+
"learning_rate": 9.180913461538462e-06,
|
| 165685 |
+
"loss": 0.4132,
|
| 165686 |
+
"step": 52130
|
| 165687 |
+
},
|
| 165688 |
+
{
|
| 165689 |
+
"epoch": 416.96,
|
| 165690 |
+
"learning_rate": 9.180833333333334e-06,
|
| 165691 |
+
"loss": 0.5226,
|
| 165692 |
+
"step": 52135
|
| 165693 |
+
},
|
| 165694 |
+
{
|
| 165695 |
+
"epoch": 417.0,
|
| 165696 |
+
"learning_rate": 9.180753205128205e-06,
|
| 165697 |
+
"loss": 1.3741,
|
| 165698 |
+
"step": 52140
|
| 165699 |
+
},
|
| 165700 |
+
{
|
| 165701 |
+
"epoch": 417.0,
|
| 165702 |
+
"eval_loss": 0.45200973749160767,
|
| 165703 |
+
"eval_runtime": 39.6524,
|
| 165704 |
+
"eval_samples_per_second": 21.184,
|
| 165705 |
+
"eval_steps_per_second": 0.681,
|
| 165706 |
+
"eval_wer": 0.20303938043398845,
|
| 165707 |
+
"step": 52140
|
| 165708 |
+
},
|
| 165709 |
+
{
|
| 165710 |
+
"epoch": 417.04,
|
| 165711 |
+
"learning_rate": 9.180673076923078e-06,
|
| 165712 |
+
"loss": 0.4385,
|
| 165713 |
+
"step": 52145
|
| 165714 |
+
},
|
| 165715 |
+
{
|
| 165716 |
+
"epoch": 417.08,
|
| 165717 |
+
"learning_rate": 9.18059294871795e-06,
|
| 165718 |
+
"loss": 0.3828,
|
| 165719 |
+
"step": 52150
|
| 165720 |
+
},
|
| 165721 |
+
{
|
| 165722 |
+
"epoch": 417.12,
|
| 165723 |
+
"learning_rate": 9.18051282051282e-06,
|
| 165724 |
+
"loss": 0.2904,
|
| 165725 |
+
"step": 52155
|
| 165726 |
+
},
|
| 165727 |
+
{
|
| 165728 |
+
"epoch": 417.16,
|
| 165729 |
+
"learning_rate": 9.180432692307692e-06,
|
| 165730 |
+
"loss": 0.5224,
|
| 165731 |
+
"step": 52160
|
| 165732 |
+
},
|
| 165733 |
+
{
|
| 165734 |
+
"epoch": 417.2,
|
| 165735 |
+
"learning_rate": 9.180352564102565e-06,
|
| 165736 |
+
"loss": 1.3969,
|
| 165737 |
+
"step": 52165
|
| 165738 |
+
},
|
| 165739 |
+
{
|
| 165740 |
+
"epoch": 417.24,
|
| 165741 |
+
"learning_rate": 9.180272435897437e-06,
|
| 165742 |
+
"loss": 0.3446,
|
| 165743 |
+
"step": 52170
|
| 165744 |
+
},
|
| 165745 |
+
{
|
| 165746 |
+
"epoch": 417.28,
|
| 165747 |
+
"learning_rate": 9.180192307692308e-06,
|
| 165748 |
+
"loss": 0.3816,
|
| 165749 |
+
"step": 52175
|
| 165750 |
+
},
|
| 165751 |
+
{
|
| 165752 |
+
"epoch": 417.32,
|
| 165753 |
+
"learning_rate": 9.180112179487181e-06,
|
| 165754 |
+
"loss": 0.4264,
|
| 165755 |
+
"step": 52180
|
| 165756 |
+
},
|
| 165757 |
+
{
|
| 165758 |
+
"epoch": 417.36,
|
| 165759 |
+
"learning_rate": 9.180032051282052e-06,
|
| 165760 |
+
"loss": 0.4443,
|
| 165761 |
+
"step": 52185
|
| 165762 |
+
},
|
| 165763 |
+
{
|
| 165764 |
+
"epoch": 417.4,
|
| 165765 |
+
"learning_rate": 9.179951923076924e-06,
|
| 165766 |
+
"loss": 1.2463,
|
| 165767 |
+
"step": 52190
|
| 165768 |
+
},
|
| 165769 |
+
{
|
| 165770 |
+
"epoch": 417.44,
|
| 165771 |
+
"learning_rate": 9.179871794871795e-06,
|
| 165772 |
+
"loss": 0.4601,
|
| 165773 |
+
"step": 52195
|
| 165774 |
+
},
|
| 165775 |
+
{
|
| 165776 |
+
"epoch": 417.48,
|
| 165777 |
+
"learning_rate": 9.179791666666668e-06,
|
| 165778 |
+
"loss": 0.3195,
|
| 165779 |
+
"step": 52200
|
| 165780 |
+
},
|
| 165781 |
+
{
|
| 165782 |
+
"epoch": 417.52,
|
| 165783 |
+
"learning_rate": 9.17971153846154e-06,
|
| 165784 |
+
"loss": 0.3873,
|
| 165785 |
+
"step": 52205
|
| 165786 |
+
},
|
| 165787 |
+
{
|
| 165788 |
+
"epoch": 417.56,
|
| 165789 |
+
"learning_rate": 9.17963141025641e-06,
|
| 165790 |
+
"loss": 0.4696,
|
| 165791 |
+
"step": 52210
|
| 165792 |
+
},
|
| 165793 |
+
{
|
| 165794 |
+
"epoch": 417.6,
|
| 165795 |
+
"learning_rate": 9.179551282051284e-06,
|
| 165796 |
+
"loss": 1.338,
|
| 165797 |
+
"step": 52215
|
| 165798 |
+
},
|
| 165799 |
+
{
|
| 165800 |
+
"epoch": 417.64,
|
| 165801 |
+
"learning_rate": 9.179471153846155e-06,
|
| 165802 |
+
"loss": 0.3751,
|
| 165803 |
+
"step": 52220
|
| 165804 |
+
},
|
| 165805 |
+
{
|
| 165806 |
+
"epoch": 417.68,
|
| 165807 |
+
"learning_rate": 9.179391025641027e-06,
|
| 165808 |
+
"loss": 0.2995,
|
| 165809 |
+
"step": 52225
|
| 165810 |
+
},
|
| 165811 |
+
{
|
| 165812 |
+
"epoch": 417.72,
|
| 165813 |
+
"learning_rate": 9.179310897435898e-06,
|
| 165814 |
+
"loss": 0.4102,
|
| 165815 |
+
"step": 52230
|
| 165816 |
+
},
|
| 165817 |
+
{
|
| 165818 |
+
"epoch": 417.76,
|
| 165819 |
+
"learning_rate": 9.179230769230771e-06,
|
| 165820 |
+
"loss": 0.5121,
|
| 165821 |
+
"step": 52235
|
| 165822 |
+
},
|
| 165823 |
+
{
|
| 165824 |
+
"epoch": 417.8,
|
| 165825 |
+
"learning_rate": 9.17915064102564e-06,
|
| 165826 |
+
"loss": 1.2505,
|
| 165827 |
+
"step": 52240
|
| 165828 |
+
},
|
| 165829 |
+
{
|
| 165830 |
+
"epoch": 417.84,
|
| 165831 |
+
"learning_rate": 9.179070512820514e-06,
|
| 165832 |
+
"loss": 0.428,
|
| 165833 |
+
"step": 52245
|
| 165834 |
+
},
|
| 165835 |
+
{
|
| 165836 |
+
"epoch": 417.88,
|
| 165837 |
+
"learning_rate": 9.178990384615385e-06,
|
| 165838 |
+
"loss": 0.3032,
|
| 165839 |
+
"step": 52250
|
| 165840 |
+
},
|
| 165841 |
+
{
|
| 165842 |
+
"epoch": 417.92,
|
| 165843 |
+
"learning_rate": 9.178910256410256e-06,
|
| 165844 |
+
"loss": 0.3661,
|
| 165845 |
+
"step": 52255
|
| 165846 |
+
},
|
| 165847 |
+
{
|
| 165848 |
+
"epoch": 417.96,
|
| 165849 |
+
"learning_rate": 9.178830128205128e-06,
|
| 165850 |
+
"loss": 0.5023,
|
| 165851 |
+
"step": 52260
|
| 165852 |
+
},
|
| 165853 |
+
{
|
| 165854 |
+
"epoch": 418.0,
|
| 165855 |
+
"learning_rate": 9.17875e-06,
|
| 165856 |
+
"loss": 1.5163,
|
| 165857 |
+
"step": 52265
|
| 165858 |
+
},
|
| 165859 |
+
{
|
| 165860 |
+
"epoch": 418.0,
|
| 165861 |
+
"eval_loss": 0.421318918466568,
|
| 165862 |
+
"eval_runtime": 37.5508,
|
| 165863 |
+
"eval_samples_per_second": 22.37,
|
| 165864 |
+
"eval_steps_per_second": 0.719,
|
| 165865 |
+
"eval_wer": 0.1988814317673378,
|
| 165866 |
+
"step": 52265
|
| 165867 |
+
},
|
| 165868 |
+
{
|
| 165869 |
+
"epoch": 418.04,
|
| 165870 |
+
"learning_rate": 9.178669871794872e-06,
|
| 165871 |
+
"loss": 0.4194,
|
| 165872 |
+
"step": 52270
|
| 165873 |
+
},
|
| 165874 |
+
{
|
| 165875 |
+
"epoch": 418.08,
|
| 165876 |
+
"learning_rate": 9.178589743589744e-06,
|
| 165877 |
+
"loss": 0.3268,
|
| 165878 |
+
"step": 52275
|
| 165879 |
+
},
|
| 165880 |
+
{
|
| 165881 |
+
"epoch": 418.12,
|
| 165882 |
+
"learning_rate": 9.178509615384617e-06,
|
| 165883 |
+
"loss": 0.3918,
|
| 165884 |
+
"step": 52280
|
| 165885 |
+
},
|
| 165886 |
+
{
|
| 165887 |
+
"epoch": 418.16,
|
| 165888 |
+
"learning_rate": 9.178429487179488e-06,
|
| 165889 |
+
"loss": 0.5551,
|
| 165890 |
+
"step": 52285
|
| 165891 |
+
},
|
| 165892 |
+
{
|
| 165893 |
+
"epoch": 418.2,
|
| 165894 |
+
"learning_rate": 9.17834935897436e-06,
|
| 165895 |
+
"loss": 1.2501,
|
| 165896 |
+
"step": 52290
|
| 165897 |
+
},
|
| 165898 |
+
{
|
| 165899 |
+
"epoch": 418.24,
|
| 165900 |
+
"learning_rate": 9.17826923076923e-06,
|
| 165901 |
+
"loss": 0.4417,
|
| 165902 |
+
"step": 52295
|
| 165903 |
+
},
|
| 165904 |
+
{
|
| 165905 |
+
"epoch": 418.28,
|
| 165906 |
+
"learning_rate": 9.178189102564104e-06,
|
| 165907 |
+
"loss": 0.3786,
|
| 165908 |
+
"step": 52300
|
| 165909 |
+
},
|
| 165910 |
+
{
|
| 165911 |
+
"epoch": 418.32,
|
| 165912 |
+
"learning_rate": 9.178108974358975e-06,
|
| 165913 |
+
"loss": 0.3834,
|
| 165914 |
+
"step": 52305
|
| 165915 |
+
},
|
| 165916 |
+
{
|
| 165917 |
+
"epoch": 418.36,
|
| 165918 |
+
"learning_rate": 9.178028846153846e-06,
|
| 165919 |
+
"loss": 0.4986,
|
| 165920 |
+
"step": 52310
|
| 165921 |
+
},
|
| 165922 |
+
{
|
| 165923 |
+
"epoch": 418.4,
|
| 165924 |
+
"learning_rate": 9.17794871794872e-06,
|
| 165925 |
+
"loss": 1.1479,
|
| 165926 |
+
"step": 52315
|
| 165927 |
+
},
|
| 165928 |
+
{
|
| 165929 |
+
"epoch": 418.44,
|
| 165930 |
+
"learning_rate": 9.177868589743591e-06,
|
| 165931 |
+
"loss": 0.3841,
|
| 165932 |
+
"step": 52320
|
| 165933 |
+
},
|
| 165934 |
+
{
|
| 165935 |
+
"epoch": 418.48,
|
| 165936 |
+
"learning_rate": 9.177788461538462e-06,
|
| 165937 |
+
"loss": 0.3993,
|
| 165938 |
+
"step": 52325
|
| 165939 |
+
},
|
| 165940 |
+
{
|
| 165941 |
+
"epoch": 418.52,
|
| 165942 |
+
"learning_rate": 9.177708333333334e-06,
|
| 165943 |
+
"loss": 0.3836,
|
| 165944 |
+
"step": 52330
|
| 165945 |
+
},
|
| 165946 |
+
{
|
| 165947 |
+
"epoch": 418.56,
|
| 165948 |
+
"learning_rate": 9.177628205128207e-06,
|
| 165949 |
+
"loss": 0.5145,
|
| 165950 |
+
"step": 52335
|
| 165951 |
+
},
|
| 165952 |
+
{
|
| 165953 |
+
"epoch": 418.6,
|
| 165954 |
+
"learning_rate": 9.177548076923078e-06,
|
| 165955 |
+
"loss": 1.4892,
|
| 165956 |
+
"step": 52340
|
| 165957 |
+
},
|
| 165958 |
+
{
|
| 165959 |
+
"epoch": 418.64,
|
| 165960 |
+
"learning_rate": 9.17746794871795e-06,
|
| 165961 |
+
"loss": 0.3854,
|
| 165962 |
+
"step": 52345
|
| 165963 |
+
},
|
| 165964 |
+
{
|
| 165965 |
+
"epoch": 418.68,
|
| 165966 |
+
"learning_rate": 9.17738782051282e-06,
|
| 165967 |
+
"loss": 0.5018,
|
| 165968 |
+
"step": 52350
|
| 165969 |
+
},
|
| 165970 |
+
{
|
| 165971 |
+
"epoch": 418.72,
|
| 165972 |
+
"learning_rate": 9.177307692307694e-06,
|
| 165973 |
+
"loss": 0.379,
|
| 165974 |
+
"step": 52355
|
| 165975 |
+
},
|
| 165976 |
+
{
|
| 165977 |
+
"epoch": 418.76,
|
| 165978 |
+
"learning_rate": 9.177227564102565e-06,
|
| 165979 |
+
"loss": 0.4675,
|
| 165980 |
+
"step": 52360
|
| 165981 |
+
},
|
| 165982 |
+
{
|
| 165983 |
+
"epoch": 418.8,
|
| 165984 |
+
"learning_rate": 9.177147435897436e-06,
|
| 165985 |
+
"loss": 1.1614,
|
| 165986 |
+
"step": 52365
|
| 165987 |
+
},
|
| 165988 |
+
{
|
| 165989 |
+
"epoch": 418.84,
|
| 165990 |
+
"learning_rate": 9.17706730769231e-06,
|
| 165991 |
+
"loss": 0.6507,
|
| 165992 |
+
"step": 52370
|
| 165993 |
+
},
|
| 165994 |
+
{
|
| 165995 |
+
"epoch": 418.88,
|
| 165996 |
+
"learning_rate": 9.176987179487181e-06,
|
| 165997 |
+
"loss": 0.357,
|
| 165998 |
+
"step": 52375
|
| 165999 |
+
},
|
| 166000 |
+
{
|
| 166001 |
+
"epoch": 418.92,
|
| 166002 |
+
"learning_rate": 9.176907051282052e-06,
|
| 166003 |
+
"loss": 0.3412,
|
| 166004 |
+
"step": 52380
|
| 166005 |
+
},
|
| 166006 |
+
{
|
| 166007 |
+
"epoch": 418.96,
|
| 166008 |
+
"learning_rate": 9.176826923076924e-06,
|
| 166009 |
+
"loss": 0.5797,
|
| 166010 |
+
"step": 52385
|
| 166011 |
+
},
|
| 166012 |
+
{
|
| 166013 |
+
"epoch": 419.0,
|
| 166014 |
+
"learning_rate": 9.176746794871797e-06,
|
| 166015 |
+
"loss": 1.4001,
|
| 166016 |
+
"step": 52390
|
| 166017 |
+
},
|
| 166018 |
+
{
|
| 166019 |
+
"epoch": 419.0,
|
| 166020 |
+
"eval_loss": 0.4143487215042114,
|
| 166021 |
+
"eval_runtime": 39.3714,
|
| 166022 |
+
"eval_samples_per_second": 21.335,
|
| 166023 |
+
"eval_steps_per_second": 0.686,
|
| 166024 |
+
"eval_wer": 0.2074479579341641,
|
| 166025 |
+
"step": 52390
|
| 166026 |
+
},
|
| 166027 |
+
{
|
| 166028 |
+
"epoch": 419.04,
|
| 166029 |
+
"learning_rate": 9.176666666666666e-06,
|
| 166030 |
+
"loss": 0.4064,
|
| 166031 |
+
"step": 52395
|
| 166032 |
+
},
|
| 166033 |
+
{
|
| 166034 |
+
"epoch": 419.08,
|
| 166035 |
+
"learning_rate": 9.17658653846154e-06,
|
| 166036 |
+
"loss": 0.4067,
|
| 166037 |
+
"step": 52400
|
| 166038 |
+
},
|
| 166039 |
+
{
|
| 166040 |
+
"epoch": 419.12,
|
| 166041 |
+
"learning_rate": 9.17650641025641e-06,
|
| 166042 |
+
"loss": 0.3486,
|
| 166043 |
+
"step": 52405
|
| 166044 |
+
},
|
| 166045 |
+
{
|
| 166046 |
+
"epoch": 419.16,
|
| 166047 |
+
"learning_rate": 9.176426282051282e-06,
|
| 166048 |
+
"loss": 0.4744,
|
| 166049 |
+
"step": 52410
|
| 166050 |
+
},
|
| 166051 |
+
{
|
| 166052 |
+
"epoch": 419.2,
|
| 166053 |
+
"learning_rate": 9.176346153846155e-06,
|
| 166054 |
+
"loss": 1.2905,
|
| 166055 |
+
"step": 52415
|
| 166056 |
+
},
|
| 166057 |
+
{
|
| 166058 |
+
"epoch": 419.24,
|
| 166059 |
+
"learning_rate": 9.176266025641027e-06,
|
| 166060 |
+
"loss": 0.3739,
|
| 166061 |
+
"step": 52420
|
| 166062 |
+
},
|
| 166063 |
+
{
|
| 166064 |
+
"epoch": 419.28,
|
| 166065 |
+
"learning_rate": 9.176185897435898e-06,
|
| 166066 |
+
"loss": 0.3682,
|
| 166067 |
+
"step": 52425
|
| 166068 |
+
},
|
| 166069 |
+
{
|
| 166070 |
+
"epoch": 419.32,
|
| 166071 |
+
"learning_rate": 9.17610576923077e-06,
|
| 166072 |
+
"loss": 0.3172,
|
| 166073 |
+
"step": 52430
|
| 166074 |
+
},
|
| 166075 |
+
{
|
| 166076 |
+
"epoch": 419.36,
|
| 166077 |
+
"learning_rate": 9.176025641025642e-06,
|
| 166078 |
+
"loss": 0.5588,
|
| 166079 |
+
"step": 52435
|
| 166080 |
+
},
|
| 166081 |
+
{
|
| 166082 |
+
"epoch": 419.4,
|
| 166083 |
+
"learning_rate": 9.175945512820514e-06,
|
| 166084 |
+
"loss": 1.3292,
|
| 166085 |
+
"step": 52440
|
| 166086 |
+
},
|
| 166087 |
+
{
|
| 166088 |
+
"epoch": 419.44,
|
| 166089 |
+
"learning_rate": 9.175865384615385e-06,
|
| 166090 |
+
"loss": 0.3841,
|
| 166091 |
+
"step": 52445
|
| 166092 |
+
},
|
| 166093 |
+
{
|
| 166094 |
+
"epoch": 419.48,
|
| 166095 |
+
"learning_rate": 9.175785256410256e-06,
|
| 166096 |
+
"loss": 0.3479,
|
| 166097 |
+
"step": 52450
|
| 166098 |
+
},
|
| 166099 |
+
{
|
| 166100 |
+
"epoch": 419.52,
|
| 166101 |
+
"learning_rate": 9.17570512820513e-06,
|
| 166102 |
+
"loss": 0.3379,
|
| 166103 |
+
"step": 52455
|
| 166104 |
+
},
|
| 166105 |
+
{
|
| 166106 |
+
"epoch": 419.56,
|
| 166107 |
+
"learning_rate": 9.175625e-06,
|
| 166108 |
+
"loss": 0.5808,
|
| 166109 |
+
"step": 52460
|
| 166110 |
+
},
|
| 166111 |
+
{
|
| 166112 |
+
"epoch": 419.6,
|
| 166113 |
+
"learning_rate": 9.175544871794872e-06,
|
| 166114 |
+
"loss": 1.3315,
|
| 166115 |
+
"step": 52465
|
| 166116 |
+
},
|
| 166117 |
+
{
|
| 166118 |
+
"epoch": 419.64,
|
| 166119 |
+
"learning_rate": 9.175464743589745e-06,
|
| 166120 |
+
"loss": 0.5561,
|
| 166121 |
+
"step": 52470
|
| 166122 |
+
},
|
| 166123 |
+
{
|
| 166124 |
+
"epoch": 419.68,
|
| 166125 |
+
"learning_rate": 9.175384615384617e-06,
|
| 166126 |
+
"loss": 0.3038,
|
| 166127 |
+
"step": 52475
|
| 166128 |
+
},
|
| 166129 |
+
{
|
| 166130 |
+
"epoch": 419.72,
|
| 166131 |
+
"learning_rate": 9.175304487179488e-06,
|
| 166132 |
+
"loss": 0.2846,
|
| 166133 |
+
"step": 52480
|
| 166134 |
+
},
|
| 166135 |
+
{
|
| 166136 |
+
"epoch": 419.76,
|
| 166137 |
+
"learning_rate": 9.17522435897436e-06,
|
| 166138 |
+
"loss": 0.5519,
|
| 166139 |
+
"step": 52485
|
| 166140 |
+
},
|
| 166141 |
+
{
|
| 166142 |
+
"epoch": 419.8,
|
| 166143 |
+
"learning_rate": 9.175144230769232e-06,
|
| 166144 |
+
"loss": 1.2826,
|
| 166145 |
+
"step": 52490
|
| 166146 |
+
},
|
| 166147 |
+
{
|
| 166148 |
+
"epoch": 419.84,
|
| 166149 |
+
"learning_rate": 9.175064102564104e-06,
|
| 166150 |
+
"loss": 0.4148,
|
| 166151 |
+
"step": 52495
|
| 166152 |
+
},
|
| 166153 |
+
{
|
| 166154 |
+
"epoch": 419.88,
|
| 166155 |
+
"learning_rate": 9.174983974358975e-06,
|
| 166156 |
+
"loss": 0.3156,
|
| 166157 |
+
"step": 52500
|
| 166158 |
+
},
|
| 166159 |
+
{
|
| 166160 |
+
"epoch": 419.92,
|
| 166161 |
+
"learning_rate": 9.174903846153846e-06,
|
| 166162 |
+
"loss": 0.3902,
|
| 166163 |
+
"step": 52505
|
| 166164 |
+
},
|
| 166165 |
+
{
|
| 166166 |
+
"epoch": 419.96,
|
| 166167 |
+
"learning_rate": 9.17482371794872e-06,
|
| 166168 |
+
"loss": 0.571,
|
| 166169 |
+
"step": 52510
|
| 166170 |
+
},
|
| 166171 |
+
{
|
| 166172 |
+
"epoch": 420.0,
|
| 166173 |
+
"learning_rate": 9.17474358974359e-06,
|
| 166174 |
+
"loss": 1.6161,
|
| 166175 |
+
"step": 52515
|
| 166176 |
+
},
|
| 166177 |
+
{
|
| 166178 |
+
"epoch": 420.0,
|
| 166179 |
+
"eval_loss": 0.40050163865089417,
|
| 166180 |
+
"eval_runtime": 38.5955,
|
| 166181 |
+
"eval_samples_per_second": 21.764,
|
| 166182 |
+
"eval_steps_per_second": 0.7,
|
| 166183 |
+
"eval_wer": 0.18992848189928482,
|
| 166184 |
+
"step": 52515
|
| 166185 |
}
|
| 166186 |
],
|
| 166187 |
+
"max_steps": 625000,
|
| 166188 |
"num_train_epochs": 5000,
|
| 166189 |
+
"total_flos": 1.4776861135717535e+20,
|
| 166190 |
"trial_name": null,
|
| 166191 |
"trial_params": null
|
| 166192 |
}
|
model-bin/finetune/base/{checkpoint-46666 β checkpoint-52515}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-52015 β checkpoint-52639}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-52015 β checkpoint-52639}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165009
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1e6078b1a25bbfb93877b366a383e3230a0601ad5cee0d6d81e8efa4520c3950
|
| 3 |
size 722165009
|
model-bin/finetune/base/{checkpoint-52015 β checkpoint-52639}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-46666 β checkpoint-52639}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5c21d02d9671d703b466f4b6d46fb6b63007714a088428656792c5d6dd669f30
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-46666 β checkpoint-52639}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:31b6a4c636c3176580ba417b7373c6cfc166eb27e1434eb9258492ab9053129e
|
| 3 |
size 14503
|
model-bin/finetune/base/{checkpoint-46666 β checkpoint-52639}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ec8225ca5a3e0ec39b9289e48fc0dc88985a398292b7dfa2abaa0c27d66f24e5
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-52015 β checkpoint-52639}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8186c12820d92b2b7071f4326ea5609e06d88ce375b7275fe1a132a598badd19
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-46666 β checkpoint-52639}/trainer_state.json
RENAMED
|
The diff for this file is too large to render.
See raw diff
|
|
|
model-bin/finetune/base/{checkpoint-52015 β checkpoint-52639}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1629752462.5473433/events.out.tfevents.1629752462.74272264b15c.932.223
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c489f511d82915b4e7f88bf4eb3211dc8aae892ff0910bd9ff5ee8a195dacadd
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629753109.9928777/events.out.tfevents.1629753109.74272264b15c.932.225
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e26ccb6cb1eeea2c72c6474688ab33f0b5ae0a460f8f963c103e5a876c73149a
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629753750.9052882/events.out.tfevents.1629753750.74272264b15c.932.227
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7b209c39007ce8023b09151816d066f42d4cb4348b104fc2923d04a11fe3de6e
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629754385.4896328/events.out.tfevents.1629754385.74272264b15c.932.229
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0b0beb0b33aca7231ecea9d3317ed34eec395d17dc2436a0b32b4e3ffbf6a94c
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629755043.4480946/events.out.tfevents.1629755043.74272264b15c.932.231
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dd61fe9a26c7e4bc4d3b9c61e42fc30f39d40646f8b419baee69f34f6b430824
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1629752462.74272264b15c.932.222
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:755a878e8436f6d5c0096545a510104544a3393d3db96b16f5e99824eeb87cd4
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629753109.74272264b15c.932.224
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ad11b22666286ca1d9937d8e0d88244eb72bb28b6aaa8f29ba17ca2b614dd526
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629753750.74272264b15c.932.226
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:38e53c9f9d3830c0ac7b059a91db4c4b8433d5eb61fa8cc0eecfa53b3401c7ee
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629754385.74272264b15c.932.228
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:645c61ea40c25a843bf499a1a4de5d29800abb44c1a3baf40c5046e3159abf49
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629755043.74272264b15c.932.230
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9d0fe88ffe689f196442935fd701efe5d0c8313eba581eea66cd7ff052819ad9
|
| 3 |
+
size 8462
|