"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-55501 β checkpoint-60604}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-55501 β checkpoint-60604}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-55501 β checkpoint-60604}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-60105 β checkpoint-60604}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-60105 β checkpoint-60604}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-60105 β checkpoint-60604}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-55501 β checkpoint-60604}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-60105 β checkpoint-60604}/trainer_state.json +633 -3
- model-bin/finetune/base/{checkpoint-55501 β checkpoint-60604}/training_args.bin +0 -0
- model-bin/finetune/base/{checkpoint-60105 β checkpoint-60729}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-60105 β checkpoint-60729}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-60105 β checkpoint-60729}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-55501 β checkpoint-60729}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-55501 β checkpoint-60729}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-55501 β checkpoint-60729}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-60105 β checkpoint-60729}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-55501 β checkpoint-60729}/trainer_state.json +0 -0
- model-bin/finetune/base/{checkpoint-60105 β checkpoint-60729}/training_args.bin +0 -0
- model-bin/finetune/base/log/1629798286.419865/events.out.tfevents.1629798286.c435e1c5ee04.920.101 +3 -0
- model-bin/finetune/base/log/1629798930.4451573/events.out.tfevents.1629798930.c435e1c5ee04.920.103 +3 -0
- model-bin/finetune/base/log/1629799575.228714/events.out.tfevents.1629799575.c435e1c5ee04.920.105 +3 -0
- model-bin/finetune/base/log/1629800218.038527/events.out.tfevents.1629800218.c435e1c5ee04.920.107 +3 -0
- model-bin/finetune/base/log/1629800871.1235416/events.out.tfevents.1629800871.c435e1c5ee04.920.109 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629798286.c435e1c5ee04.920.100 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629798930.c435e1c5ee04.920.102 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629799575.c435e1c5ee04.920.104 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629800218.c435e1c5ee04.920.106 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629800871.c435e1c5ee04.920.108 +3 -0
model-bin/finetune/base/{checkpoint-55501 β checkpoint-60604}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-55501 β checkpoint-60604}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165009
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:54d742d39c8446eace3eb84f293d54d3c92367bac539b2c6aac2610c3b5cd216
|
| 3 |
size 722165009
|
model-bin/finetune/base/{checkpoint-55501 β checkpoint-60604}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-60105 β checkpoint-60604}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b5929eb300a5e7bc9be64e52e4b06762401ab913925a2a1b0e17dcaad3e94476
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-60105 β checkpoint-60604}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6c1a87fab1e51e42d80278684e7ba62284a23bae14afa60f95e959e29d7d01e1
|
| 3 |
size 14503
|
model-bin/finetune/base/{checkpoint-60105 β checkpoint-60604}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:789e2be1e3c06a3e2bb8e0d942a5bad1e52db387c9c9778d8d055b91e46a0dc3
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-55501 β checkpoint-60604}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d917b2672d75e0efb76bc510eaa2d3141b6f755a5d5799309de61ece06544b7f
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-60105 β checkpoint-60604}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.18631571186315712,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-55501",
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -175839,11 +175839,641 @@
|
|
| 175839 |
"eval_steps_per_second": 0.68,
|
| 175840 |
"eval_wer": 0.19872743415211602,
|
| 175841 |
"step": 60105
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 175842 |
}
|
| 175843 |
],
|
| 175844 |
"max_steps": 620000,
|
| 175845 |
"num_train_epochs": 5000,
|
| 175846 |
-
"total_flos": 1.
|
| 175847 |
"trial_name": null,
|
| 175848 |
"trial_params": null
|
| 175849 |
}
|
|
|
|
| 1 |
{
|
| 2 |
"best_metric": 0.18631571186315712,
|
| 3 |
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-55501",
|
| 4 |
+
"epoch": 488.0,
|
| 5 |
+
"global_step": 60604,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 175839 |
"eval_steps_per_second": 0.68,
|
| 175840 |
"eval_wer": 0.19872743415211602,
|
| 175841 |
"step": 60105
|
| 175842 |
+
},
|
| 175843 |
+
{
|
| 175844 |
+
"epoch": 480.04,
|
| 175845 |
+
"learning_rate": 9.053076923076923e-06,
|
| 175846 |
+
"loss": 0.3265,
|
| 175847 |
+
"step": 60110
|
| 175848 |
+
},
|
| 175849 |
+
{
|
| 175850 |
+
"epoch": 480.08,
|
| 175851 |
+
"learning_rate": 9.052996794871796e-06,
|
| 175852 |
+
"loss": 0.29,
|
| 175853 |
+
"step": 60115
|
| 175854 |
+
},
|
| 175855 |
+
{
|
| 175856 |
+
"epoch": 480.12,
|
| 175857 |
+
"learning_rate": 9.052916666666668e-06,
|
| 175858 |
+
"loss": 0.3835,
|
| 175859 |
+
"step": 60120
|
| 175860 |
+
},
|
| 175861 |
+
{
|
| 175862 |
+
"epoch": 480.16,
|
| 175863 |
+
"learning_rate": 9.052836538461539e-06,
|
| 175864 |
+
"loss": 0.4613,
|
| 175865 |
+
"step": 60125
|
| 175866 |
+
},
|
| 175867 |
+
{
|
| 175868 |
+
"epoch": 480.2,
|
| 175869 |
+
"learning_rate": 9.052756410256412e-06,
|
| 175870 |
+
"loss": 1.1667,
|
| 175871 |
+
"step": 60130
|
| 175872 |
+
},
|
| 175873 |
+
{
|
| 175874 |
+
"epoch": 480.24,
|
| 175875 |
+
"learning_rate": 9.052676282051282e-06,
|
| 175876 |
+
"loss": 0.3625,
|
| 175877 |
+
"step": 60135
|
| 175878 |
+
},
|
| 175879 |
+
{
|
| 175880 |
+
"epoch": 480.28,
|
| 175881 |
+
"learning_rate": 9.052596153846155e-06,
|
| 175882 |
+
"loss": 0.3146,
|
| 175883 |
+
"step": 60140
|
| 175884 |
+
},
|
| 175885 |
+
{
|
| 175886 |
+
"epoch": 480.32,
|
| 175887 |
+
"learning_rate": 9.052516025641026e-06,
|
| 175888 |
+
"loss": 0.3837,
|
| 175889 |
+
"step": 60145
|
| 175890 |
+
},
|
| 175891 |
+
{
|
| 175892 |
+
"epoch": 480.36,
|
| 175893 |
+
"learning_rate": 9.052435897435898e-06,
|
| 175894 |
+
"loss": 0.5827,
|
| 175895 |
+
"step": 60150
|
| 175896 |
+
},
|
| 175897 |
+
{
|
| 175898 |
+
"epoch": 480.4,
|
| 175899 |
+
"learning_rate": 9.052355769230769e-06,
|
| 175900 |
+
"loss": 1.3282,
|
| 175901 |
+
"step": 60155
|
| 175902 |
+
},
|
| 175903 |
+
{
|
| 175904 |
+
"epoch": 480.44,
|
| 175905 |
+
"learning_rate": 9.052275641025642e-06,
|
| 175906 |
+
"loss": 0.3936,
|
| 175907 |
+
"step": 60160
|
| 175908 |
+
},
|
| 175909 |
+
{
|
| 175910 |
+
"epoch": 480.48,
|
| 175911 |
+
"learning_rate": 9.052195512820513e-06,
|
| 175912 |
+
"loss": 0.3122,
|
| 175913 |
+
"step": 60165
|
| 175914 |
+
},
|
| 175915 |
+
{
|
| 175916 |
+
"epoch": 480.52,
|
| 175917 |
+
"learning_rate": 9.052115384615385e-06,
|
| 175918 |
+
"loss": 0.3587,
|
| 175919 |
+
"step": 60170
|
| 175920 |
+
},
|
| 175921 |
+
{
|
| 175922 |
+
"epoch": 480.56,
|
| 175923 |
+
"learning_rate": 9.052035256410258e-06,
|
| 175924 |
+
"loss": 0.5124,
|
| 175925 |
+
"step": 60175
|
| 175926 |
+
},
|
| 175927 |
+
{
|
| 175928 |
+
"epoch": 480.6,
|
| 175929 |
+
"learning_rate": 9.051955128205129e-06,
|
| 175930 |
+
"loss": 1.1635,
|
| 175931 |
+
"step": 60180
|
| 175932 |
+
},
|
| 175933 |
+
{
|
| 175934 |
+
"epoch": 480.64,
|
| 175935 |
+
"learning_rate": 9.051875e-06,
|
| 175936 |
+
"loss": 0.3471,
|
| 175937 |
+
"step": 60185
|
| 175938 |
+
},
|
| 175939 |
+
{
|
| 175940 |
+
"epoch": 480.68,
|
| 175941 |
+
"learning_rate": 9.051794871794872e-06,
|
| 175942 |
+
"loss": 0.3232,
|
| 175943 |
+
"step": 60190
|
| 175944 |
+
},
|
| 175945 |
+
{
|
| 175946 |
+
"epoch": 480.72,
|
| 175947 |
+
"learning_rate": 9.051714743589745e-06,
|
| 175948 |
+
"loss": 0.2884,
|
| 175949 |
+
"step": 60195
|
| 175950 |
+
},
|
| 175951 |
+
{
|
| 175952 |
+
"epoch": 480.76,
|
| 175953 |
+
"learning_rate": 9.051634615384616e-06,
|
| 175954 |
+
"loss": 0.4555,
|
| 175955 |
+
"step": 60200
|
| 175956 |
+
},
|
| 175957 |
+
{
|
| 175958 |
+
"epoch": 480.8,
|
| 175959 |
+
"learning_rate": 9.051554487179488e-06,
|
| 175960 |
+
"loss": 1.1592,
|
| 175961 |
+
"step": 60205
|
| 175962 |
+
},
|
| 175963 |
+
{
|
| 175964 |
+
"epoch": 480.84,
|
| 175965 |
+
"learning_rate": 9.051474358974359e-06,
|
| 175966 |
+
"loss": 0.3908,
|
| 175967 |
+
"step": 60210
|
| 175968 |
+
},
|
| 175969 |
+
{
|
| 175970 |
+
"epoch": 480.88,
|
| 175971 |
+
"learning_rate": 9.051394230769232e-06,
|
| 175972 |
+
"loss": 0.3136,
|
| 175973 |
+
"step": 60215
|
| 175974 |
+
},
|
| 175975 |
+
{
|
| 175976 |
+
"epoch": 480.92,
|
| 175977 |
+
"learning_rate": 9.051314102564103e-06,
|
| 175978 |
+
"loss": 0.357,
|
| 175979 |
+
"step": 60220
|
| 175980 |
+
},
|
| 175981 |
+
{
|
| 175982 |
+
"epoch": 480.96,
|
| 175983 |
+
"learning_rate": 9.051233974358975e-06,
|
| 175984 |
+
"loss": 0.4976,
|
| 175985 |
+
"step": 60225
|
| 175986 |
+
},
|
| 175987 |
+
{
|
| 175988 |
+
"epoch": 481.0,
|
| 175989 |
+
"learning_rate": 9.051153846153848e-06,
|
| 175990 |
+
"loss": 1.3006,
|
| 175991 |
+
"step": 60230
|
| 175992 |
+
},
|
| 175993 |
+
{
|
| 175994 |
+
"epoch": 481.0,
|
| 175995 |
+
"eval_loss": 0.43538910150527954,
|
| 175996 |
+
"eval_runtime": 40.2473,
|
| 175997 |
+
"eval_samples_per_second": 20.821,
|
| 175998 |
+
"eval_steps_per_second": 0.671,
|
| 175999 |
+
"eval_wer": 0.19678979771328056,
|
| 176000 |
+
"step": 60230
|
| 176001 |
+
},
|
| 176002 |
+
{
|
| 176003 |
+
"epoch": 481.04,
|
| 176004 |
+
"learning_rate": 9.051073717948719e-06,
|
| 176005 |
+
"loss": 0.3275,
|
| 176006 |
+
"step": 60235
|
| 176007 |
+
},
|
| 176008 |
+
{
|
| 176009 |
+
"epoch": 481.08,
|
| 176010 |
+
"learning_rate": 9.05099358974359e-06,
|
| 176011 |
+
"loss": 0.3484,
|
| 176012 |
+
"step": 60240
|
| 176013 |
+
},
|
| 176014 |
+
{
|
| 176015 |
+
"epoch": 481.12,
|
| 176016 |
+
"learning_rate": 9.050913461538462e-06,
|
| 176017 |
+
"loss": 0.3509,
|
| 176018 |
+
"step": 60245
|
| 176019 |
+
},
|
| 176020 |
+
{
|
| 176021 |
+
"epoch": 481.16,
|
| 176022 |
+
"learning_rate": 9.050833333333335e-06,
|
| 176023 |
+
"loss": 0.5427,
|
| 176024 |
+
"step": 60250
|
| 176025 |
+
},
|
| 176026 |
+
{
|
| 176027 |
+
"epoch": 481.2,
|
| 176028 |
+
"learning_rate": 9.050753205128206e-06,
|
| 176029 |
+
"loss": 1.181,
|
| 176030 |
+
"step": 60255
|
| 176031 |
+
},
|
| 176032 |
+
{
|
| 176033 |
+
"epoch": 481.24,
|
| 176034 |
+
"learning_rate": 9.050673076923078e-06,
|
| 176035 |
+
"loss": 0.4515,
|
| 176036 |
+
"step": 60260
|
| 176037 |
+
},
|
| 176038 |
+
{
|
| 176039 |
+
"epoch": 481.28,
|
| 176040 |
+
"learning_rate": 9.050592948717949e-06,
|
| 176041 |
+
"loss": 0.2992,
|
| 176042 |
+
"step": 60265
|
| 176043 |
+
},
|
| 176044 |
+
{
|
| 176045 |
+
"epoch": 481.32,
|
| 176046 |
+
"learning_rate": 9.050512820512822e-06,
|
| 176047 |
+
"loss": 0.398,
|
| 176048 |
+
"step": 60270
|
| 176049 |
+
},
|
| 176050 |
+
{
|
| 176051 |
+
"epoch": 481.36,
|
| 176052 |
+
"learning_rate": 9.050432692307693e-06,
|
| 176053 |
+
"loss": 0.4436,
|
| 176054 |
+
"step": 60275
|
| 176055 |
+
},
|
| 176056 |
+
{
|
| 176057 |
+
"epoch": 481.4,
|
| 176058 |
+
"learning_rate": 9.050352564102565e-06,
|
| 176059 |
+
"loss": 1.1835,
|
| 176060 |
+
"step": 60280
|
| 176061 |
+
},
|
| 176062 |
+
{
|
| 176063 |
+
"epoch": 481.44,
|
| 176064 |
+
"learning_rate": 9.050272435897438e-06,
|
| 176065 |
+
"loss": 0.3682,
|
| 176066 |
+
"step": 60285
|
| 176067 |
+
},
|
| 176068 |
+
{
|
| 176069 |
+
"epoch": 481.48,
|
| 176070 |
+
"learning_rate": 9.050192307692307e-06,
|
| 176071 |
+
"loss": 0.4315,
|
| 176072 |
+
"step": 60290
|
| 176073 |
+
},
|
| 176074 |
+
{
|
| 176075 |
+
"epoch": 481.52,
|
| 176076 |
+
"learning_rate": 9.05011217948718e-06,
|
| 176077 |
+
"loss": 0.3953,
|
| 176078 |
+
"step": 60295
|
| 176079 |
+
},
|
| 176080 |
+
{
|
| 176081 |
+
"epoch": 481.56,
|
| 176082 |
+
"learning_rate": 9.050032051282052e-06,
|
| 176083 |
+
"loss": 0.5403,
|
| 176084 |
+
"step": 60300
|
| 176085 |
+
},
|
| 176086 |
+
{
|
| 176087 |
+
"epoch": 481.6,
|
| 176088 |
+
"learning_rate": 9.049951923076923e-06,
|
| 176089 |
+
"loss": 1.2516,
|
| 176090 |
+
"step": 60305
|
| 176091 |
+
},
|
| 176092 |
+
{
|
| 176093 |
+
"epoch": 481.64,
|
| 176094 |
+
"learning_rate": 9.049871794871795e-06,
|
| 176095 |
+
"loss": 0.3523,
|
| 176096 |
+
"step": 60310
|
| 176097 |
+
},
|
| 176098 |
+
{
|
| 176099 |
+
"epoch": 481.68,
|
| 176100 |
+
"learning_rate": 9.049791666666668e-06,
|
| 176101 |
+
"loss": 0.34,
|
| 176102 |
+
"step": 60315
|
| 176103 |
+
},
|
| 176104 |
+
{
|
| 176105 |
+
"epoch": 481.72,
|
| 176106 |
+
"learning_rate": 9.049711538461539e-06,
|
| 176107 |
+
"loss": 0.3594,
|
| 176108 |
+
"step": 60320
|
| 176109 |
+
},
|
| 176110 |
+
{
|
| 176111 |
+
"epoch": 481.76,
|
| 176112 |
+
"learning_rate": 9.04963141025641e-06,
|
| 176113 |
+
"loss": 0.5865,
|
| 176114 |
+
"step": 60325
|
| 176115 |
+
},
|
| 176116 |
+
{
|
| 176117 |
+
"epoch": 481.8,
|
| 176118 |
+
"learning_rate": 9.049551282051283e-06,
|
| 176119 |
+
"loss": 1.1754,
|
| 176120 |
+
"step": 60330
|
| 176121 |
+
},
|
| 176122 |
+
{
|
| 176123 |
+
"epoch": 481.84,
|
| 176124 |
+
"learning_rate": 9.049471153846155e-06,
|
| 176125 |
+
"loss": 0.3405,
|
| 176126 |
+
"step": 60335
|
| 176127 |
+
},
|
| 176128 |
+
{
|
| 176129 |
+
"epoch": 481.88,
|
| 176130 |
+
"learning_rate": 9.049391025641026e-06,
|
| 176131 |
+
"loss": 0.656,
|
| 176132 |
+
"step": 60340
|
| 176133 |
+
},
|
| 176134 |
+
{
|
| 176135 |
+
"epoch": 481.92,
|
| 176136 |
+
"learning_rate": 9.049310897435897e-06,
|
| 176137 |
+
"loss": 0.342,
|
| 176138 |
+
"step": 60345
|
| 176139 |
+
},
|
| 176140 |
+
{
|
| 176141 |
+
"epoch": 481.96,
|
| 176142 |
+
"learning_rate": 9.04923076923077e-06,
|
| 176143 |
+
"loss": 0.4822,
|
| 176144 |
+
"step": 60350
|
| 176145 |
+
},
|
| 176146 |
+
{
|
| 176147 |
+
"epoch": 482.0,
|
| 176148 |
+
"learning_rate": 9.049150641025642e-06,
|
| 176149 |
+
"loss": 1.2516,
|
| 176150 |
+
"step": 60355
|
| 176151 |
+
},
|
| 176152 |
+
{
|
| 176153 |
+
"epoch": 482.0,
|
| 176154 |
+
"eval_loss": 0.4162551164627075,
|
| 176155 |
+
"eval_runtime": 40.1501,
|
| 176156 |
+
"eval_samples_per_second": 20.872,
|
| 176157 |
+
"eval_steps_per_second": 0.672,
|
| 176158 |
+
"eval_wer": 0.19967876177265095,
|
| 176159 |
+
"step": 60355
|
| 176160 |
+
},
|
| 176161 |
+
{
|
| 176162 |
+
"epoch": 482.04,
|
| 176163 |
+
"learning_rate": 9.049070512820513e-06,
|
| 176164 |
+
"loss": 0.4382,
|
| 176165 |
+
"step": 60360
|
| 176166 |
+
},
|
| 176167 |
+
{
|
| 176168 |
+
"epoch": 482.08,
|
| 176169 |
+
"learning_rate": 9.048990384615385e-06,
|
| 176170 |
+
"loss": 0.3064,
|
| 176171 |
+
"step": 60365
|
| 176172 |
+
},
|
| 176173 |
+
{
|
| 176174 |
+
"epoch": 482.12,
|
| 176175 |
+
"learning_rate": 9.048910256410258e-06,
|
| 176176 |
+
"loss": 0.348,
|
| 176177 |
+
"step": 60370
|
| 176178 |
+
},
|
| 176179 |
+
{
|
| 176180 |
+
"epoch": 482.16,
|
| 176181 |
+
"learning_rate": 9.048830128205129e-06,
|
| 176182 |
+
"loss": 0.4735,
|
| 176183 |
+
"step": 60375
|
| 176184 |
+
},
|
| 176185 |
+
{
|
| 176186 |
+
"epoch": 482.2,
|
| 176187 |
+
"learning_rate": 9.04875e-06,
|
| 176188 |
+
"loss": 1.3066,
|
| 176189 |
+
"step": 60380
|
| 176190 |
+
},
|
| 176191 |
+
{
|
| 176192 |
+
"epoch": 482.24,
|
| 176193 |
+
"learning_rate": 9.048669871794873e-06,
|
| 176194 |
+
"loss": 0.4388,
|
| 176195 |
+
"step": 60385
|
| 176196 |
+
},
|
| 176197 |
+
{
|
| 176198 |
+
"epoch": 482.28,
|
| 176199 |
+
"learning_rate": 9.048589743589745e-06,
|
| 176200 |
+
"loss": 0.3116,
|
| 176201 |
+
"step": 60390
|
| 176202 |
+
},
|
| 176203 |
+
{
|
| 176204 |
+
"epoch": 482.32,
|
| 176205 |
+
"learning_rate": 9.048509615384616e-06,
|
| 176206 |
+
"loss": 0.3442,
|
| 176207 |
+
"step": 60395
|
| 176208 |
+
},
|
| 176209 |
+
{
|
| 176210 |
+
"epoch": 482.36,
|
| 176211 |
+
"learning_rate": 9.048429487179488e-06,
|
| 176212 |
+
"loss": 0.483,
|
| 176213 |
+
"step": 60400
|
| 176214 |
+
},
|
| 176215 |
+
{
|
| 176216 |
+
"epoch": 482.4,
|
| 176217 |
+
"learning_rate": 9.04834935897436e-06,
|
| 176218 |
+
"loss": 1.329,
|
| 176219 |
+
"step": 60405
|
| 176220 |
+
},
|
| 176221 |
+
{
|
| 176222 |
+
"epoch": 482.44,
|
| 176223 |
+
"learning_rate": 9.04826923076923e-06,
|
| 176224 |
+
"loss": 0.3322,
|
| 176225 |
+
"step": 60410
|
| 176226 |
+
},
|
| 176227 |
+
{
|
| 176228 |
+
"epoch": 482.48,
|
| 176229 |
+
"learning_rate": 9.048189102564103e-06,
|
| 176230 |
+
"loss": 0.3371,
|
| 176231 |
+
"step": 60415
|
| 176232 |
+
},
|
| 176233 |
+
{
|
| 176234 |
+
"epoch": 482.52,
|
| 176235 |
+
"learning_rate": 9.048108974358976e-06,
|
| 176236 |
+
"loss": 0.3224,
|
| 176237 |
+
"step": 60420
|
| 176238 |
+
},
|
| 176239 |
+
{
|
| 176240 |
+
"epoch": 482.56,
|
| 176241 |
+
"learning_rate": 9.048028846153846e-06,
|
| 176242 |
+
"loss": 0.5545,
|
| 176243 |
+
"step": 60425
|
| 176244 |
+
},
|
| 176245 |
+
{
|
| 176246 |
+
"epoch": 482.6,
|
| 176247 |
+
"learning_rate": 9.047948717948719e-06,
|
| 176248 |
+
"loss": 1.3621,
|
| 176249 |
+
"step": 60430
|
| 176250 |
+
},
|
| 176251 |
+
{
|
| 176252 |
+
"epoch": 482.64,
|
| 176253 |
+
"learning_rate": 9.04786858974359e-06,
|
| 176254 |
+
"loss": 0.3918,
|
| 176255 |
+
"step": 60435
|
| 176256 |
+
},
|
| 176257 |
+
{
|
| 176258 |
+
"epoch": 482.68,
|
| 176259 |
+
"learning_rate": 9.047788461538462e-06,
|
| 176260 |
+
"loss": 0.3828,
|
| 176261 |
+
"step": 60440
|
| 176262 |
+
},
|
| 176263 |
+
{
|
| 176264 |
+
"epoch": 482.72,
|
| 176265 |
+
"learning_rate": 9.047708333333333e-06,
|
| 176266 |
+
"loss": 0.3523,
|
| 176267 |
+
"step": 60445
|
| 176268 |
+
},
|
| 176269 |
+
{
|
| 176270 |
+
"epoch": 482.76,
|
| 176271 |
+
"learning_rate": 9.047628205128206e-06,
|
| 176272 |
+
"loss": 0.448,
|
| 176273 |
+
"step": 60450
|
| 176274 |
+
},
|
| 176275 |
+
{
|
| 176276 |
+
"epoch": 482.8,
|
| 176277 |
+
"learning_rate": 9.047548076923078e-06,
|
| 176278 |
+
"loss": 1.3147,
|
| 176279 |
+
"step": 60455
|
| 176280 |
+
},
|
| 176281 |
+
{
|
| 176282 |
+
"epoch": 482.84,
|
| 176283 |
+
"learning_rate": 9.047467948717949e-06,
|
| 176284 |
+
"loss": 0.3676,
|
| 176285 |
+
"step": 60460
|
| 176286 |
+
},
|
| 176287 |
+
{
|
| 176288 |
+
"epoch": 482.88,
|
| 176289 |
+
"learning_rate": 9.04738782051282e-06,
|
| 176290 |
+
"loss": 0.3621,
|
| 176291 |
+
"step": 60465
|
| 176292 |
+
},
|
| 176293 |
+
{
|
| 176294 |
+
"epoch": 482.92,
|
| 176295 |
+
"learning_rate": 9.047307692307693e-06,
|
| 176296 |
+
"loss": 0.3729,
|
| 176297 |
+
"step": 60470
|
| 176298 |
+
},
|
| 176299 |
+
{
|
| 176300 |
+
"epoch": 482.96,
|
| 176301 |
+
"learning_rate": 9.047227564102565e-06,
|
| 176302 |
+
"loss": 0.5614,
|
| 176303 |
+
"step": 60475
|
| 176304 |
+
},
|
| 176305 |
+
{
|
| 176306 |
+
"epoch": 483.0,
|
| 176307 |
+
"learning_rate": 9.047147435897436e-06,
|
| 176308 |
+
"loss": 1.4147,
|
| 176309 |
+
"step": 60480
|
| 176310 |
+
},
|
| 176311 |
+
{
|
| 176312 |
+
"epoch": 483.0,
|
| 176313 |
+
"eval_loss": 0.43452778458595276,
|
| 176314 |
+
"eval_runtime": 40.3349,
|
| 176315 |
+
"eval_samples_per_second": 20.776,
|
| 176316 |
+
"eval_steps_per_second": 0.669,
|
| 176317 |
+
"eval_wer": 0.19634636463280963,
|
| 176318 |
+
"step": 60480
|
| 176319 |
+
},
|
| 176320 |
+
{
|
| 176321 |
+
"epoch": 487.04,
|
| 176322 |
+
"learning_rate": 9.047067307692309e-06,
|
| 176323 |
+
"loss": 0.4135,
|
| 176324 |
+
"step": 60485
|
| 176325 |
+
},
|
| 176326 |
+
{
|
| 176327 |
+
"epoch": 487.08,
|
| 176328 |
+
"learning_rate": 9.04698717948718e-06,
|
| 176329 |
+
"loss": 0.347,
|
| 176330 |
+
"step": 60490
|
| 176331 |
+
},
|
| 176332 |
+
{
|
| 176333 |
+
"epoch": 487.12,
|
| 176334 |
+
"learning_rate": 9.046907051282052e-06,
|
| 176335 |
+
"loss": 0.5046,
|
| 176336 |
+
"step": 60495
|
| 176337 |
+
},
|
| 176338 |
+
{
|
| 176339 |
+
"epoch": 487.16,
|
| 176340 |
+
"learning_rate": 9.046826923076923e-06,
|
| 176341 |
+
"loss": 0.4958,
|
| 176342 |
+
"step": 60500
|
| 176343 |
+
},
|
| 176344 |
+
{
|
| 176345 |
+
"epoch": 487.2,
|
| 176346 |
+
"learning_rate": 9.046746794871796e-06,
|
| 176347 |
+
"loss": 1.1384,
|
| 176348 |
+
"step": 60505
|
| 176349 |
+
},
|
| 176350 |
+
{
|
| 176351 |
+
"epoch": 487.24,
|
| 176352 |
+
"learning_rate": 9.046666666666668e-06,
|
| 176353 |
+
"loss": 0.386,
|
| 176354 |
+
"step": 60510
|
| 176355 |
+
},
|
| 176356 |
+
{
|
| 176357 |
+
"epoch": 487.28,
|
| 176358 |
+
"learning_rate": 9.046586538461539e-06,
|
| 176359 |
+
"loss": 0.3101,
|
| 176360 |
+
"step": 60515
|
| 176361 |
+
},
|
| 176362 |
+
{
|
| 176363 |
+
"epoch": 487.32,
|
| 176364 |
+
"learning_rate": 9.046506410256412e-06,
|
| 176365 |
+
"loss": 0.3785,
|
| 176366 |
+
"step": 60520
|
| 176367 |
+
},
|
| 176368 |
+
{
|
| 176369 |
+
"epoch": 487.36,
|
| 176370 |
+
"learning_rate": 9.046426282051283e-06,
|
| 176371 |
+
"loss": 0.5308,
|
| 176372 |
+
"step": 60525
|
| 176373 |
+
},
|
| 176374 |
+
{
|
| 176375 |
+
"epoch": 487.4,
|
| 176376 |
+
"learning_rate": 9.046346153846155e-06,
|
| 176377 |
+
"loss": 1.2602,
|
| 176378 |
+
"step": 60530
|
| 176379 |
+
},
|
| 176380 |
+
{
|
| 176381 |
+
"epoch": 487.44,
|
| 176382 |
+
"learning_rate": 9.046266025641026e-06,
|
| 176383 |
+
"loss": 0.3258,
|
| 176384 |
+
"step": 60535
|
| 176385 |
+
},
|
| 176386 |
+
{
|
| 176387 |
+
"epoch": 487.48,
|
| 176388 |
+
"learning_rate": 9.046185897435899e-06,
|
| 176389 |
+
"loss": 0.305,
|
| 176390 |
+
"step": 60540
|
| 176391 |
+
},
|
| 176392 |
+
{
|
| 176393 |
+
"epoch": 487.52,
|
| 176394 |
+
"learning_rate": 9.04610576923077e-06,
|
| 176395 |
+
"loss": 0.336,
|
| 176396 |
+
"step": 60545
|
| 176397 |
+
},
|
| 176398 |
+
{
|
| 176399 |
+
"epoch": 487.56,
|
| 176400 |
+
"learning_rate": 9.046025641025642e-06,
|
| 176401 |
+
"loss": 0.5471,
|
| 176402 |
+
"step": 60550
|
| 176403 |
+
},
|
| 176404 |
+
{
|
| 176405 |
+
"epoch": 487.6,
|
| 176406 |
+
"learning_rate": 9.045945512820513e-06,
|
| 176407 |
+
"loss": 1.2149,
|
| 176408 |
+
"step": 60555
|
| 176409 |
+
},
|
| 176410 |
+
{
|
| 176411 |
+
"epoch": 487.65,
|
| 176412 |
+
"learning_rate": 9.045865384615386e-06,
|
| 176413 |
+
"loss": 0.3799,
|
| 176414 |
+
"step": 60560
|
| 176415 |
+
},
|
| 176416 |
+
{
|
| 176417 |
+
"epoch": 487.69,
|
| 176418 |
+
"learning_rate": 9.045785256410256e-06,
|
| 176419 |
+
"loss": 0.2942,
|
| 176420 |
+
"step": 60565
|
| 176421 |
+
},
|
| 176422 |
+
{
|
| 176423 |
+
"epoch": 487.73,
|
| 176424 |
+
"learning_rate": 9.045705128205129e-06,
|
| 176425 |
+
"loss": 0.3494,
|
| 176426 |
+
"step": 60570
|
| 176427 |
+
},
|
| 176428 |
+
{
|
| 176429 |
+
"epoch": 487.77,
|
| 176430 |
+
"learning_rate": 9.045625000000002e-06,
|
| 176431 |
+
"loss": 0.4582,
|
| 176432 |
+
"step": 60575
|
| 176433 |
+
},
|
| 176434 |
+
{
|
| 176435 |
+
"epoch": 487.81,
|
| 176436 |
+
"learning_rate": 9.045560897435898e-06,
|
| 176437 |
+
"loss": 1.3935,
|
| 176438 |
+
"step": 60580
|
| 176439 |
+
},
|
| 176440 |
+
{
|
| 176441 |
+
"epoch": 487.85,
|
| 176442 |
+
"learning_rate": 9.04548076923077e-06,
|
| 176443 |
+
"loss": 0.4015,
|
| 176444 |
+
"step": 60585
|
| 176445 |
+
},
|
| 176446 |
+
{
|
| 176447 |
+
"epoch": 487.89,
|
| 176448 |
+
"learning_rate": 9.045400641025642e-06,
|
| 176449 |
+
"loss": 0.2689,
|
| 176450 |
+
"step": 60590
|
| 176451 |
+
},
|
| 176452 |
+
{
|
| 176453 |
+
"epoch": 487.93,
|
| 176454 |
+
"learning_rate": 9.045320512820514e-06,
|
| 176455 |
+
"loss": 0.3782,
|
| 176456 |
+
"step": 60595
|
| 176457 |
+
},
|
| 176458 |
+
{
|
| 176459 |
+
"epoch": 487.97,
|
| 176460 |
+
"learning_rate": 9.045240384615387e-06,
|
| 176461 |
+
"loss": 0.5354,
|
| 176462 |
+
"step": 60600
|
| 176463 |
+
},
|
| 176464 |
+
{
|
| 176465 |
+
"epoch": 488.0,
|
| 176466 |
+
"eval_loss": 0.38258084654808044,
|
| 176467 |
+
"eval_runtime": 39.0437,
|
| 176468 |
+
"eval_samples_per_second": 21.463,
|
| 176469 |
+
"eval_steps_per_second": 0.692,
|
| 176470 |
+
"eval_wer": 0.19932432432432431,
|
| 176471 |
+
"step": 60604
|
| 176472 |
}
|
| 176473 |
],
|
| 176474 |
"max_steps": 620000,
|
| 176475 |
"num_train_epochs": 5000,
|
| 176476 |
+
"total_flos": 1.7056102466424242e+20,
|
| 176477 |
"trial_name": null,
|
| 176478 |
"trial_params": null
|
| 176479 |
}
|
model-bin/finetune/base/{checkpoint-55501 β checkpoint-60604}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-60105 β checkpoint-60729}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-60105 β checkpoint-60729}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165009
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0406650db1393fb9692f8d43b2e0d3fd7ac1ccba07e5597f49b7e009002820b6
|
| 3 |
size 722165009
|
model-bin/finetune/base/{checkpoint-60105 β checkpoint-60729}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-55501 β checkpoint-60729}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2839b98fe4b47ad20f5b8b1bae2552bd34df8084d7cb7ec7a9efadbef996fd76
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-55501 β checkpoint-60729}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e4b858c2d8d13cecbbedfc9bcb6fd9b338c4dab35127f6a04749c2971bdf13f5
|
| 3 |
size 14503
|
model-bin/finetune/base/{checkpoint-55501 β checkpoint-60729}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8726c92bb7ef60e388546e9fd18baec8ae7c170a0be794a916a6e8ed8249c65c
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-60105 β checkpoint-60729}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bd58c4ff7961fcf183c96f47a9e93eb6ea21e444f358cbb7cda92b3831704324
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-55501 β checkpoint-60729}/trainer_state.json
RENAMED
|
The diff for this file is too large to render.
See raw diff
|
|
|
model-bin/finetune/base/{checkpoint-60105 β checkpoint-60729}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1629798286.419865/events.out.tfevents.1629798286.c435e1c5ee04.920.101
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8cdff19b05c5968077560f33eef23f6875ca4369a594138a22f689d4e5110b1f
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629798930.4451573/events.out.tfevents.1629798930.c435e1c5ee04.920.103
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c3c25008b862a6b83c49da8da3e9cf89f99feeaa5e5dd15b7671b3a36095c3e1
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629799575.228714/events.out.tfevents.1629799575.c435e1c5ee04.920.105
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:cf181fc102c8aa4d48a39d2da285c497cfaf12a7875db52f035c81bfe74de9f2
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629800218.038527/events.out.tfevents.1629800218.c435e1c5ee04.920.107
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a5f981abdb766a16f2729062859a1afb40cb90685c56de17aa042b8517f669a1
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629800871.1235416/events.out.tfevents.1629800871.c435e1c5ee04.920.109
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a20fd083fb3c7be7c7944010abc2e1b9fa23a305d56148813f97d5af1bb8c279
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1629798286.c435e1c5ee04.920.100
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b3f3ed544a80037d3a4a4ff4bf81759482a236f489fb1992f204e3aa98d05195
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629798930.c435e1c5ee04.920.102
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:377c33ce99cb86655a50d4fcd7b766ff3cb1dfce42c4d694049996756a81d123
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629799575.c435e1c5ee04.920.104
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:22913b623bf7a65164b7fff2d9e6a4d08d222fd5509d4a2aa46b358922f1bfdf
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629800218.c435e1c5ee04.920.106
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:cee26fb6ace669d99253f6d5365bdfff21dc68cef4451a6e5d3730bced531825
|
| 3 |
+
size 8462
|
model-bin/finetune/base/log/events.out.tfevents.1629800871.c435e1c5ee04.920.108
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bbb8fff7bcd951291a1f98c7d9279705af59cd699ed617ad824f6084531a7843
|
| 3 |
+
size 8622
|