"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-132910 β checkpoint-143240}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-132910 β checkpoint-143240}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-132910 β checkpoint-143240}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-142742 β checkpoint-143240}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-142742 β checkpoint-143240}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-142742 β checkpoint-143240}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-132910 β checkpoint-143240}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-142742 β checkpoint-143240}/trainer_state.json +641 -5
- model-bin/finetune/base/{checkpoint-132910 β checkpoint-143240}/training_args.bin +0 -0
- model-bin/finetune/base/{checkpoint-142742 β checkpoint-143365}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-142742 β checkpoint-143365}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-142742 β checkpoint-143365}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-132910 β checkpoint-143365}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-132910 β checkpoint-143365}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-132910 β checkpoint-143365}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-142742 β checkpoint-143365}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-132910 β checkpoint-143365}/trainer_state.json +0 -0
- model-bin/finetune/base/{checkpoint-142742 β checkpoint-143365}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630217087.7673898/events.out.tfevents.1630217087.cc93b136ebf5.1086.59 +3 -0
- model-bin/finetune/base/log/1630217533.1449723/events.out.tfevents.1630217533.cc93b136ebf5.1086.61 +3 -0
- model-bin/finetune/base/log/1630217981.5941153/events.out.tfevents.1630217981.cc93b136ebf5.1086.63 +3 -0
- model-bin/finetune/base/log/1630218424.3441827/events.out.tfevents.1630218424.cc93b136ebf5.1086.65 +3 -0
- model-bin/finetune/base/log/1630218868.8727193/events.out.tfevents.1630218868.cc93b136ebf5.1086.67 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630217087.cc93b136ebf5.1086.58 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630217533.cc93b136ebf5.1086.60 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630217981.cc93b136ebf5.1086.62 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630218424.cc93b136ebf5.1086.64 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630218868.cc93b136ebf5.1086.66 +3 -0
model-bin/finetune/base/{checkpoint-132910 β checkpoint-143240}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-132910 β checkpoint-143240}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3b07eb70666c98efa81b1ae25d30d8dc0ef685547e99aa9c605653d13ffc1a26
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-132910 β checkpoint-143240}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-142742 β checkpoint-143240}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:719c64023422102f774c4b415dbe42084ebf730e04e436945dbc23a39747dc12
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-142742 β checkpoint-143240}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b4d59aaa1d1129f5ce7777bdc711ff6665db55d55d5210162fcd2ecfe304ca32
|
| 3 |
size 14503
|
model-bin/finetune/base/{checkpoint-142742 β checkpoint-143240}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:392bcd906a1be97f8bf6b539921d31e1244067c313fa0a9adb8dd019a4c75fa3
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-132910 β checkpoint-143240}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:317d976693bb649858811cc275bd77488d50936643f55da2776daa72308f421c
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-142742 β checkpoint-143240}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
-
"best_metric": 0.
|
| 3 |
-
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -280977,11 +280977,647 @@
|
|
| 280977 |
"eval_steps_per_second": 0.655,
|
| 280978 |
"eval_wer": 0.18999632217727105,
|
| 280979 |
"step": 142742
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 280980 |
}
|
| 280981 |
],
|
| 280982 |
"max_steps": 625000,
|
| 280983 |
"num_train_epochs": 5000,
|
| 280984 |
-
"total_flos": 4.
|
| 280985 |
"trial_name": null,
|
| 280986 |
"trial_params": null
|
| 280987 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"best_metric": 0.16819527695529718,
|
| 3 |
+
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-143240",
|
| 4 |
+
"epoch": 1145.0,
|
| 5 |
+
"global_step": 143240,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 280977 |
"eval_steps_per_second": 0.655,
|
| 280978 |
"eval_wer": 0.18999632217727105,
|
| 280979 |
"step": 142742
|
| 280980 |
+
},
|
| 280981 |
+
{
|
| 280982 |
+
"epoch": 1151.02,
|
| 280983 |
+
"learning_rate": 7.711017770597739e-06,
|
| 280984 |
+
"loss": 0.3134,
|
| 280985 |
+
"step": 142745
|
| 280986 |
+
},
|
| 280987 |
+
{
|
| 280988 |
+
"epoch": 1151.06,
|
| 280989 |
+
"learning_rate": 7.710936995153475e-06,
|
| 280990 |
+
"loss": 0.2984,
|
| 280991 |
+
"step": 142750
|
| 280992 |
+
},
|
| 280993 |
+
{
|
| 280994 |
+
"epoch": 1151.1,
|
| 280995 |
+
"learning_rate": 7.710856219709209e-06,
|
| 280996 |
+
"loss": 0.286,
|
| 280997 |
+
"step": 142755
|
| 280998 |
+
},
|
| 280999 |
+
{
|
| 281000 |
+
"epoch": 1151.14,
|
| 281001 |
+
"learning_rate": 7.710775444264945e-06,
|
| 281002 |
+
"loss": 0.2923,
|
| 281003 |
+
"step": 142760
|
| 281004 |
+
},
|
| 281005 |
+
{
|
| 281006 |
+
"epoch": 1151.18,
|
| 281007 |
+
"learning_rate": 7.71069466882068e-06,
|
| 281008 |
+
"loss": 0.7799,
|
| 281009 |
+
"step": 142765
|
| 281010 |
+
},
|
| 281011 |
+
{
|
| 281012 |
+
"epoch": 1151.22,
|
| 281013 |
+
"learning_rate": 7.710613893376414e-06,
|
| 281014 |
+
"loss": 0.8129,
|
| 281015 |
+
"step": 142770
|
| 281016 |
+
},
|
| 281017 |
+
{
|
| 281018 |
+
"epoch": 1151.27,
|
| 281019 |
+
"learning_rate": 7.71053311793215e-06,
|
| 281020 |
+
"loss": 0.3205,
|
| 281021 |
+
"step": 142775
|
| 281022 |
+
},
|
| 281023 |
+
{
|
| 281024 |
+
"epoch": 1151.31,
|
| 281025 |
+
"learning_rate": 7.710452342487884e-06,
|
| 281026 |
+
"loss": 0.2673,
|
| 281027 |
+
"step": 142780
|
| 281028 |
+
},
|
| 281029 |
+
{
|
| 281030 |
+
"epoch": 1151.35,
|
| 281031 |
+
"learning_rate": 7.71037156704362e-06,
|
| 281032 |
+
"loss": 0.3895,
|
| 281033 |
+
"step": 142785
|
| 281034 |
+
},
|
| 281035 |
+
{
|
| 281036 |
+
"epoch": 1151.39,
|
| 281037 |
+
"learning_rate": 7.710290791599354e-06,
|
| 281038 |
+
"loss": 0.6514,
|
| 281039 |
+
"step": 142790
|
| 281040 |
+
},
|
| 281041 |
+
{
|
| 281042 |
+
"epoch": 1151.43,
|
| 281043 |
+
"learning_rate": 7.71021001615509e-06,
|
| 281044 |
+
"loss": 0.8698,
|
| 281045 |
+
"step": 142795
|
| 281046 |
+
},
|
| 281047 |
+
{
|
| 281048 |
+
"epoch": 1151.47,
|
| 281049 |
+
"learning_rate": 7.710129240710824e-06,
|
| 281050 |
+
"loss": 0.283,
|
| 281051 |
+
"step": 142800
|
| 281052 |
+
},
|
| 281053 |
+
{
|
| 281054 |
+
"epoch": 1151.51,
|
| 281055 |
+
"learning_rate": 7.71004846526656e-06,
|
| 281056 |
+
"loss": 0.3581,
|
| 281057 |
+
"step": 142805
|
| 281058 |
+
},
|
| 281059 |
+
{
|
| 281060 |
+
"epoch": 1151.55,
|
| 281061 |
+
"learning_rate": 7.709967689822294e-06,
|
| 281062 |
+
"loss": 0.3568,
|
| 281063 |
+
"step": 142810
|
| 281064 |
+
},
|
| 281065 |
+
{
|
| 281066 |
+
"epoch": 1151.59,
|
| 281067 |
+
"learning_rate": 7.70988691437803e-06,
|
| 281068 |
+
"loss": 0.7716,
|
| 281069 |
+
"step": 142815
|
| 281070 |
+
},
|
| 281071 |
+
{
|
| 281072 |
+
"epoch": 1151.63,
|
| 281073 |
+
"learning_rate": 7.709806138933766e-06,
|
| 281074 |
+
"loss": 0.8879,
|
| 281075 |
+
"step": 142820
|
| 281076 |
+
},
|
| 281077 |
+
{
|
| 281078 |
+
"epoch": 1151.67,
|
| 281079 |
+
"learning_rate": 7.7097253634895e-06,
|
| 281080 |
+
"loss": 0.2742,
|
| 281081 |
+
"step": 142825
|
| 281082 |
+
},
|
| 281083 |
+
{
|
| 281084 |
+
"epoch": 1151.71,
|
| 281085 |
+
"learning_rate": 7.709644588045236e-06,
|
| 281086 |
+
"loss": 0.2852,
|
| 281087 |
+
"step": 142830
|
| 281088 |
+
},
|
| 281089 |
+
{
|
| 281090 |
+
"epoch": 1151.75,
|
| 281091 |
+
"learning_rate": 7.70956381260097e-06,
|
| 281092 |
+
"loss": 0.3155,
|
| 281093 |
+
"step": 142835
|
| 281094 |
+
},
|
| 281095 |
+
{
|
| 281096 |
+
"epoch": 1151.79,
|
| 281097 |
+
"learning_rate": 7.709483037156706e-06,
|
| 281098 |
+
"loss": 0.6991,
|
| 281099 |
+
"step": 142840
|
| 281100 |
+
},
|
| 281101 |
+
{
|
| 281102 |
+
"epoch": 1151.83,
|
| 281103 |
+
"learning_rate": 7.70940226171244e-06,
|
| 281104 |
+
"loss": 0.8038,
|
| 281105 |
+
"step": 142845
|
| 281106 |
+
},
|
| 281107 |
+
{
|
| 281108 |
+
"epoch": 1151.87,
|
| 281109 |
+
"learning_rate": 7.709321486268176e-06,
|
| 281110 |
+
"loss": 0.2299,
|
| 281111 |
+
"step": 142850
|
| 281112 |
+
},
|
| 281113 |
+
{
|
| 281114 |
+
"epoch": 1151.91,
|
| 281115 |
+
"learning_rate": 7.70924071082391e-06,
|
| 281116 |
+
"loss": 0.2772,
|
| 281117 |
+
"step": 142855
|
| 281118 |
+
},
|
| 281119 |
+
{
|
| 281120 |
+
"epoch": 1151.95,
|
| 281121 |
+
"learning_rate": 7.709159935379646e-06,
|
| 281122 |
+
"loss": 0.354,
|
| 281123 |
+
"step": 142860
|
| 281124 |
+
},
|
| 281125 |
+
{
|
| 281126 |
+
"epoch": 1151.99,
|
| 281127 |
+
"learning_rate": 7.70907915993538e-06,
|
| 281128 |
+
"loss": 0.7848,
|
| 281129 |
+
"step": 142865
|
| 281130 |
+
},
|
| 281131 |
+
{
|
| 281132 |
+
"epoch": 1152.0,
|
| 281133 |
+
"eval_loss": 0.3314845860004425,
|
| 281134 |
+
"eval_runtime": 41.7288,
|
| 281135 |
+
"eval_samples_per_second": 20.058,
|
| 281136 |
+
"eval_steps_per_second": 0.647,
|
| 281137 |
+
"eval_wer": 0.18063857705204842,
|
| 281138 |
+
"step": 142866
|
| 281139 |
+
},
|
| 281140 |
+
{
|
| 281141 |
+
"epoch": 1142.03,
|
| 281142 |
+
"learning_rate": 7.708998384491116e-06,
|
| 281143 |
+
"loss": 0.37,
|
| 281144 |
+
"step": 142870
|
| 281145 |
+
},
|
| 281146 |
+
{
|
| 281147 |
+
"epoch": 1142.07,
|
| 281148 |
+
"learning_rate": 7.708917609046852e-06,
|
| 281149 |
+
"loss": 0.268,
|
| 281150 |
+
"step": 142875
|
| 281151 |
+
},
|
| 281152 |
+
{
|
| 281153 |
+
"epoch": 1142.11,
|
| 281154 |
+
"learning_rate": 7.708836833602586e-06,
|
| 281155 |
+
"loss": 0.3045,
|
| 281156 |
+
"step": 142880
|
| 281157 |
+
},
|
| 281158 |
+
{
|
| 281159 |
+
"epoch": 1142.15,
|
| 281160 |
+
"learning_rate": 7.708756058158322e-06,
|
| 281161 |
+
"loss": 0.3686,
|
| 281162 |
+
"step": 142885
|
| 281163 |
+
},
|
| 281164 |
+
{
|
| 281165 |
+
"epoch": 1142.19,
|
| 281166 |
+
"learning_rate": 7.708675282714056e-06,
|
| 281167 |
+
"loss": 0.7276,
|
| 281168 |
+
"step": 142890
|
| 281169 |
+
},
|
| 281170 |
+
{
|
| 281171 |
+
"epoch": 1142.23,
|
| 281172 |
+
"learning_rate": 7.708594507269792e-06,
|
| 281173 |
+
"loss": 0.6929,
|
| 281174 |
+
"step": 142895
|
| 281175 |
+
},
|
| 281176 |
+
{
|
| 281177 |
+
"epoch": 1142.27,
|
| 281178 |
+
"learning_rate": 7.708513731825526e-06,
|
| 281179 |
+
"loss": 0.3062,
|
| 281180 |
+
"step": 142900
|
| 281181 |
+
},
|
| 281182 |
+
{
|
| 281183 |
+
"epoch": 1142.31,
|
| 281184 |
+
"learning_rate": 7.708432956381262e-06,
|
| 281185 |
+
"loss": 0.2826,
|
| 281186 |
+
"step": 142905
|
| 281187 |
+
},
|
| 281188 |
+
{
|
| 281189 |
+
"epoch": 1142.35,
|
| 281190 |
+
"learning_rate": 7.708352180936996e-06,
|
| 281191 |
+
"loss": 0.3328,
|
| 281192 |
+
"step": 142910
|
| 281193 |
+
},
|
| 281194 |
+
{
|
| 281195 |
+
"epoch": 1142.39,
|
| 281196 |
+
"learning_rate": 7.708271405492732e-06,
|
| 281197 |
+
"loss": 0.8634,
|
| 281198 |
+
"step": 142915
|
| 281199 |
+
},
|
| 281200 |
+
{
|
| 281201 |
+
"epoch": 1142.43,
|
| 281202 |
+
"learning_rate": 7.708190630048466e-06,
|
| 281203 |
+
"loss": 0.7105,
|
| 281204 |
+
"step": 142920
|
| 281205 |
+
},
|
| 281206 |
+
{
|
| 281207 |
+
"epoch": 1142.47,
|
| 281208 |
+
"learning_rate": 7.708109854604202e-06,
|
| 281209 |
+
"loss": 0.274,
|
| 281210 |
+
"step": 142925
|
| 281211 |
+
},
|
| 281212 |
+
{
|
| 281213 |
+
"epoch": 1142.51,
|
| 281214 |
+
"learning_rate": 7.708029079159936e-06,
|
| 281215 |
+
"loss": 0.3273,
|
| 281216 |
+
"step": 142930
|
| 281217 |
+
},
|
| 281218 |
+
{
|
| 281219 |
+
"epoch": 1142.55,
|
| 281220 |
+
"learning_rate": 7.707948303715672e-06,
|
| 281221 |
+
"loss": 0.41,
|
| 281222 |
+
"step": 142935
|
| 281223 |
+
},
|
| 281224 |
+
{
|
| 281225 |
+
"epoch": 1142.59,
|
| 281226 |
+
"learning_rate": 7.707867528271407e-06,
|
| 281227 |
+
"loss": 0.8705,
|
| 281228 |
+
"step": 142940
|
| 281229 |
+
},
|
| 281230 |
+
{
|
| 281231 |
+
"epoch": 1142.63,
|
| 281232 |
+
"learning_rate": 7.707786752827142e-06,
|
| 281233 |
+
"loss": 0.6095,
|
| 281234 |
+
"step": 142945
|
| 281235 |
+
},
|
| 281236 |
+
{
|
| 281237 |
+
"epoch": 1142.67,
|
| 281238 |
+
"learning_rate": 7.707705977382877e-06,
|
| 281239 |
+
"loss": 0.3094,
|
| 281240 |
+
"step": 142950
|
| 281241 |
+
},
|
| 281242 |
+
{
|
| 281243 |
+
"epoch": 1142.71,
|
| 281244 |
+
"learning_rate": 7.707625201938612e-06,
|
| 281245 |
+
"loss": 0.3067,
|
| 281246 |
+
"step": 142955
|
| 281247 |
+
},
|
| 281248 |
+
{
|
| 281249 |
+
"epoch": 1142.75,
|
| 281250 |
+
"learning_rate": 7.707544426494347e-06,
|
| 281251 |
+
"loss": 0.3958,
|
| 281252 |
+
"step": 142960
|
| 281253 |
+
},
|
| 281254 |
+
{
|
| 281255 |
+
"epoch": 1142.79,
|
| 281256 |
+
"learning_rate": 7.707463651050082e-06,
|
| 281257 |
+
"loss": 0.9498,
|
| 281258 |
+
"step": 142965
|
| 281259 |
+
},
|
| 281260 |
+
{
|
| 281261 |
+
"epoch": 1142.83,
|
| 281262 |
+
"learning_rate": 7.707382875605817e-06,
|
| 281263 |
+
"loss": 0.6889,
|
| 281264 |
+
"step": 142970
|
| 281265 |
+
},
|
| 281266 |
+
{
|
| 281267 |
+
"epoch": 1142.87,
|
| 281268 |
+
"learning_rate": 7.707302100161552e-06,
|
| 281269 |
+
"loss": 0.2645,
|
| 281270 |
+
"step": 142975
|
| 281271 |
+
},
|
| 281272 |
+
{
|
| 281273 |
+
"epoch": 1142.91,
|
| 281274 |
+
"learning_rate": 7.707221324717287e-06,
|
| 281275 |
+
"loss": 0.291,
|
| 281276 |
+
"step": 142980
|
| 281277 |
+
},
|
| 281278 |
+
{
|
| 281279 |
+
"epoch": 1142.95,
|
| 281280 |
+
"learning_rate": 7.707140549273021e-06,
|
| 281281 |
+
"loss": 0.4395,
|
| 281282 |
+
"step": 142985
|
| 281283 |
+
},
|
| 281284 |
+
{
|
| 281285 |
+
"epoch": 1142.99,
|
| 281286 |
+
"learning_rate": 7.707059773828757e-06,
|
| 281287 |
+
"loss": 1.038,
|
| 281288 |
+
"step": 142990
|
| 281289 |
+
},
|
| 281290 |
+
{
|
| 281291 |
+
"epoch": 1143.0,
|
| 281292 |
+
"eval_loss": 0.3770361542701721,
|
| 281293 |
+
"eval_runtime": 41.6024,
|
| 281294 |
+
"eval_samples_per_second": 20.143,
|
| 281295 |
+
"eval_steps_per_second": 0.649,
|
| 281296 |
+
"eval_wer": 0.176540196731011,
|
| 281297 |
+
"step": 142991
|
| 281298 |
+
},
|
| 281299 |
+
{
|
| 281300 |
+
"epoch": 1153.03,
|
| 281301 |
+
"learning_rate": 7.706978998384491e-06,
|
| 281302 |
+
"loss": 0.2919,
|
| 281303 |
+
"step": 142995
|
| 281304 |
+
},
|
| 281305 |
+
{
|
| 281306 |
+
"epoch": 1153.07,
|
| 281307 |
+
"learning_rate": 7.706898222940227e-06,
|
| 281308 |
+
"loss": 0.243,
|
| 281309 |
+
"step": 143000
|
| 281310 |
+
},
|
| 281311 |
+
{
|
| 281312 |
+
"epoch": 1153.11,
|
| 281313 |
+
"learning_rate": 7.706817447495963e-06,
|
| 281314 |
+
"loss": 0.2585,
|
| 281315 |
+
"step": 143005
|
| 281316 |
+
},
|
| 281317 |
+
{
|
| 281318 |
+
"epoch": 1153.15,
|
| 281319 |
+
"learning_rate": 7.706736672051697e-06,
|
| 281320 |
+
"loss": 0.4033,
|
| 281321 |
+
"step": 143010
|
| 281322 |
+
},
|
| 281323 |
+
{
|
| 281324 |
+
"epoch": 1153.19,
|
| 281325 |
+
"learning_rate": 7.706655896607433e-06,
|
| 281326 |
+
"loss": 0.854,
|
| 281327 |
+
"step": 143015
|
| 281328 |
+
},
|
| 281329 |
+
{
|
| 281330 |
+
"epoch": 1153.23,
|
| 281331 |
+
"learning_rate": 7.706575121163167e-06,
|
| 281332 |
+
"loss": 0.7609,
|
| 281333 |
+
"step": 143020
|
| 281334 |
+
},
|
| 281335 |
+
{
|
| 281336 |
+
"epoch": 1153.27,
|
| 281337 |
+
"learning_rate": 7.706494345718903e-06,
|
| 281338 |
+
"loss": 0.2552,
|
| 281339 |
+
"step": 143025
|
| 281340 |
+
},
|
| 281341 |
+
{
|
| 281342 |
+
"epoch": 1153.31,
|
| 281343 |
+
"learning_rate": 7.706413570274637e-06,
|
| 281344 |
+
"loss": 0.3037,
|
| 281345 |
+
"step": 143030
|
| 281346 |
+
},
|
| 281347 |
+
{
|
| 281348 |
+
"epoch": 1153.35,
|
| 281349 |
+
"learning_rate": 7.706332794830373e-06,
|
| 281350 |
+
"loss": 0.3576,
|
| 281351 |
+
"step": 143035
|
| 281352 |
+
},
|
| 281353 |
+
{
|
| 281354 |
+
"epoch": 1153.39,
|
| 281355 |
+
"learning_rate": 7.706252019386107e-06,
|
| 281356 |
+
"loss": 0.8974,
|
| 281357 |
+
"step": 143040
|
| 281358 |
+
},
|
| 281359 |
+
{
|
| 281360 |
+
"epoch": 1153.43,
|
| 281361 |
+
"learning_rate": 7.706171243941843e-06,
|
| 281362 |
+
"loss": 0.6808,
|
| 281363 |
+
"step": 143045
|
| 281364 |
+
},
|
| 281365 |
+
{
|
| 281366 |
+
"epoch": 1153.47,
|
| 281367 |
+
"learning_rate": 7.706090468497577e-06,
|
| 281368 |
+
"loss": 0.2624,
|
| 281369 |
+
"step": 143050
|
| 281370 |
+
},
|
| 281371 |
+
{
|
| 281372 |
+
"epoch": 1153.51,
|
| 281373 |
+
"learning_rate": 7.706009693053313e-06,
|
| 281374 |
+
"loss": 0.3187,
|
| 281375 |
+
"step": 143055
|
| 281376 |
+
},
|
| 281377 |
+
{
|
| 281378 |
+
"epoch": 1153.55,
|
| 281379 |
+
"learning_rate": 7.705928917609047e-06,
|
| 281380 |
+
"loss": 0.3324,
|
| 281381 |
+
"step": 143060
|
| 281382 |
+
},
|
| 281383 |
+
{
|
| 281384 |
+
"epoch": 1153.59,
|
| 281385 |
+
"learning_rate": 7.705848142164783e-06,
|
| 281386 |
+
"loss": 0.8548,
|
| 281387 |
+
"step": 143065
|
| 281388 |
+
},
|
| 281389 |
+
{
|
| 281390 |
+
"epoch": 1153.63,
|
| 281391 |
+
"learning_rate": 7.705767366720517e-06,
|
| 281392 |
+
"loss": 0.5806,
|
| 281393 |
+
"step": 143070
|
| 281394 |
+
},
|
| 281395 |
+
{
|
| 281396 |
+
"epoch": 1153.67,
|
| 281397 |
+
"learning_rate": 7.705686591276253e-06,
|
| 281398 |
+
"loss": 0.2566,
|
| 281399 |
+
"step": 143075
|
| 281400 |
+
},
|
| 281401 |
+
{
|
| 281402 |
+
"epoch": 1153.71,
|
| 281403 |
+
"learning_rate": 7.705605815831989e-06,
|
| 281404 |
+
"loss": 0.2849,
|
| 281405 |
+
"step": 143080
|
| 281406 |
+
},
|
| 281407 |
+
{
|
| 281408 |
+
"epoch": 1153.76,
|
| 281409 |
+
"learning_rate": 7.705525040387723e-06,
|
| 281410 |
+
"loss": 0.3476,
|
| 281411 |
+
"step": 143085
|
| 281412 |
+
},
|
| 281413 |
+
{
|
| 281414 |
+
"epoch": 1153.8,
|
| 281415 |
+
"learning_rate": 7.705444264943459e-06,
|
| 281416 |
+
"loss": 0.8085,
|
| 281417 |
+
"step": 143090
|
| 281418 |
+
},
|
| 281419 |
+
{
|
| 281420 |
+
"epoch": 1153.84,
|
| 281421 |
+
"learning_rate": 7.705363489499193e-06,
|
| 281422 |
+
"loss": 0.6814,
|
| 281423 |
+
"step": 143095
|
| 281424 |
+
},
|
| 281425 |
+
{
|
| 281426 |
+
"epoch": 1153.88,
|
| 281427 |
+
"learning_rate": 7.705282714054929e-06,
|
| 281428 |
+
"loss": 0.2852,
|
| 281429 |
+
"step": 143100
|
| 281430 |
+
},
|
| 281431 |
+
{
|
| 281432 |
+
"epoch": 1153.92,
|
| 281433 |
+
"learning_rate": 7.705201938610663e-06,
|
| 281434 |
+
"loss": 0.2844,
|
| 281435 |
+
"step": 143105
|
| 281436 |
+
},
|
| 281437 |
+
{
|
| 281438 |
+
"epoch": 1153.96,
|
| 281439 |
+
"learning_rate": 7.705121163166399e-06,
|
| 281440 |
+
"loss": 0.4279,
|
| 281441 |
+
"step": 143110
|
| 281442 |
+
},
|
| 281443 |
+
{
|
| 281444 |
+
"epoch": 1154.0,
|
| 281445 |
+
"learning_rate": 7.705040387722133e-06,
|
| 281446 |
+
"loss": 1.0141,
|
| 281447 |
+
"step": 143115
|
| 281448 |
+
},
|
| 281449 |
+
{
|
| 281450 |
+
"epoch": 1154.0,
|
| 281451 |
+
"eval_loss": 0.31175512075424194,
|
| 281452 |
+
"eval_runtime": 41.2685,
|
| 281453 |
+
"eval_samples_per_second": 20.306,
|
| 281454 |
+
"eval_steps_per_second": 0.654,
|
| 281455 |
+
"eval_wer": 0.1807184750733138,
|
| 281456 |
+
"step": 143115
|
| 281457 |
+
},
|
| 281458 |
+
{
|
| 281459 |
+
"epoch": 1144.04,
|
| 281460 |
+
"learning_rate": 7.704959612277869e-06,
|
| 281461 |
+
"loss": 0.3268,
|
| 281462 |
+
"step": 143120
|
| 281463 |
+
},
|
| 281464 |
+
{
|
| 281465 |
+
"epoch": 1144.08,
|
| 281466 |
+
"learning_rate": 7.704878836833603e-06,
|
| 281467 |
+
"loss": 0.2871,
|
| 281468 |
+
"step": 143125
|
| 281469 |
+
},
|
| 281470 |
+
{
|
| 281471 |
+
"epoch": 1144.12,
|
| 281472 |
+
"learning_rate": 7.704798061389339e-06,
|
| 281473 |
+
"loss": 0.2926,
|
| 281474 |
+
"step": 143130
|
| 281475 |
+
},
|
| 281476 |
+
{
|
| 281477 |
+
"epoch": 1144.16,
|
| 281478 |
+
"learning_rate": 7.704717285945073e-06,
|
| 281479 |
+
"loss": 0.4266,
|
| 281480 |
+
"step": 143135
|
| 281481 |
+
},
|
| 281482 |
+
{
|
| 281483 |
+
"epoch": 1144.2,
|
| 281484 |
+
"learning_rate": 7.704636510500809e-06,
|
| 281485 |
+
"loss": 1.2651,
|
| 281486 |
+
"step": 143140
|
| 281487 |
+
},
|
| 281488 |
+
{
|
| 281489 |
+
"epoch": 1144.24,
|
| 281490 |
+
"learning_rate": 7.704555735056544e-06,
|
| 281491 |
+
"loss": 0.3356,
|
| 281492 |
+
"step": 143145
|
| 281493 |
+
},
|
| 281494 |
+
{
|
| 281495 |
+
"epoch": 1144.28,
|
| 281496 |
+
"learning_rate": 7.704474959612279e-06,
|
| 281497 |
+
"loss": 0.2414,
|
| 281498 |
+
"step": 143150
|
| 281499 |
+
},
|
| 281500 |
+
{
|
| 281501 |
+
"epoch": 1144.32,
|
| 281502 |
+
"learning_rate": 7.704394184168014e-06,
|
| 281503 |
+
"loss": 0.2823,
|
| 281504 |
+
"step": 143155
|
| 281505 |
+
},
|
| 281506 |
+
{
|
| 281507 |
+
"epoch": 1144.36,
|
| 281508 |
+
"learning_rate": 7.704313408723749e-06,
|
| 281509 |
+
"loss": 0.4548,
|
| 281510 |
+
"step": 143160
|
| 281511 |
+
},
|
| 281512 |
+
{
|
| 281513 |
+
"epoch": 1144.4,
|
| 281514 |
+
"learning_rate": 7.704232633279484e-06,
|
| 281515 |
+
"loss": 1.1442,
|
| 281516 |
+
"step": 143165
|
| 281517 |
+
},
|
| 281518 |
+
{
|
| 281519 |
+
"epoch": 1144.44,
|
| 281520 |
+
"learning_rate": 7.704151857835219e-06,
|
| 281521 |
+
"loss": 0.2731,
|
| 281522 |
+
"step": 143170
|
| 281523 |
+
},
|
| 281524 |
+
{
|
| 281525 |
+
"epoch": 1144.48,
|
| 281526 |
+
"learning_rate": 7.704071082390954e-06,
|
| 281527 |
+
"loss": 0.2685,
|
| 281528 |
+
"step": 143175
|
| 281529 |
+
},
|
| 281530 |
+
{
|
| 281531 |
+
"epoch": 1144.52,
|
| 281532 |
+
"learning_rate": 7.703990306946689e-06,
|
| 281533 |
+
"loss": 0.2771,
|
| 281534 |
+
"step": 143180
|
| 281535 |
+
},
|
| 281536 |
+
{
|
| 281537 |
+
"epoch": 1144.56,
|
| 281538 |
+
"learning_rate": 7.703909531502424e-06,
|
| 281539 |
+
"loss": 0.4552,
|
| 281540 |
+
"step": 143185
|
| 281541 |
+
},
|
| 281542 |
+
{
|
| 281543 |
+
"epoch": 1144.6,
|
| 281544 |
+
"learning_rate": 7.703828756058159e-06,
|
| 281545 |
+
"loss": 1.3455,
|
| 281546 |
+
"step": 143190
|
| 281547 |
+
},
|
| 281548 |
+
{
|
| 281549 |
+
"epoch": 1144.64,
|
| 281550 |
+
"learning_rate": 7.703747980613894e-06,
|
| 281551 |
+
"loss": 0.3369,
|
| 281552 |
+
"step": 143195
|
| 281553 |
+
},
|
| 281554 |
+
{
|
| 281555 |
+
"epoch": 1144.68,
|
| 281556 |
+
"learning_rate": 7.703667205169628e-06,
|
| 281557 |
+
"loss": 0.2937,
|
| 281558 |
+
"step": 143200
|
| 281559 |
+
},
|
| 281560 |
+
{
|
| 281561 |
+
"epoch": 1144.72,
|
| 281562 |
+
"learning_rate": 7.703586429725364e-06,
|
| 281563 |
+
"loss": 0.3249,
|
| 281564 |
+
"step": 143205
|
| 281565 |
+
},
|
| 281566 |
+
{
|
| 281567 |
+
"epoch": 1144.76,
|
| 281568 |
+
"learning_rate": 7.7035056542811e-06,
|
| 281569 |
+
"loss": 0.4669,
|
| 281570 |
+
"step": 143210
|
| 281571 |
+
},
|
| 281572 |
+
{
|
| 281573 |
+
"epoch": 1144.8,
|
| 281574 |
+
"learning_rate": 7.703424878836834e-06,
|
| 281575 |
+
"loss": 1.1763,
|
| 281576 |
+
"step": 143215
|
| 281577 |
+
},
|
| 281578 |
+
{
|
| 281579 |
+
"epoch": 1144.84,
|
| 281580 |
+
"learning_rate": 7.70334410339257e-06,
|
| 281581 |
+
"loss": 0.2941,
|
| 281582 |
+
"step": 143220
|
| 281583 |
+
},
|
| 281584 |
+
{
|
| 281585 |
+
"epoch": 1144.88,
|
| 281586 |
+
"learning_rate": 7.703263327948304e-06,
|
| 281587 |
+
"loss": 0.4048,
|
| 281588 |
+
"step": 143225
|
| 281589 |
+
},
|
| 281590 |
+
{
|
| 281591 |
+
"epoch": 1144.92,
|
| 281592 |
+
"learning_rate": 7.70318255250404e-06,
|
| 281593 |
+
"loss": 0.3128,
|
| 281594 |
+
"step": 143230
|
| 281595 |
+
},
|
| 281596 |
+
{
|
| 281597 |
+
"epoch": 1144.96,
|
| 281598 |
+
"learning_rate": 7.703101777059774e-06,
|
| 281599 |
+
"loss": 0.4631,
|
| 281600 |
+
"step": 143235
|
| 281601 |
+
},
|
| 281602 |
+
{
|
| 281603 |
+
"epoch": 1145.0,
|
| 281604 |
+
"learning_rate": 7.70302100161551e-06,
|
| 281605 |
+
"loss": 1.2677,
|
| 281606 |
+
"step": 143240
|
| 281607 |
+
},
|
| 281608 |
+
{
|
| 281609 |
+
"epoch": 1145.0,
|
| 281610 |
+
"eval_loss": 0.35869625210762024,
|
| 281611 |
+
"eval_runtime": 41.8075,
|
| 281612 |
+
"eval_samples_per_second": 20.044,
|
| 281613 |
+
"eval_steps_per_second": 0.646,
|
| 281614 |
+
"eval_wer": 0.16819527695529718,
|
| 281615 |
+
"step": 143240
|
| 281616 |
}
|
| 281617 |
],
|
| 281618 |
"max_steps": 625000,
|
| 281619 |
"num_train_epochs": 5000,
|
| 281620 |
+
"total_flos": 4.0307380836057815e+20,
|
| 281621 |
"trial_name": null,
|
| 281622 |
"trial_params": null
|
| 281623 |
}
|
model-bin/finetune/base/{checkpoint-132910 β checkpoint-143240}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-142742 β checkpoint-143365}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-142742 β checkpoint-143365}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9b4ae2f824fc44e74d5f922c65c684da85a296c2dd928113161d2e4f8fd01f68
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-142742 β checkpoint-143365}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-132910 β checkpoint-143365}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7e1221ef028ae6ba23c0df2fe3cfe15fa48e2d55a99582cb94df564ba0f055f1
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-132910 β checkpoint-143365}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:67d1a327091bff6c6eb8436fcc27644978a972446a5eebb7ff5c3adc62b63046
|
| 3 |
size 14503
|
model-bin/finetune/base/{checkpoint-132910 β checkpoint-143365}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dbaca6535201eae6dc1d868a0f5be4c8789d1fe82762cfc6215c355aacb74f62
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-142742 β checkpoint-143365}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f2b05ae9c3a5bf446ba77acdd3c9ade4ed6e50b98322ae9d0848f215de2ab73f
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-132910 β checkpoint-143365}/trainer_state.json
RENAMED
|
The diff for this file is too large to render.
See raw diff
|
|
|
model-bin/finetune/base/{checkpoint-142742 β checkpoint-143365}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1630217087.7673898/events.out.tfevents.1630217087.cc93b136ebf5.1086.59
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:09b8084a2ac97413aaf76a88b7e3092a97e277067c3796295860c9c4caf09d02
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630217533.1449723/events.out.tfevents.1630217533.cc93b136ebf5.1086.61
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:21212778424bd51abb4aacfd5252daebd31ce1f97f1f09ea1fcce286c46f9c93
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630217981.5941153/events.out.tfevents.1630217981.cc93b136ebf5.1086.63
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5d6b8848122be1124612dcd0db04e18faf74347bacb94c76b956338a454d50ad
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630218424.3441827/events.out.tfevents.1630218424.cc93b136ebf5.1086.65
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6c924e5278bf26844662007c8bd5145ebe4ea3994598afbebe9e83f677faefec
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630218868.8727193/events.out.tfevents.1630218868.cc93b136ebf5.1086.67
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3b19fa218f1be9449a5d32a64885bfecc29770580523ffc10dfeed6ccb873be0
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630217087.cc93b136ebf5.1086.58
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:907cf29111876985d236cafe3f953035e84c59b474c9dbd9dff4e969ab83f1c3
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630217533.cc93b136ebf5.1086.60
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ee0d13fac16746b67f94ac3c316b9c00b8b3dbf21d6bcf5e02048b7b1a93ea26
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630217981.cc93b136ebf5.1086.62
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c8b3555234e7b0bc800f9349f325108795cf1047ae6183e6ed8e6afb5c0295d9
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630218424.cc93b136ebf5.1086.64
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fccc790c4cb6a6b399c77a5a7fccd5ca694c52a3a290cd8636074e64786e9d37
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630218868.cc93b136ebf5.1086.66
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9ae733cf523888acb79e1284050d06ee1259396ee832443898d7c9e8c29a64f9
|
| 3 |
+
size 8622
|