"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-92342 β checkpoint-94333}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-92342 β checkpoint-94333}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-92342 β checkpoint-94333}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-93960 β checkpoint-94333}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-93960 β checkpoint-94333}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-93960 β checkpoint-94333}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-92342 β checkpoint-94333}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-93960 β checkpoint-94333}/trainer_state.json +476 -5
- model-bin/finetune/base/{checkpoint-92342 β checkpoint-94333}/training_args.bin +0 -0
- model-bin/finetune/base/{checkpoint-93960 β checkpoint-94581}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-93960 β checkpoint-94581}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-93960 β checkpoint-94581}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-92342 β checkpoint-94581}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-92342 β checkpoint-94581}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-92342 β checkpoint-94581}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-93960 β checkpoint-94581}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-92342 β checkpoint-94581}/trainer_state.json +2856 -6
- model-bin/finetune/base/{checkpoint-93960 β checkpoint-94581}/training_args.bin +0 -0
- model-bin/finetune/base/log/1629959918.4620056/events.out.tfevents.1629959918.8e89bd551565.924.91 +3 -0
- model-bin/finetune/base/log/1629960372.9279814/events.out.tfevents.1629960372.8e89bd551565.924.93 +3 -0
- model-bin/finetune/base/log/1629960821.1668708/events.out.tfevents.1629960821.8e89bd551565.924.95 +3 -0
- model-bin/finetune/base/log/1629961261.3675253/events.out.tfevents.1629961261.8e89bd551565.924.97 +3 -0
- model-bin/finetune/base/log/1629961689.1069834/events.out.tfevents.1629961689.8e89bd551565.924.99 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629959918.8e89bd551565.924.90 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629960372.8e89bd551565.924.92 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629960819.8e89bd551565.924.94 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629961261.8e89bd551565.924.96 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1629961689.8e89bd551565.924.98 +3 -0
model-bin/finetune/base/{checkpoint-92342 β checkpoint-94333}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-92342 β checkpoint-94333}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b6f6316e87ed422056e99effada8ed7f2016d5c18b208bfb5a8a30a7adc08f15
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-92342 β checkpoint-94333}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-93960 β checkpoint-94333}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bf4522fbe4ab5ca25ec5a37d3ebc83c3938b07c7cf4375a5a1f5ce0cb152b4b8
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-93960 β checkpoint-94333}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5523d5efd244168a713147374038c59a064ddbb8ee9b5072348dc04053f99bc8
|
| 3 |
size 14503
|
model-bin/finetune/base/{checkpoint-93960 β checkpoint-94333}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:db292c316a79f456016322ef81169ebe4f14d1db899cb806a80f69948b3990ec
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-92342 β checkpoint-94333}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:325366acd924839de48bc8b31155f3b34fc1d3125a0ea4e341fdc188f1dc8329
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-93960 β checkpoint-94333}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
-
"best_metric": 0.
|
| 3 |
-
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -218913,11 +218913,482 @@
|
|
| 218913 |
"eval_steps_per_second": 0.677,
|
| 218914 |
"eval_wer": 0.18971618401097712,
|
| 218915 |
"step": 93960
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 218916 |
}
|
| 218917 |
],
|
| 218918 |
"max_steps": 620000,
|
| 218919 |
"num_train_epochs": 5000,
|
| 218920 |
-
"total_flos": 2.
|
| 218921 |
"trial_name": null,
|
| 218922 |
"trial_params": null
|
| 218923 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"best_metric": 0.17565318086415285,
|
| 3 |
+
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-94333",
|
| 4 |
+
"epoch": 760.0,
|
| 5 |
+
"global_step": 94333,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 218913 |
"eval_steps_per_second": 0.677,
|
| 218914 |
"eval_wer": 0.18971618401097712,
|
| 218915 |
"step": 93960
|
| 218916 |
+
},
|
| 218917 |
+
{
|
| 218918 |
+
"epoch": 751.04,
|
| 218919 |
+
"learning_rate": 8.51076923076923e-06,
|
| 218920 |
+
"loss": 0.3278,
|
| 218921 |
+
"step": 93965
|
| 218922 |
+
},
|
| 218923 |
+
{
|
| 218924 |
+
"epoch": 751.08,
|
| 218925 |
+
"learning_rate": 8.510689102564104e-06,
|
| 218926 |
+
"loss": 0.2895,
|
| 218927 |
+
"step": 93970
|
| 218928 |
+
},
|
| 218929 |
+
{
|
| 218930 |
+
"epoch": 751.12,
|
| 218931 |
+
"learning_rate": 8.510608974358975e-06,
|
| 218932 |
+
"loss": 0.3708,
|
| 218933 |
+
"step": 93975
|
| 218934 |
+
},
|
| 218935 |
+
{
|
| 218936 |
+
"epoch": 751.16,
|
| 218937 |
+
"learning_rate": 8.510528846153846e-06,
|
| 218938 |
+
"loss": 0.5214,
|
| 218939 |
+
"step": 93980
|
| 218940 |
+
},
|
| 218941 |
+
{
|
| 218942 |
+
"epoch": 751.2,
|
| 218943 |
+
"learning_rate": 8.510448717948718e-06,
|
| 218944 |
+
"loss": 1.45,
|
| 218945 |
+
"step": 93985
|
| 218946 |
+
},
|
| 218947 |
+
{
|
| 218948 |
+
"epoch": 751.24,
|
| 218949 |
+
"learning_rate": 8.510368589743591e-06,
|
| 218950 |
+
"loss": 0.2901,
|
| 218951 |
+
"step": 93990
|
| 218952 |
+
},
|
| 218953 |
+
{
|
| 218954 |
+
"epoch": 751.28,
|
| 218955 |
+
"learning_rate": 8.510288461538462e-06,
|
| 218956 |
+
"loss": 0.2749,
|
| 218957 |
+
"step": 93995
|
| 218958 |
+
},
|
| 218959 |
+
{
|
| 218960 |
+
"epoch": 751.32,
|
| 218961 |
+
"learning_rate": 8.510208333333334e-06,
|
| 218962 |
+
"loss": 0.3866,
|
| 218963 |
+
"step": 94000
|
| 218964 |
+
},
|
| 218965 |
+
{
|
| 218966 |
+
"epoch": 751.36,
|
| 218967 |
+
"learning_rate": 8.510128205128207e-06,
|
| 218968 |
+
"loss": 0.5277,
|
| 218969 |
+
"step": 94005
|
| 218970 |
+
},
|
| 218971 |
+
{
|
| 218972 |
+
"epoch": 751.4,
|
| 218973 |
+
"learning_rate": 8.510048076923078e-06,
|
| 218974 |
+
"loss": 1.0501,
|
| 218975 |
+
"step": 94010
|
| 218976 |
+
},
|
| 218977 |
+
{
|
| 218978 |
+
"epoch": 751.44,
|
| 218979 |
+
"learning_rate": 8.50996794871795e-06,
|
| 218980 |
+
"loss": 0.3593,
|
| 218981 |
+
"step": 94015
|
| 218982 |
+
},
|
| 218983 |
+
{
|
| 218984 |
+
"epoch": 751.48,
|
| 218985 |
+
"learning_rate": 8.50988782051282e-06,
|
| 218986 |
+
"loss": 0.2661,
|
| 218987 |
+
"step": 94020
|
| 218988 |
+
},
|
| 218989 |
+
{
|
| 218990 |
+
"epoch": 751.52,
|
| 218991 |
+
"learning_rate": 8.509807692307694e-06,
|
| 218992 |
+
"loss": 0.3336,
|
| 218993 |
+
"step": 94025
|
| 218994 |
+
},
|
| 218995 |
+
{
|
| 218996 |
+
"epoch": 751.56,
|
| 218997 |
+
"learning_rate": 8.509727564102565e-06,
|
| 218998 |
+
"loss": 0.4038,
|
| 218999 |
+
"step": 94030
|
| 219000 |
+
},
|
| 219001 |
+
{
|
| 219002 |
+
"epoch": 751.6,
|
| 219003 |
+
"learning_rate": 8.509647435897437e-06,
|
| 219004 |
+
"loss": 1.0587,
|
| 219005 |
+
"step": 94035
|
| 219006 |
+
},
|
| 219007 |
+
{
|
| 219008 |
+
"epoch": 751.64,
|
| 219009 |
+
"learning_rate": 8.509567307692308e-06,
|
| 219010 |
+
"loss": 0.3351,
|
| 219011 |
+
"step": 94040
|
| 219012 |
+
},
|
| 219013 |
+
{
|
| 219014 |
+
"epoch": 751.68,
|
| 219015 |
+
"learning_rate": 8.509487179487181e-06,
|
| 219016 |
+
"loss": 0.2673,
|
| 219017 |
+
"step": 94045
|
| 219018 |
+
},
|
| 219019 |
+
{
|
| 219020 |
+
"epoch": 751.72,
|
| 219021 |
+
"learning_rate": 8.509407051282052e-06,
|
| 219022 |
+
"loss": 0.3,
|
| 219023 |
+
"step": 94050
|
| 219024 |
+
},
|
| 219025 |
+
{
|
| 219026 |
+
"epoch": 751.76,
|
| 219027 |
+
"learning_rate": 8.509326923076924e-06,
|
| 219028 |
+
"loss": 0.4632,
|
| 219029 |
+
"step": 94055
|
| 219030 |
+
},
|
| 219031 |
+
{
|
| 219032 |
+
"epoch": 751.8,
|
| 219033 |
+
"learning_rate": 8.509246794871797e-06,
|
| 219034 |
+
"loss": 1.0868,
|
| 219035 |
+
"step": 94060
|
| 219036 |
+
},
|
| 219037 |
+
{
|
| 219038 |
+
"epoch": 751.84,
|
| 219039 |
+
"learning_rate": 8.509166666666668e-06,
|
| 219040 |
+
"loss": 0.3317,
|
| 219041 |
+
"step": 94065
|
| 219042 |
+
},
|
| 219043 |
+
{
|
| 219044 |
+
"epoch": 751.88,
|
| 219045 |
+
"learning_rate": 8.50908653846154e-06,
|
| 219046 |
+
"loss": 0.2943,
|
| 219047 |
+
"step": 94070
|
| 219048 |
+
},
|
| 219049 |
+
{
|
| 219050 |
+
"epoch": 751.92,
|
| 219051 |
+
"learning_rate": 8.50900641025641e-06,
|
| 219052 |
+
"loss": 0.3133,
|
| 219053 |
+
"step": 94075
|
| 219054 |
+
},
|
| 219055 |
+
{
|
| 219056 |
+
"epoch": 751.96,
|
| 219057 |
+
"learning_rate": 8.508926282051284e-06,
|
| 219058 |
+
"loss": 0.498,
|
| 219059 |
+
"step": 94080
|
| 219060 |
+
},
|
| 219061 |
+
{
|
| 219062 |
+
"epoch": 752.0,
|
| 219063 |
+
"learning_rate": 8.508846153846153e-06,
|
| 219064 |
+
"loss": 1.3055,
|
| 219065 |
+
"step": 94085
|
| 219066 |
+
},
|
| 219067 |
+
{
|
| 219068 |
+
"epoch": 752.0,
|
| 219069 |
+
"eval_loss": 0.3784423768520355,
|
| 219070 |
+
"eval_runtime": 39.658,
|
| 219071 |
+
"eval_samples_per_second": 21.131,
|
| 219072 |
+
"eval_steps_per_second": 0.681,
|
| 219073 |
+
"eval_wer": 0.1819756115159824,
|
| 219074 |
+
"step": 94085
|
| 219075 |
+
},
|
| 219076 |
+
{
|
| 219077 |
+
"epoch": 758.04,
|
| 219078 |
+
"learning_rate": 8.508766025641027e-06,
|
| 219079 |
+
"loss": 0.3594,
|
| 219080 |
+
"step": 94090
|
| 219081 |
+
},
|
| 219082 |
+
{
|
| 219083 |
+
"epoch": 758.08,
|
| 219084 |
+
"learning_rate": 8.508685897435898e-06,
|
| 219085 |
+
"loss": 0.3195,
|
| 219086 |
+
"step": 94095
|
| 219087 |
+
},
|
| 219088 |
+
{
|
| 219089 |
+
"epoch": 758.12,
|
| 219090 |
+
"learning_rate": 8.50860576923077e-06,
|
| 219091 |
+
"loss": 0.3556,
|
| 219092 |
+
"step": 94100
|
| 219093 |
+
},
|
| 219094 |
+
{
|
| 219095 |
+
"epoch": 758.16,
|
| 219096 |
+
"learning_rate": 8.508525641025642e-06,
|
| 219097 |
+
"loss": 0.4796,
|
| 219098 |
+
"step": 94105
|
| 219099 |
+
},
|
| 219100 |
+
{
|
| 219101 |
+
"epoch": 758.2,
|
| 219102 |
+
"learning_rate": 8.508445512820514e-06,
|
| 219103 |
+
"loss": 1.2387,
|
| 219104 |
+
"step": 94110
|
| 219105 |
+
},
|
| 219106 |
+
{
|
| 219107 |
+
"epoch": 758.24,
|
| 219108 |
+
"learning_rate": 8.508365384615385e-06,
|
| 219109 |
+
"loss": 0.3018,
|
| 219110 |
+
"step": 94115
|
| 219111 |
+
},
|
| 219112 |
+
{
|
| 219113 |
+
"epoch": 758.28,
|
| 219114 |
+
"learning_rate": 8.508285256410256e-06,
|
| 219115 |
+
"loss": 0.2822,
|
| 219116 |
+
"step": 94120
|
| 219117 |
+
},
|
| 219118 |
+
{
|
| 219119 |
+
"epoch": 758.32,
|
| 219120 |
+
"learning_rate": 8.50820512820513e-06,
|
| 219121 |
+
"loss": 0.352,
|
| 219122 |
+
"step": 94125
|
| 219123 |
+
},
|
| 219124 |
+
{
|
| 219125 |
+
"epoch": 758.36,
|
| 219126 |
+
"learning_rate": 8.508125e-06,
|
| 219127 |
+
"loss": 0.5246,
|
| 219128 |
+
"step": 94130
|
| 219129 |
+
},
|
| 219130 |
+
{
|
| 219131 |
+
"epoch": 758.4,
|
| 219132 |
+
"learning_rate": 8.508044871794872e-06,
|
| 219133 |
+
"loss": 1.2258,
|
| 219134 |
+
"step": 94135
|
| 219135 |
+
},
|
| 219136 |
+
{
|
| 219137 |
+
"epoch": 758.44,
|
| 219138 |
+
"learning_rate": 8.507964743589744e-06,
|
| 219139 |
+
"loss": 0.3109,
|
| 219140 |
+
"step": 94140
|
| 219141 |
+
},
|
| 219142 |
+
{
|
| 219143 |
+
"epoch": 758.48,
|
| 219144 |
+
"learning_rate": 8.507884615384617e-06,
|
| 219145 |
+
"loss": 0.2892,
|
| 219146 |
+
"step": 94145
|
| 219147 |
+
},
|
| 219148 |
+
{
|
| 219149 |
+
"epoch": 758.52,
|
| 219150 |
+
"learning_rate": 8.507804487179488e-06,
|
| 219151 |
+
"loss": 0.3321,
|
| 219152 |
+
"step": 94150
|
| 219153 |
+
},
|
| 219154 |
+
{
|
| 219155 |
+
"epoch": 758.56,
|
| 219156 |
+
"learning_rate": 8.50772435897436e-06,
|
| 219157 |
+
"loss": 0.468,
|
| 219158 |
+
"step": 94155
|
| 219159 |
+
},
|
| 219160 |
+
{
|
| 219161 |
+
"epoch": 758.6,
|
| 219162 |
+
"learning_rate": 8.507644230769232e-06,
|
| 219163 |
+
"loss": 1.1824,
|
| 219164 |
+
"step": 94160
|
| 219165 |
+
},
|
| 219166 |
+
{
|
| 219167 |
+
"epoch": 758.64,
|
| 219168 |
+
"learning_rate": 8.507564102564104e-06,
|
| 219169 |
+
"loss": 0.3597,
|
| 219170 |
+
"step": 94165
|
| 219171 |
+
},
|
| 219172 |
+
{
|
| 219173 |
+
"epoch": 758.68,
|
| 219174 |
+
"learning_rate": 8.507483974358975e-06,
|
| 219175 |
+
"loss": 0.331,
|
| 219176 |
+
"step": 94170
|
| 219177 |
+
},
|
| 219178 |
+
{
|
| 219179 |
+
"epoch": 758.72,
|
| 219180 |
+
"learning_rate": 8.507403846153846e-06,
|
| 219181 |
+
"loss": 0.3436,
|
| 219182 |
+
"step": 94175
|
| 219183 |
+
},
|
| 219184 |
+
{
|
| 219185 |
+
"epoch": 758.76,
|
| 219186 |
+
"learning_rate": 8.50732371794872e-06,
|
| 219187 |
+
"loss": 0.5242,
|
| 219188 |
+
"step": 94180
|
| 219189 |
+
},
|
| 219190 |
+
{
|
| 219191 |
+
"epoch": 758.8,
|
| 219192 |
+
"learning_rate": 8.50724358974359e-06,
|
| 219193 |
+
"loss": 1.3293,
|
| 219194 |
+
"step": 94185
|
| 219195 |
+
},
|
| 219196 |
+
{
|
| 219197 |
+
"epoch": 758.84,
|
| 219198 |
+
"learning_rate": 8.507163461538462e-06,
|
| 219199 |
+
"loss": 0.3482,
|
| 219200 |
+
"step": 94190
|
| 219201 |
+
},
|
| 219202 |
+
{
|
| 219203 |
+
"epoch": 758.88,
|
| 219204 |
+
"learning_rate": 8.507083333333334e-06,
|
| 219205 |
+
"loss": 0.28,
|
| 219206 |
+
"step": 94195
|
| 219207 |
+
},
|
| 219208 |
+
{
|
| 219209 |
+
"epoch": 758.92,
|
| 219210 |
+
"learning_rate": 8.507003205128207e-06,
|
| 219211 |
+
"loss": 0.287,
|
| 219212 |
+
"step": 94200
|
| 219213 |
+
},
|
| 219214 |
+
{
|
| 219215 |
+
"epoch": 758.96,
|
| 219216 |
+
"learning_rate": 8.506923076923078e-06,
|
| 219217 |
+
"loss": 0.5444,
|
| 219218 |
+
"step": 94205
|
| 219219 |
+
},
|
| 219220 |
+
{
|
| 219221 |
+
"epoch": 759.0,
|
| 219222 |
+
"eval_loss": 0.43794122338294983,
|
| 219223 |
+
"eval_runtime": 38.5307,
|
| 219224 |
+
"eval_samples_per_second": 21.749,
|
| 219225 |
+
"eval_steps_per_second": 0.701,
|
| 219226 |
+
"eval_wer": 0.1961866902237927,
|
| 219227 |
+
"step": 94209
|
| 219228 |
+
},
|
| 219229 |
+
{
|
| 219230 |
+
"epoch": 759.01,
|
| 219231 |
+
"learning_rate": 8.50684294871795e-06,
|
| 219232 |
+
"loss": 0.5366,
|
| 219233 |
+
"step": 94210
|
| 219234 |
+
},
|
| 219235 |
+
{
|
| 219236 |
+
"epoch": 759.05,
|
| 219237 |
+
"learning_rate": 8.506762820512822e-06,
|
| 219238 |
+
"loss": 0.3192,
|
| 219239 |
+
"step": 94215
|
| 219240 |
+
},
|
| 219241 |
+
{
|
| 219242 |
+
"epoch": 759.09,
|
| 219243 |
+
"learning_rate": 8.506682692307692e-06,
|
| 219244 |
+
"loss": 0.2532,
|
| 219245 |
+
"step": 94220
|
| 219246 |
+
},
|
| 219247 |
+
{
|
| 219248 |
+
"epoch": 759.13,
|
| 219249 |
+
"learning_rate": 8.506602564102565e-06,
|
| 219250 |
+
"loss": 0.3294,
|
| 219251 |
+
"step": 94225
|
| 219252 |
+
},
|
| 219253 |
+
{
|
| 219254 |
+
"epoch": 759.17,
|
| 219255 |
+
"learning_rate": 8.506522435897436e-06,
|
| 219256 |
+
"loss": 0.5267,
|
| 219257 |
+
"step": 94230
|
| 219258 |
+
},
|
| 219259 |
+
{
|
| 219260 |
+
"epoch": 759.21,
|
| 219261 |
+
"learning_rate": 8.506442307692308e-06,
|
| 219262 |
+
"loss": 1.1953,
|
| 219263 |
+
"step": 94235
|
| 219264 |
+
},
|
| 219265 |
+
{
|
| 219266 |
+
"epoch": 759.25,
|
| 219267 |
+
"learning_rate": 8.50636217948718e-06,
|
| 219268 |
+
"loss": 0.33,
|
| 219269 |
+
"step": 94240
|
| 219270 |
+
},
|
| 219271 |
+
{
|
| 219272 |
+
"epoch": 759.29,
|
| 219273 |
+
"learning_rate": 8.506282051282052e-06,
|
| 219274 |
+
"loss": 0.3258,
|
| 219275 |
+
"step": 94245
|
| 219276 |
+
},
|
| 219277 |
+
{
|
| 219278 |
+
"epoch": 759.33,
|
| 219279 |
+
"learning_rate": 8.506201923076924e-06,
|
| 219280 |
+
"loss": 0.4281,
|
| 219281 |
+
"step": 94250
|
| 219282 |
+
},
|
| 219283 |
+
{
|
| 219284 |
+
"epoch": 759.37,
|
| 219285 |
+
"learning_rate": 8.506121794871795e-06,
|
| 219286 |
+
"loss": 0.5508,
|
| 219287 |
+
"step": 94255
|
| 219288 |
+
},
|
| 219289 |
+
{
|
| 219290 |
+
"epoch": 759.41,
|
| 219291 |
+
"learning_rate": 8.506041666666668e-06,
|
| 219292 |
+
"loss": 1.1474,
|
| 219293 |
+
"step": 94260
|
| 219294 |
+
},
|
| 219295 |
+
{
|
| 219296 |
+
"epoch": 759.45,
|
| 219297 |
+
"learning_rate": 8.50596153846154e-06,
|
| 219298 |
+
"loss": 0.3027,
|
| 219299 |
+
"step": 94265
|
| 219300 |
+
},
|
| 219301 |
+
{
|
| 219302 |
+
"epoch": 759.49,
|
| 219303 |
+
"learning_rate": 8.50588141025641e-06,
|
| 219304 |
+
"loss": 0.3026,
|
| 219305 |
+
"step": 94270
|
| 219306 |
+
},
|
| 219307 |
+
{
|
| 219308 |
+
"epoch": 759.53,
|
| 219309 |
+
"learning_rate": 8.505801282051282e-06,
|
| 219310 |
+
"loss": 0.3823,
|
| 219311 |
+
"step": 94275
|
| 219312 |
+
},
|
| 219313 |
+
{
|
| 219314 |
+
"epoch": 759.57,
|
| 219315 |
+
"learning_rate": 8.505721153846155e-06,
|
| 219316 |
+
"loss": 0.5083,
|
| 219317 |
+
"step": 94280
|
| 219318 |
+
},
|
| 219319 |
+
{
|
| 219320 |
+
"epoch": 759.61,
|
| 219321 |
+
"learning_rate": 8.505641025641026e-06,
|
| 219322 |
+
"loss": 1.4258,
|
| 219323 |
+
"step": 94285
|
| 219324 |
+
},
|
| 219325 |
+
{
|
| 219326 |
+
"epoch": 759.65,
|
| 219327 |
+
"learning_rate": 8.505560897435898e-06,
|
| 219328 |
+
"loss": 0.3108,
|
| 219329 |
+
"step": 94290
|
| 219330 |
+
},
|
| 219331 |
+
{
|
| 219332 |
+
"epoch": 759.69,
|
| 219333 |
+
"learning_rate": 8.50548076923077e-06,
|
| 219334 |
+
"loss": 0.3102,
|
| 219335 |
+
"step": 94295
|
| 219336 |
+
},
|
| 219337 |
+
{
|
| 219338 |
+
"epoch": 759.73,
|
| 219339 |
+
"learning_rate": 8.505400641025642e-06,
|
| 219340 |
+
"loss": 0.3525,
|
| 219341 |
+
"step": 94300
|
| 219342 |
+
},
|
| 219343 |
+
{
|
| 219344 |
+
"epoch": 759.77,
|
| 219345 |
+
"learning_rate": 8.505320512820514e-06,
|
| 219346 |
+
"loss": 0.5332,
|
| 219347 |
+
"step": 94305
|
| 219348 |
+
},
|
| 219349 |
+
{
|
| 219350 |
+
"epoch": 759.81,
|
| 219351 |
+
"learning_rate": 8.505240384615385e-06,
|
| 219352 |
+
"loss": 1.0223,
|
| 219353 |
+
"step": 94310
|
| 219354 |
+
},
|
| 219355 |
+
{
|
| 219356 |
+
"epoch": 759.85,
|
| 219357 |
+
"learning_rate": 8.505160256410258e-06,
|
| 219358 |
+
"loss": 0.296,
|
| 219359 |
+
"step": 94315
|
| 219360 |
+
},
|
| 219361 |
+
{
|
| 219362 |
+
"epoch": 759.9,
|
| 219363 |
+
"learning_rate": 8.50508012820513e-06,
|
| 219364 |
+
"loss": 0.3009,
|
| 219365 |
+
"step": 94320
|
| 219366 |
+
},
|
| 219367 |
+
{
|
| 219368 |
+
"epoch": 759.94,
|
| 219369 |
+
"learning_rate": 8.505e-06,
|
| 219370 |
+
"loss": 0.3876,
|
| 219371 |
+
"step": 94325
|
| 219372 |
+
},
|
| 219373 |
+
{
|
| 219374 |
+
"epoch": 759.98,
|
| 219375 |
+
"learning_rate": 8.504919871794872e-06,
|
| 219376 |
+
"loss": 0.6779,
|
| 219377 |
+
"step": 94330
|
| 219378 |
+
},
|
| 219379 |
+
{
|
| 219380 |
+
"epoch": 760.0,
|
| 219381 |
+
"eval_loss": 0.358523428440094,
|
| 219382 |
+
"eval_runtime": 40.0339,
|
| 219383 |
+
"eval_samples_per_second": 20.932,
|
| 219384 |
+
"eval_steps_per_second": 0.674,
|
| 219385 |
+
"eval_wer": 0.17565318086415285,
|
| 219386 |
+
"step": 94333
|
| 219387 |
}
|
| 219388 |
],
|
| 219389 |
"max_steps": 620000,
|
| 219390 |
"num_train_epochs": 5000,
|
| 219391 |
+
"total_flos": 2.654619731204131e+20,
|
| 219392 |
"trial_name": null,
|
| 219393 |
"trial_params": null
|
| 219394 |
}
|
model-bin/finetune/base/{checkpoint-92342 β checkpoint-94333}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-93960 β checkpoint-94581}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-93960 β checkpoint-94581}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d4d02ee328683843640348c701073e14687bfccfdc71dd9fac4672e9fd9f9112
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-93960 β checkpoint-94581}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-92342 β checkpoint-94581}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0f720cc011181397b9e12a242b445abc9374bf1844864f9647c8460e0122ddba
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-92342 β checkpoint-94581}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9c83d1cd1f2f56df5b6678bf4ea9faf6ece318c361cfade58002b290f3bef31e
|
| 3 |
size 14503
|
model-bin/finetune/base/{checkpoint-92342 β checkpoint-94581}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:235b7cb3eba8a035f2a61e31c99daf589d6fc366080c7ca91c504d71b69970bd
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-93960 β checkpoint-94581}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e879444cbf49a7fb371b024bb1c67573a08cf2ce27706931638f5e2dc0eb5745
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-92342 β checkpoint-94581}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
-
"best_metric": 0.
|
| 3 |
-
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -216852,11 +216852,2861 @@
|
|
| 216852 |
"eval_steps_per_second": 0.703,
|
| 216853 |
"eval_wer": 0.17637692697401752,
|
| 216854 |
"step": 92342
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 216855 |
}
|
| 216856 |
],
|
| 216857 |
-
"max_steps":
|
| 216858 |
"num_train_epochs": 5000,
|
| 216859 |
-
"total_flos": 2.
|
| 216860 |
"trial_name": null,
|
| 216861 |
"trial_params": null
|
| 216862 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"best_metric": 0.17565318086415285,
|
| 3 |
+
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-94333",
|
| 4 |
+
"epoch": 761.995983935743,
|
| 5 |
+
"global_step": 94581,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 216852 |
"eval_steps_per_second": 0.703,
|
| 216853 |
"eval_wer": 0.17637692697401752,
|
| 216854 |
"step": 92342
|
| 216855 |
+
},
|
| 216856 |
+
{
|
| 216857 |
+
"epoch": 738.02,
|
| 216858 |
+
"learning_rate": 8.53673076923077e-06,
|
| 216859 |
+
"loss": 0.3277,
|
| 216860 |
+
"step": 92345
|
| 216861 |
+
},
|
| 216862 |
+
{
|
| 216863 |
+
"epoch": 738.06,
|
| 216864 |
+
"learning_rate": 8.536650641025643e-06,
|
| 216865 |
+
"loss": 0.3399,
|
| 216866 |
+
"step": 92350
|
| 216867 |
+
},
|
| 216868 |
+
{
|
| 216869 |
+
"epoch": 738.1,
|
| 216870 |
+
"learning_rate": 8.536570512820513e-06,
|
| 216871 |
+
"loss": 0.2699,
|
| 216872 |
+
"step": 92355
|
| 216873 |
+
},
|
| 216874 |
+
{
|
| 216875 |
+
"epoch": 738.14,
|
| 216876 |
+
"learning_rate": 8.536490384615386e-06,
|
| 216877 |
+
"loss": 0.3721,
|
| 216878 |
+
"step": 92360
|
| 216879 |
+
},
|
| 216880 |
+
{
|
| 216881 |
+
"epoch": 738.18,
|
| 216882 |
+
"learning_rate": 8.536410256410257e-06,
|
| 216883 |
+
"loss": 0.8481,
|
| 216884 |
+
"step": 92365
|
| 216885 |
+
},
|
| 216886 |
+
{
|
| 216887 |
+
"epoch": 738.22,
|
| 216888 |
+
"learning_rate": 8.536330128205128e-06,
|
| 216889 |
+
"loss": 1.0099,
|
| 216890 |
+
"step": 92370
|
| 216891 |
+
},
|
| 216892 |
+
{
|
| 216893 |
+
"epoch": 738.26,
|
| 216894 |
+
"learning_rate": 8.53625e-06,
|
| 216895 |
+
"loss": 0.2858,
|
| 216896 |
+
"step": 92375
|
| 216897 |
+
},
|
| 216898 |
+
{
|
| 216899 |
+
"epoch": 738.3,
|
| 216900 |
+
"learning_rate": 8.536169871794873e-06,
|
| 216901 |
+
"loss": 0.2905,
|
| 216902 |
+
"step": 92380
|
| 216903 |
+
},
|
| 216904 |
+
{
|
| 216905 |
+
"epoch": 738.34,
|
| 216906 |
+
"learning_rate": 8.536089743589744e-06,
|
| 216907 |
+
"loss": 0.4036,
|
| 216908 |
+
"step": 92385
|
| 216909 |
+
},
|
| 216910 |
+
{
|
| 216911 |
+
"epoch": 738.38,
|
| 216912 |
+
"learning_rate": 8.536009615384616e-06,
|
| 216913 |
+
"loss": 0.7687,
|
| 216914 |
+
"step": 92390
|
| 216915 |
+
},
|
| 216916 |
+
{
|
| 216917 |
+
"epoch": 738.42,
|
| 216918 |
+
"learning_rate": 8.535929487179487e-06,
|
| 216919 |
+
"loss": 0.8671,
|
| 216920 |
+
"step": 92395
|
| 216921 |
+
},
|
| 216922 |
+
{
|
| 216923 |
+
"epoch": 738.46,
|
| 216924 |
+
"learning_rate": 8.53584935897436e-06,
|
| 216925 |
+
"loss": 0.2939,
|
| 216926 |
+
"step": 92400
|
| 216927 |
+
},
|
| 216928 |
+
{
|
| 216929 |
+
"epoch": 738.5,
|
| 216930 |
+
"learning_rate": 8.535769230769231e-06,
|
| 216931 |
+
"loss": 0.278,
|
| 216932 |
+
"step": 92405
|
| 216933 |
+
},
|
| 216934 |
+
{
|
| 216935 |
+
"epoch": 738.54,
|
| 216936 |
+
"learning_rate": 8.535689102564103e-06,
|
| 216937 |
+
"loss": 0.3251,
|
| 216938 |
+
"step": 92410
|
| 216939 |
+
},
|
| 216940 |
+
{
|
| 216941 |
+
"epoch": 738.58,
|
| 216942 |
+
"learning_rate": 8.535608974358976e-06,
|
| 216943 |
+
"loss": 0.8,
|
| 216944 |
+
"step": 92415
|
| 216945 |
+
},
|
| 216946 |
+
{
|
| 216947 |
+
"epoch": 738.62,
|
| 216948 |
+
"learning_rate": 8.535528846153847e-06,
|
| 216949 |
+
"loss": 0.9726,
|
| 216950 |
+
"step": 92420
|
| 216951 |
+
},
|
| 216952 |
+
{
|
| 216953 |
+
"epoch": 738.66,
|
| 216954 |
+
"learning_rate": 8.535448717948718e-06,
|
| 216955 |
+
"loss": 0.283,
|
| 216956 |
+
"step": 92425
|
| 216957 |
+
},
|
| 216958 |
+
{
|
| 216959 |
+
"epoch": 738.7,
|
| 216960 |
+
"learning_rate": 8.53536858974359e-06,
|
| 216961 |
+
"loss": 0.2983,
|
| 216962 |
+
"step": 92430
|
| 216963 |
+
},
|
| 216964 |
+
{
|
| 216965 |
+
"epoch": 738.74,
|
| 216966 |
+
"learning_rate": 8.535288461538463e-06,
|
| 216967 |
+
"loss": 0.4433,
|
| 216968 |
+
"step": 92435
|
| 216969 |
+
},
|
| 216970 |
+
{
|
| 216971 |
+
"epoch": 738.78,
|
| 216972 |
+
"learning_rate": 8.535208333333334e-06,
|
| 216973 |
+
"loss": 0.7029,
|
| 216974 |
+
"step": 92440
|
| 216975 |
+
},
|
| 216976 |
+
{
|
| 216977 |
+
"epoch": 738.82,
|
| 216978 |
+
"learning_rate": 8.535128205128206e-06,
|
| 216979 |
+
"loss": 0.8049,
|
| 216980 |
+
"step": 92445
|
| 216981 |
+
},
|
| 216982 |
+
{
|
| 216983 |
+
"epoch": 738.86,
|
| 216984 |
+
"learning_rate": 8.535048076923079e-06,
|
| 216985 |
+
"loss": 0.4668,
|
| 216986 |
+
"step": 92450
|
| 216987 |
+
},
|
| 216988 |
+
{
|
| 216989 |
+
"epoch": 738.9,
|
| 216990 |
+
"learning_rate": 8.53496794871795e-06,
|
| 216991 |
+
"loss": 0.2845,
|
| 216992 |
+
"step": 92455
|
| 216993 |
+
},
|
| 216994 |
+
{
|
| 216995 |
+
"epoch": 738.94,
|
| 216996 |
+
"learning_rate": 8.534887820512821e-06,
|
| 216997 |
+
"loss": 0.3941,
|
| 216998 |
+
"step": 92460
|
| 216999 |
+
},
|
| 217000 |
+
{
|
| 217001 |
+
"epoch": 738.98,
|
| 217002 |
+
"learning_rate": 8.534807692307693e-06,
|
| 217003 |
+
"loss": 0.9213,
|
| 217004 |
+
"step": 92465
|
| 217005 |
+
},
|
| 217006 |
+
{
|
| 217007 |
+
"epoch": 739.0,
|
| 217008 |
+
"eval_loss": 0.4267621338367462,
|
| 217009 |
+
"eval_runtime": 40.1319,
|
| 217010 |
+
"eval_samples_per_second": 20.881,
|
| 217011 |
+
"eval_steps_per_second": 0.673,
|
| 217012 |
+
"eval_wer": 0.1924955595026643,
|
| 217013 |
+
"step": 92467
|
| 217014 |
+
},
|
| 217015 |
+
{
|
| 217016 |
+
"epoch": 739.02,
|
| 217017 |
+
"learning_rate": 8.534727564102566e-06,
|
| 217018 |
+
"loss": 0.4611,
|
| 217019 |
+
"step": 92470
|
| 217020 |
+
},
|
| 217021 |
+
{
|
| 217022 |
+
"epoch": 739.06,
|
| 217023 |
+
"learning_rate": 8.534647435897435e-06,
|
| 217024 |
+
"loss": 0.2818,
|
| 217025 |
+
"step": 92475
|
| 217026 |
+
},
|
| 217027 |
+
{
|
| 217028 |
+
"epoch": 739.1,
|
| 217029 |
+
"learning_rate": 8.534567307692308e-06,
|
| 217030 |
+
"loss": 0.3168,
|
| 217031 |
+
"step": 92480
|
| 217032 |
+
},
|
| 217033 |
+
{
|
| 217034 |
+
"epoch": 739.14,
|
| 217035 |
+
"learning_rate": 8.53448717948718e-06,
|
| 217036 |
+
"loss": 0.3464,
|
| 217037 |
+
"step": 92485
|
| 217038 |
+
},
|
| 217039 |
+
{
|
| 217040 |
+
"epoch": 739.18,
|
| 217041 |
+
"learning_rate": 8.534407051282051e-06,
|
| 217042 |
+
"loss": 0.7226,
|
| 217043 |
+
"step": 92490
|
| 217044 |
+
},
|
| 217045 |
+
{
|
| 217046 |
+
"epoch": 739.22,
|
| 217047 |
+
"learning_rate": 8.534326923076923e-06,
|
| 217048 |
+
"loss": 0.8778,
|
| 217049 |
+
"step": 92495
|
| 217050 |
+
},
|
| 217051 |
+
{
|
| 217052 |
+
"epoch": 739.26,
|
| 217053 |
+
"learning_rate": 8.534246794871796e-06,
|
| 217054 |
+
"loss": 0.3689,
|
| 217055 |
+
"step": 92500
|
| 217056 |
+
},
|
| 217057 |
+
{
|
| 217058 |
+
"epoch": 739.3,
|
| 217059 |
+
"learning_rate": 8.534166666666667e-06,
|
| 217060 |
+
"loss": 0.3223,
|
| 217061 |
+
"step": 92505
|
| 217062 |
+
},
|
| 217063 |
+
{
|
| 217064 |
+
"epoch": 739.34,
|
| 217065 |
+
"learning_rate": 8.534086538461538e-06,
|
| 217066 |
+
"loss": 0.3403,
|
| 217067 |
+
"step": 92510
|
| 217068 |
+
},
|
| 217069 |
+
{
|
| 217070 |
+
"epoch": 739.38,
|
| 217071 |
+
"learning_rate": 8.534006410256411e-06,
|
| 217072 |
+
"loss": 0.7436,
|
| 217073 |
+
"step": 92515
|
| 217074 |
+
},
|
| 217075 |
+
{
|
| 217076 |
+
"epoch": 739.42,
|
| 217077 |
+
"learning_rate": 8.533926282051283e-06,
|
| 217078 |
+
"loss": 0.9088,
|
| 217079 |
+
"step": 92520
|
| 217080 |
+
},
|
| 217081 |
+
{
|
| 217082 |
+
"epoch": 739.46,
|
| 217083 |
+
"learning_rate": 8.533846153846154e-06,
|
| 217084 |
+
"loss": 0.2664,
|
| 217085 |
+
"step": 92525
|
| 217086 |
+
},
|
| 217087 |
+
{
|
| 217088 |
+
"epoch": 739.5,
|
| 217089 |
+
"learning_rate": 8.533766025641025e-06,
|
| 217090 |
+
"loss": 0.2987,
|
| 217091 |
+
"step": 92530
|
| 217092 |
+
},
|
| 217093 |
+
{
|
| 217094 |
+
"epoch": 739.54,
|
| 217095 |
+
"learning_rate": 8.533685897435899e-06,
|
| 217096 |
+
"loss": 0.3624,
|
| 217097 |
+
"step": 92535
|
| 217098 |
+
},
|
| 217099 |
+
{
|
| 217100 |
+
"epoch": 739.58,
|
| 217101 |
+
"learning_rate": 8.53360576923077e-06,
|
| 217102 |
+
"loss": 0.8594,
|
| 217103 |
+
"step": 92540
|
| 217104 |
+
},
|
| 217105 |
+
{
|
| 217106 |
+
"epoch": 739.62,
|
| 217107 |
+
"learning_rate": 8.533525641025641e-06,
|
| 217108 |
+
"loss": 0.9225,
|
| 217109 |
+
"step": 92545
|
| 217110 |
+
},
|
| 217111 |
+
{
|
| 217112 |
+
"epoch": 739.66,
|
| 217113 |
+
"learning_rate": 8.533445512820514e-06,
|
| 217114 |
+
"loss": 0.3437,
|
| 217115 |
+
"step": 92550
|
| 217116 |
+
},
|
| 217117 |
+
{
|
| 217118 |
+
"epoch": 739.7,
|
| 217119 |
+
"learning_rate": 8.533365384615386e-06,
|
| 217120 |
+
"loss": 0.2656,
|
| 217121 |
+
"step": 92555
|
| 217122 |
+
},
|
| 217123 |
+
{
|
| 217124 |
+
"epoch": 739.74,
|
| 217125 |
+
"learning_rate": 8.533285256410257e-06,
|
| 217126 |
+
"loss": 0.4011,
|
| 217127 |
+
"step": 92560
|
| 217128 |
+
},
|
| 217129 |
+
{
|
| 217130 |
+
"epoch": 739.78,
|
| 217131 |
+
"learning_rate": 8.533205128205128e-06,
|
| 217132 |
+
"loss": 0.7358,
|
| 217133 |
+
"step": 92565
|
| 217134 |
+
},
|
| 217135 |
+
{
|
| 217136 |
+
"epoch": 739.82,
|
| 217137 |
+
"learning_rate": 8.533125000000001e-06,
|
| 217138 |
+
"loss": 0.8849,
|
| 217139 |
+
"step": 92570
|
| 217140 |
+
},
|
| 217141 |
+
{
|
| 217142 |
+
"epoch": 739.86,
|
| 217143 |
+
"learning_rate": 8.533044871794873e-06,
|
| 217144 |
+
"loss": 0.2759,
|
| 217145 |
+
"step": 92575
|
| 217146 |
+
},
|
| 217147 |
+
{
|
| 217148 |
+
"epoch": 739.9,
|
| 217149 |
+
"learning_rate": 8.532964743589744e-06,
|
| 217150 |
+
"loss": 0.3249,
|
| 217151 |
+
"step": 92580
|
| 217152 |
+
},
|
| 217153 |
+
{
|
| 217154 |
+
"epoch": 739.94,
|
| 217155 |
+
"learning_rate": 8.532884615384615e-06,
|
| 217156 |
+
"loss": 0.3239,
|
| 217157 |
+
"step": 92585
|
| 217158 |
+
},
|
| 217159 |
+
{
|
| 217160 |
+
"epoch": 739.98,
|
| 217161 |
+
"learning_rate": 8.532804487179489e-06,
|
| 217162 |
+
"loss": 0.8108,
|
| 217163 |
+
"step": 92590
|
| 217164 |
+
},
|
| 217165 |
+
{
|
| 217166 |
+
"epoch": 740.0,
|
| 217167 |
+
"eval_loss": 0.3362525403499603,
|
| 217168 |
+
"eval_runtime": 39.9292,
|
| 217169 |
+
"eval_samples_per_second": 21.012,
|
| 217170 |
+
"eval_steps_per_second": 0.676,
|
| 217171 |
+
"eval_wer": 0.18948505581562838,
|
| 217172 |
+
"step": 92592
|
| 217173 |
+
},
|
| 217174 |
+
{
|
| 217175 |
+
"epoch": 746.02,
|
| 217176 |
+
"learning_rate": 8.53272435897436e-06,
|
| 217177 |
+
"loss": 0.3747,
|
| 217178 |
+
"step": 92595
|
| 217179 |
+
},
|
| 217180 |
+
{
|
| 217181 |
+
"epoch": 746.06,
|
| 217182 |
+
"learning_rate": 8.532644230769231e-06,
|
| 217183 |
+
"loss": 0.2946,
|
| 217184 |
+
"step": 92600
|
| 217185 |
+
},
|
| 217186 |
+
{
|
| 217187 |
+
"epoch": 746.1,
|
| 217188 |
+
"learning_rate": 8.532564102564104e-06,
|
| 217189 |
+
"loss": 0.3287,
|
| 217190 |
+
"step": 92605
|
| 217191 |
+
},
|
| 217192 |
+
{
|
| 217193 |
+
"epoch": 746.15,
|
| 217194 |
+
"learning_rate": 8.532483974358976e-06,
|
| 217195 |
+
"loss": 0.3471,
|
| 217196 |
+
"step": 92610
|
| 217197 |
+
},
|
| 217198 |
+
{
|
| 217199 |
+
"epoch": 746.19,
|
| 217200 |
+
"learning_rate": 8.532403846153847e-06,
|
| 217201 |
+
"loss": 0.847,
|
| 217202 |
+
"step": 92615
|
| 217203 |
+
},
|
| 217204 |
+
{
|
| 217205 |
+
"epoch": 746.23,
|
| 217206 |
+
"learning_rate": 8.532323717948718e-06,
|
| 217207 |
+
"loss": 0.9097,
|
| 217208 |
+
"step": 92620
|
| 217209 |
+
},
|
| 217210 |
+
{
|
| 217211 |
+
"epoch": 746.27,
|
| 217212 |
+
"learning_rate": 8.532243589743591e-06,
|
| 217213 |
+
"loss": 0.291,
|
| 217214 |
+
"step": 92625
|
| 217215 |
+
},
|
| 217216 |
+
{
|
| 217217 |
+
"epoch": 746.31,
|
| 217218 |
+
"learning_rate": 8.532163461538461e-06,
|
| 217219 |
+
"loss": 0.3372,
|
| 217220 |
+
"step": 92630
|
| 217221 |
+
},
|
| 217222 |
+
{
|
| 217223 |
+
"epoch": 746.35,
|
| 217224 |
+
"learning_rate": 8.532083333333334e-06,
|
| 217225 |
+
"loss": 0.4195,
|
| 217226 |
+
"step": 92635
|
| 217227 |
+
},
|
| 217228 |
+
{
|
| 217229 |
+
"epoch": 746.39,
|
| 217230 |
+
"learning_rate": 8.532003205128206e-06,
|
| 217231 |
+
"loss": 0.7703,
|
| 217232 |
+
"step": 92640
|
| 217233 |
+
},
|
| 217234 |
+
{
|
| 217235 |
+
"epoch": 746.43,
|
| 217236 |
+
"learning_rate": 8.531923076923077e-06,
|
| 217237 |
+
"loss": 0.8169,
|
| 217238 |
+
"step": 92645
|
| 217239 |
+
},
|
| 217240 |
+
{
|
| 217241 |
+
"epoch": 746.47,
|
| 217242 |
+
"learning_rate": 8.53184294871795e-06,
|
| 217243 |
+
"loss": 0.2922,
|
| 217244 |
+
"step": 92650
|
| 217245 |
+
},
|
| 217246 |
+
{
|
| 217247 |
+
"epoch": 746.51,
|
| 217248 |
+
"learning_rate": 8.531762820512821e-06,
|
| 217249 |
+
"loss": 0.3593,
|
| 217250 |
+
"step": 92655
|
| 217251 |
+
},
|
| 217252 |
+
{
|
| 217253 |
+
"epoch": 746.55,
|
| 217254 |
+
"learning_rate": 8.531682692307693e-06,
|
| 217255 |
+
"loss": 0.3671,
|
| 217256 |
+
"step": 92660
|
| 217257 |
+
},
|
| 217258 |
+
{
|
| 217259 |
+
"epoch": 746.59,
|
| 217260 |
+
"learning_rate": 8.531602564102564e-06,
|
| 217261 |
+
"loss": 0.691,
|
| 217262 |
+
"step": 92665
|
| 217263 |
+
},
|
| 217264 |
+
{
|
| 217265 |
+
"epoch": 746.63,
|
| 217266 |
+
"learning_rate": 8.531522435897437e-06,
|
| 217267 |
+
"loss": 0.8604,
|
| 217268 |
+
"step": 92670
|
| 217269 |
+
},
|
| 217270 |
+
{
|
| 217271 |
+
"epoch": 746.67,
|
| 217272 |
+
"learning_rate": 8.531442307692308e-06,
|
| 217273 |
+
"loss": 0.2629,
|
| 217274 |
+
"step": 92675
|
| 217275 |
+
},
|
| 217276 |
+
{
|
| 217277 |
+
"epoch": 746.71,
|
| 217278 |
+
"learning_rate": 8.53136217948718e-06,
|
| 217279 |
+
"loss": 0.2998,
|
| 217280 |
+
"step": 92680
|
| 217281 |
+
},
|
| 217282 |
+
{
|
| 217283 |
+
"epoch": 746.75,
|
| 217284 |
+
"learning_rate": 8.531282051282051e-06,
|
| 217285 |
+
"loss": 0.4048,
|
| 217286 |
+
"step": 92685
|
| 217287 |
+
},
|
| 217288 |
+
{
|
| 217289 |
+
"epoch": 746.79,
|
| 217290 |
+
"learning_rate": 8.531201923076924e-06,
|
| 217291 |
+
"loss": 0.8322,
|
| 217292 |
+
"step": 92690
|
| 217293 |
+
},
|
| 217294 |
+
{
|
| 217295 |
+
"epoch": 746.83,
|
| 217296 |
+
"learning_rate": 8.531121794871796e-06,
|
| 217297 |
+
"loss": 0.8773,
|
| 217298 |
+
"step": 92695
|
| 217299 |
+
},
|
| 217300 |
+
{
|
| 217301 |
+
"epoch": 746.87,
|
| 217302 |
+
"learning_rate": 8.531041666666667e-06,
|
| 217303 |
+
"loss": 0.2898,
|
| 217304 |
+
"step": 92700
|
| 217305 |
+
},
|
| 217306 |
+
{
|
| 217307 |
+
"epoch": 746.91,
|
| 217308 |
+
"learning_rate": 8.53096153846154e-06,
|
| 217309 |
+
"loss": 0.2973,
|
| 217310 |
+
"step": 92705
|
| 217311 |
+
},
|
| 217312 |
+
{
|
| 217313 |
+
"epoch": 746.95,
|
| 217314 |
+
"learning_rate": 8.530881410256411e-06,
|
| 217315 |
+
"loss": 0.4066,
|
| 217316 |
+
"step": 92710
|
| 217317 |
+
},
|
| 217318 |
+
{
|
| 217319 |
+
"epoch": 746.99,
|
| 217320 |
+
"learning_rate": 8.530801282051283e-06,
|
| 217321 |
+
"loss": 0.9941,
|
| 217322 |
+
"step": 92715
|
| 217323 |
+
},
|
| 217324 |
+
{
|
| 217325 |
+
"epoch": 747.0,
|
| 217326 |
+
"eval_loss": 0.49049338698387146,
|
| 217327 |
+
"eval_runtime": 38.7792,
|
| 217328 |
+
"eval_samples_per_second": 21.635,
|
| 217329 |
+
"eval_steps_per_second": 0.696,
|
| 217330 |
+
"eval_wer": 0.19165727170236754,
|
| 217331 |
+
"step": 92716
|
| 217332 |
+
},
|
| 217333 |
+
{
|
| 217334 |
+
"epoch": 753.03,
|
| 217335 |
+
"learning_rate": 8.530721153846154e-06,
|
| 217336 |
+
"loss": 0.3505,
|
| 217337 |
+
"step": 92720
|
| 217338 |
+
},
|
| 217339 |
+
{
|
| 217340 |
+
"epoch": 753.07,
|
| 217341 |
+
"learning_rate": 8.530641025641027e-06,
|
| 217342 |
+
"loss": 0.295,
|
| 217343 |
+
"step": 92725
|
| 217344 |
+
},
|
| 217345 |
+
{
|
| 217346 |
+
"epoch": 753.11,
|
| 217347 |
+
"learning_rate": 8.530560897435898e-06,
|
| 217348 |
+
"loss": 0.3735,
|
| 217349 |
+
"step": 92730
|
| 217350 |
+
},
|
| 217351 |
+
{
|
| 217352 |
+
"epoch": 753.15,
|
| 217353 |
+
"learning_rate": 8.53048076923077e-06,
|
| 217354 |
+
"loss": 0.4566,
|
| 217355 |
+
"step": 92735
|
| 217356 |
+
},
|
| 217357 |
+
{
|
| 217358 |
+
"epoch": 753.19,
|
| 217359 |
+
"learning_rate": 8.530400641025641e-06,
|
| 217360 |
+
"loss": 0.9224,
|
| 217361 |
+
"step": 92740
|
| 217362 |
+
},
|
| 217363 |
+
{
|
| 217364 |
+
"epoch": 753.23,
|
| 217365 |
+
"learning_rate": 8.530320512820514e-06,
|
| 217366 |
+
"loss": 0.6817,
|
| 217367 |
+
"step": 92745
|
| 217368 |
+
},
|
| 217369 |
+
{
|
| 217370 |
+
"epoch": 753.28,
|
| 217371 |
+
"learning_rate": 8.530240384615386e-06,
|
| 217372 |
+
"loss": 0.3159,
|
| 217373 |
+
"step": 92750
|
| 217374 |
+
},
|
| 217375 |
+
{
|
| 217376 |
+
"epoch": 753.32,
|
| 217377 |
+
"learning_rate": 8.530160256410257e-06,
|
| 217378 |
+
"loss": 0.3457,
|
| 217379 |
+
"step": 92755
|
| 217380 |
+
},
|
| 217381 |
+
{
|
| 217382 |
+
"epoch": 753.36,
|
| 217383 |
+
"learning_rate": 8.53008012820513e-06,
|
| 217384 |
+
"loss": 0.4083,
|
| 217385 |
+
"step": 92760
|
| 217386 |
+
},
|
| 217387 |
+
{
|
| 217388 |
+
"epoch": 753.4,
|
| 217389 |
+
"learning_rate": 8.530000000000001e-06,
|
| 217390 |
+
"loss": 1.0235,
|
| 217391 |
+
"step": 92765
|
| 217392 |
+
},
|
| 217393 |
+
{
|
| 217394 |
+
"epoch": 753.44,
|
| 217395 |
+
"learning_rate": 8.529919871794873e-06,
|
| 217396 |
+
"loss": 0.6671,
|
| 217397 |
+
"step": 92770
|
| 217398 |
+
},
|
| 217399 |
+
{
|
| 217400 |
+
"epoch": 753.48,
|
| 217401 |
+
"learning_rate": 8.529839743589744e-06,
|
| 217402 |
+
"loss": 0.3144,
|
| 217403 |
+
"step": 92775
|
| 217404 |
+
},
|
| 217405 |
+
{
|
| 217406 |
+
"epoch": 753.52,
|
| 217407 |
+
"learning_rate": 8.529759615384617e-06,
|
| 217408 |
+
"loss": 0.3815,
|
| 217409 |
+
"step": 92780
|
| 217410 |
+
},
|
| 217411 |
+
{
|
| 217412 |
+
"epoch": 753.56,
|
| 217413 |
+
"learning_rate": 8.529679487179487e-06,
|
| 217414 |
+
"loss": 0.4122,
|
| 217415 |
+
"step": 92785
|
| 217416 |
+
},
|
| 217417 |
+
{
|
| 217418 |
+
"epoch": 753.6,
|
| 217419 |
+
"learning_rate": 8.52959935897436e-06,
|
| 217420 |
+
"loss": 0.9775,
|
| 217421 |
+
"step": 92790
|
| 217422 |
+
},
|
| 217423 |
+
{
|
| 217424 |
+
"epoch": 753.64,
|
| 217425 |
+
"learning_rate": 8.529519230769231e-06,
|
| 217426 |
+
"loss": 0.65,
|
| 217427 |
+
"step": 92795
|
| 217428 |
+
},
|
| 217429 |
+
{
|
| 217430 |
+
"epoch": 753.68,
|
| 217431 |
+
"learning_rate": 8.529439102564103e-06,
|
| 217432 |
+
"loss": 0.2865,
|
| 217433 |
+
"step": 92800
|
| 217434 |
+
},
|
| 217435 |
+
{
|
| 217436 |
+
"epoch": 753.72,
|
| 217437 |
+
"learning_rate": 8.529358974358976e-06,
|
| 217438 |
+
"loss": 0.3356,
|
| 217439 |
+
"step": 92805
|
| 217440 |
+
},
|
| 217441 |
+
{
|
| 217442 |
+
"epoch": 753.76,
|
| 217443 |
+
"learning_rate": 8.529278846153847e-06,
|
| 217444 |
+
"loss": 0.3791,
|
| 217445 |
+
"step": 92810
|
| 217446 |
+
},
|
| 217447 |
+
{
|
| 217448 |
+
"epoch": 753.8,
|
| 217449 |
+
"learning_rate": 8.529198717948718e-06,
|
| 217450 |
+
"loss": 0.9281,
|
| 217451 |
+
"step": 92815
|
| 217452 |
+
},
|
| 217453 |
+
{
|
| 217454 |
+
"epoch": 753.84,
|
| 217455 |
+
"learning_rate": 8.52911858974359e-06,
|
| 217456 |
+
"loss": 0.7974,
|
| 217457 |
+
"step": 92820
|
| 217458 |
+
},
|
| 217459 |
+
{
|
| 217460 |
+
"epoch": 753.88,
|
| 217461 |
+
"learning_rate": 8.529038461538463e-06,
|
| 217462 |
+
"loss": 0.343,
|
| 217463 |
+
"step": 92825
|
| 217464 |
+
},
|
| 217465 |
+
{
|
| 217466 |
+
"epoch": 753.92,
|
| 217467 |
+
"learning_rate": 8.528958333333334e-06,
|
| 217468 |
+
"loss": 0.2876,
|
| 217469 |
+
"step": 92830
|
| 217470 |
+
},
|
| 217471 |
+
{
|
| 217472 |
+
"epoch": 753.96,
|
| 217473 |
+
"learning_rate": 8.528878205128205e-06,
|
| 217474 |
+
"loss": 0.4759,
|
| 217475 |
+
"step": 92835
|
| 217476 |
+
},
|
| 217477 |
+
{
|
| 217478 |
+
"epoch": 754.0,
|
| 217479 |
+
"eval_loss": 0.3695593476295471,
|
| 217480 |
+
"eval_runtime": 39.7408,
|
| 217481 |
+
"eval_samples_per_second": 21.087,
|
| 217482 |
+
"eval_steps_per_second": 0.679,
|
| 217483 |
+
"eval_wer": 0.1869410310477714,
|
| 217484 |
+
"step": 92839
|
| 217485 |
+
},
|
| 217486 |
+
{
|
| 217487 |
+
"epoch": 742.01,
|
| 217488 |
+
"learning_rate": 8.528798076923077e-06,
|
| 217489 |
+
"loss": 0.4816,
|
| 217490 |
+
"step": 92840
|
| 217491 |
+
},
|
| 217492 |
+
{
|
| 217493 |
+
"epoch": 742.05,
|
| 217494 |
+
"learning_rate": 8.52871794871795e-06,
|
| 217495 |
+
"loss": 0.356,
|
| 217496 |
+
"step": 92845
|
| 217497 |
+
},
|
| 217498 |
+
{
|
| 217499 |
+
"epoch": 742.09,
|
| 217500 |
+
"learning_rate": 8.528637820512821e-06,
|
| 217501 |
+
"loss": 0.2479,
|
| 217502 |
+
"step": 92850
|
| 217503 |
+
},
|
| 217504 |
+
{
|
| 217505 |
+
"epoch": 742.13,
|
| 217506 |
+
"learning_rate": 8.528557692307693e-06,
|
| 217507 |
+
"loss": 0.3506,
|
| 217508 |
+
"step": 92855
|
| 217509 |
+
},
|
| 217510 |
+
{
|
| 217511 |
+
"epoch": 742.17,
|
| 217512 |
+
"learning_rate": 8.528477564102566e-06,
|
| 217513 |
+
"loss": 0.5059,
|
| 217514 |
+
"step": 92860
|
| 217515 |
+
},
|
| 217516 |
+
{
|
| 217517 |
+
"epoch": 742.21,
|
| 217518 |
+
"learning_rate": 8.528397435897437e-06,
|
| 217519 |
+
"loss": 1.0797,
|
| 217520 |
+
"step": 92865
|
| 217521 |
+
},
|
| 217522 |
+
{
|
| 217523 |
+
"epoch": 742.25,
|
| 217524 |
+
"learning_rate": 8.528317307692308e-06,
|
| 217525 |
+
"loss": 0.3273,
|
| 217526 |
+
"step": 92870
|
| 217527 |
+
},
|
| 217528 |
+
{
|
| 217529 |
+
"epoch": 742.29,
|
| 217530 |
+
"learning_rate": 8.52823717948718e-06,
|
| 217531 |
+
"loss": 0.3208,
|
| 217532 |
+
"step": 92875
|
| 217533 |
+
},
|
| 217534 |
+
{
|
| 217535 |
+
"epoch": 742.33,
|
| 217536 |
+
"learning_rate": 8.528157051282053e-06,
|
| 217537 |
+
"loss": 0.3685,
|
| 217538 |
+
"step": 92880
|
| 217539 |
+
},
|
| 217540 |
+
{
|
| 217541 |
+
"epoch": 742.37,
|
| 217542 |
+
"learning_rate": 8.528076923076924e-06,
|
| 217543 |
+
"loss": 0.553,
|
| 217544 |
+
"step": 92885
|
| 217545 |
+
},
|
| 217546 |
+
{
|
| 217547 |
+
"epoch": 742.41,
|
| 217548 |
+
"learning_rate": 8.527996794871795e-06,
|
| 217549 |
+
"loss": 1.1638,
|
| 217550 |
+
"step": 92890
|
| 217551 |
+
},
|
| 217552 |
+
{
|
| 217553 |
+
"epoch": 742.45,
|
| 217554 |
+
"learning_rate": 8.527916666666667e-06,
|
| 217555 |
+
"loss": 0.3191,
|
| 217556 |
+
"step": 92895
|
| 217557 |
+
},
|
| 217558 |
+
{
|
| 217559 |
+
"epoch": 742.49,
|
| 217560 |
+
"learning_rate": 8.52783653846154e-06,
|
| 217561 |
+
"loss": 0.3565,
|
| 217562 |
+
"step": 92900
|
| 217563 |
+
},
|
| 217564 |
+
{
|
| 217565 |
+
"epoch": 742.53,
|
| 217566 |
+
"learning_rate": 8.527756410256411e-06,
|
| 217567 |
+
"loss": 0.3219,
|
| 217568 |
+
"step": 92905
|
| 217569 |
+
},
|
| 217570 |
+
{
|
| 217571 |
+
"epoch": 742.57,
|
| 217572 |
+
"learning_rate": 8.527676282051283e-06,
|
| 217573 |
+
"loss": 0.4828,
|
| 217574 |
+
"step": 92910
|
| 217575 |
+
},
|
| 217576 |
+
{
|
| 217577 |
+
"epoch": 742.61,
|
| 217578 |
+
"learning_rate": 8.527596153846156e-06,
|
| 217579 |
+
"loss": 1.0651,
|
| 217580 |
+
"step": 92915
|
| 217581 |
+
},
|
| 217582 |
+
{
|
| 217583 |
+
"epoch": 742.65,
|
| 217584 |
+
"learning_rate": 8.527516025641025e-06,
|
| 217585 |
+
"loss": 0.3005,
|
| 217586 |
+
"step": 92920
|
| 217587 |
+
},
|
| 217588 |
+
{
|
| 217589 |
+
"epoch": 742.69,
|
| 217590 |
+
"learning_rate": 8.527435897435898e-06,
|
| 217591 |
+
"loss": 0.2958,
|
| 217592 |
+
"step": 92925
|
| 217593 |
+
},
|
| 217594 |
+
{
|
| 217595 |
+
"epoch": 742.73,
|
| 217596 |
+
"learning_rate": 8.52735576923077e-06,
|
| 217597 |
+
"loss": 0.3653,
|
| 217598 |
+
"step": 92930
|
| 217599 |
+
},
|
| 217600 |
+
{
|
| 217601 |
+
"epoch": 742.77,
|
| 217602 |
+
"learning_rate": 8.527275641025641e-06,
|
| 217603 |
+
"loss": 0.5593,
|
| 217604 |
+
"step": 92935
|
| 217605 |
+
},
|
| 217606 |
+
{
|
| 217607 |
+
"epoch": 742.81,
|
| 217608 |
+
"learning_rate": 8.527195512820512e-06,
|
| 217609 |
+
"loss": 1.1913,
|
| 217610 |
+
"step": 92940
|
| 217611 |
+
},
|
| 217612 |
+
{
|
| 217613 |
+
"epoch": 742.85,
|
| 217614 |
+
"learning_rate": 8.527115384615385e-06,
|
| 217615 |
+
"loss": 0.346,
|
| 217616 |
+
"step": 92945
|
| 217617 |
+
},
|
| 217618 |
+
{
|
| 217619 |
+
"epoch": 742.89,
|
| 217620 |
+
"learning_rate": 8.527035256410257e-06,
|
| 217621 |
+
"loss": 0.3054,
|
| 217622 |
+
"step": 92950
|
| 217623 |
+
},
|
| 217624 |
+
{
|
| 217625 |
+
"epoch": 742.93,
|
| 217626 |
+
"learning_rate": 8.526955128205128e-06,
|
| 217627 |
+
"loss": 0.3304,
|
| 217628 |
+
"step": 92955
|
| 217629 |
+
},
|
| 217630 |
+
{
|
| 217631 |
+
"epoch": 742.97,
|
| 217632 |
+
"learning_rate": 8.526875000000001e-06,
|
| 217633 |
+
"loss": 0.547,
|
| 217634 |
+
"step": 92960
|
| 217635 |
+
},
|
| 217636 |
+
{
|
| 217637 |
+
"epoch": 743.0,
|
| 217638 |
+
"eval_loss": 0.3991490602493286,
|
| 217639 |
+
"eval_runtime": 37.5272,
|
| 217640 |
+
"eval_samples_per_second": 22.33,
|
| 217641 |
+
"eval_steps_per_second": 0.719,
|
| 217642 |
+
"eval_wer": 0.19152162923376906,
|
| 217643 |
+
"step": 92964
|
| 217644 |
+
},
|
| 217645 |
+
{
|
| 217646 |
+
"epoch": 743.01,
|
| 217647 |
+
"learning_rate": 8.526794871794873e-06,
|
| 217648 |
+
"loss": 0.3283,
|
| 217649 |
+
"step": 92965
|
| 217650 |
+
},
|
| 217651 |
+
{
|
| 217652 |
+
"epoch": 743.05,
|
| 217653 |
+
"learning_rate": 8.526714743589744e-06,
|
| 217654 |
+
"loss": 0.2986,
|
| 217655 |
+
"step": 92970
|
| 217656 |
+
},
|
| 217657 |
+
{
|
| 217658 |
+
"epoch": 743.09,
|
| 217659 |
+
"learning_rate": 8.526634615384615e-06,
|
| 217660 |
+
"loss": 0.3763,
|
| 217661 |
+
"step": 92975
|
| 217662 |
+
},
|
| 217663 |
+
{
|
| 217664 |
+
"epoch": 743.13,
|
| 217665 |
+
"learning_rate": 8.526554487179488e-06,
|
| 217666 |
+
"loss": 0.3062,
|
| 217667 |
+
"step": 92980
|
| 217668 |
+
},
|
| 217669 |
+
{
|
| 217670 |
+
"epoch": 743.17,
|
| 217671 |
+
"learning_rate": 8.52647435897436e-06,
|
| 217672 |
+
"loss": 0.5076,
|
| 217673 |
+
"step": 92985
|
| 217674 |
+
},
|
| 217675 |
+
{
|
| 217676 |
+
"epoch": 743.21,
|
| 217677 |
+
"learning_rate": 8.526394230769231e-06,
|
| 217678 |
+
"loss": 1.1239,
|
| 217679 |
+
"step": 92990
|
| 217680 |
+
},
|
| 217681 |
+
{
|
| 217682 |
+
"epoch": 743.25,
|
| 217683 |
+
"learning_rate": 8.526314102564102e-06,
|
| 217684 |
+
"loss": 0.3523,
|
| 217685 |
+
"step": 92995
|
| 217686 |
+
},
|
| 217687 |
+
{
|
| 217688 |
+
"epoch": 743.29,
|
| 217689 |
+
"learning_rate": 8.526233974358976e-06,
|
| 217690 |
+
"loss": 0.2903,
|
| 217691 |
+
"step": 93000
|
| 217692 |
+
},
|
| 217693 |
+
{
|
| 217694 |
+
"epoch": 743.33,
|
| 217695 |
+
"learning_rate": 8.526153846153847e-06,
|
| 217696 |
+
"loss": 0.367,
|
| 217697 |
+
"step": 93005
|
| 217698 |
+
},
|
| 217699 |
+
{
|
| 217700 |
+
"epoch": 743.37,
|
| 217701 |
+
"learning_rate": 8.526073717948718e-06,
|
| 217702 |
+
"loss": 0.6214,
|
| 217703 |
+
"step": 93010
|
| 217704 |
+
},
|
| 217705 |
+
{
|
| 217706 |
+
"epoch": 743.41,
|
| 217707 |
+
"learning_rate": 8.525993589743591e-06,
|
| 217708 |
+
"loss": 1.1313,
|
| 217709 |
+
"step": 93015
|
| 217710 |
+
},
|
| 217711 |
+
{
|
| 217712 |
+
"epoch": 743.45,
|
| 217713 |
+
"learning_rate": 8.525913461538463e-06,
|
| 217714 |
+
"loss": 0.2883,
|
| 217715 |
+
"step": 93020
|
| 217716 |
+
},
|
| 217717 |
+
{
|
| 217718 |
+
"epoch": 743.49,
|
| 217719 |
+
"learning_rate": 8.525833333333334e-06,
|
| 217720 |
+
"loss": 0.2828,
|
| 217721 |
+
"step": 93025
|
| 217722 |
+
},
|
| 217723 |
+
{
|
| 217724 |
+
"epoch": 743.53,
|
| 217725 |
+
"learning_rate": 8.525753205128205e-06,
|
| 217726 |
+
"loss": 0.3696,
|
| 217727 |
+
"step": 93030
|
| 217728 |
+
},
|
| 217729 |
+
{
|
| 217730 |
+
"epoch": 743.57,
|
| 217731 |
+
"learning_rate": 8.525673076923078e-06,
|
| 217732 |
+
"loss": 0.4927,
|
| 217733 |
+
"step": 93035
|
| 217734 |
+
},
|
| 217735 |
+
{
|
| 217736 |
+
"epoch": 743.61,
|
| 217737 |
+
"learning_rate": 8.52559294871795e-06,
|
| 217738 |
+
"loss": 1.1815,
|
| 217739 |
+
"step": 93040
|
| 217740 |
+
},
|
| 217741 |
+
{
|
| 217742 |
+
"epoch": 743.65,
|
| 217743 |
+
"learning_rate": 8.525512820512821e-06,
|
| 217744 |
+
"loss": 0.3192,
|
| 217745 |
+
"step": 93045
|
| 217746 |
+
},
|
| 217747 |
+
{
|
| 217748 |
+
"epoch": 743.69,
|
| 217749 |
+
"learning_rate": 8.525432692307694e-06,
|
| 217750 |
+
"loss": 0.659,
|
| 217751 |
+
"step": 93050
|
| 217752 |
+
},
|
| 217753 |
+
{
|
| 217754 |
+
"epoch": 743.73,
|
| 217755 |
+
"learning_rate": 8.525352564102566e-06,
|
| 217756 |
+
"loss": 0.3991,
|
| 217757 |
+
"step": 93055
|
| 217758 |
+
},
|
| 217759 |
+
{
|
| 217760 |
+
"epoch": 743.77,
|
| 217761 |
+
"learning_rate": 8.525272435897437e-06,
|
| 217762 |
+
"loss": 0.5503,
|
| 217763 |
+
"step": 93060
|
| 217764 |
+
},
|
| 217765 |
+
{
|
| 217766 |
+
"epoch": 743.81,
|
| 217767 |
+
"learning_rate": 8.525192307692308e-06,
|
| 217768 |
+
"loss": 1.1523,
|
| 217769 |
+
"step": 93065
|
| 217770 |
+
},
|
| 217771 |
+
{
|
| 217772 |
+
"epoch": 743.85,
|
| 217773 |
+
"learning_rate": 8.525112179487181e-06,
|
| 217774 |
+
"loss": 0.2851,
|
| 217775 |
+
"step": 93070
|
| 217776 |
+
},
|
| 217777 |
+
{
|
| 217778 |
+
"epoch": 743.89,
|
| 217779 |
+
"learning_rate": 8.525032051282051e-06,
|
| 217780 |
+
"loss": 0.364,
|
| 217781 |
+
"step": 93075
|
| 217782 |
+
},
|
| 217783 |
+
{
|
| 217784 |
+
"epoch": 743.93,
|
| 217785 |
+
"learning_rate": 8.524951923076924e-06,
|
| 217786 |
+
"loss": 0.31,
|
| 217787 |
+
"step": 93080
|
| 217788 |
+
},
|
| 217789 |
+
{
|
| 217790 |
+
"epoch": 743.97,
|
| 217791 |
+
"learning_rate": 8.524871794871795e-06,
|
| 217792 |
+
"loss": 0.5022,
|
| 217793 |
+
"step": 93085
|
| 217794 |
+
},
|
| 217795 |
+
{
|
| 217796 |
+
"epoch": 744.0,
|
| 217797 |
+
"eval_loss": 0.3836924135684967,
|
| 217798 |
+
"eval_runtime": 38.2104,
|
| 217799 |
+
"eval_samples_per_second": 21.931,
|
| 217800 |
+
"eval_steps_per_second": 0.707,
|
| 217801 |
+
"eval_wer": 0.1841402698607403,
|
| 217802 |
+
"step": 93089
|
| 217803 |
+
},
|
| 217804 |
+
{
|
| 217805 |
+
"epoch": 750.01,
|
| 217806 |
+
"learning_rate": 8.524791666666667e-06,
|
| 217807 |
+
"loss": 0.3855,
|
| 217808 |
+
"step": 93090
|
| 217809 |
+
},
|
| 217810 |
+
{
|
| 217811 |
+
"epoch": 750.05,
|
| 217812 |
+
"learning_rate": 8.524711538461538e-06,
|
| 217813 |
+
"loss": 0.3155,
|
| 217814 |
+
"step": 93095
|
| 217815 |
+
},
|
| 217816 |
+
{
|
| 217817 |
+
"epoch": 750.09,
|
| 217818 |
+
"learning_rate": 8.524631410256411e-06,
|
| 217819 |
+
"loss": 0.2896,
|
| 217820 |
+
"step": 93100
|
| 217821 |
+
},
|
| 217822 |
+
{
|
| 217823 |
+
"epoch": 750.13,
|
| 217824 |
+
"learning_rate": 8.524551282051283e-06,
|
| 217825 |
+
"loss": 0.3521,
|
| 217826 |
+
"step": 93105
|
| 217827 |
+
},
|
| 217828 |
+
{
|
| 217829 |
+
"epoch": 750.17,
|
| 217830 |
+
"learning_rate": 8.524471153846154e-06,
|
| 217831 |
+
"loss": 0.5719,
|
| 217832 |
+
"step": 93110
|
| 217833 |
+
},
|
| 217834 |
+
{
|
| 217835 |
+
"epoch": 750.21,
|
| 217836 |
+
"learning_rate": 8.524391025641027e-06,
|
| 217837 |
+
"loss": 1.1404,
|
| 217838 |
+
"step": 93115
|
| 217839 |
+
},
|
| 217840 |
+
{
|
| 217841 |
+
"epoch": 750.25,
|
| 217842 |
+
"learning_rate": 8.524310897435898e-06,
|
| 217843 |
+
"loss": 0.3249,
|
| 217844 |
+
"step": 93120
|
| 217845 |
+
},
|
| 217846 |
+
{
|
| 217847 |
+
"epoch": 750.29,
|
| 217848 |
+
"learning_rate": 8.52423076923077e-06,
|
| 217849 |
+
"loss": 0.2555,
|
| 217850 |
+
"step": 93125
|
| 217851 |
+
},
|
| 217852 |
+
{
|
| 217853 |
+
"epoch": 750.33,
|
| 217854 |
+
"learning_rate": 8.524150641025641e-06,
|
| 217855 |
+
"loss": 0.3787,
|
| 217856 |
+
"step": 93130
|
| 217857 |
+
},
|
| 217858 |
+
{
|
| 217859 |
+
"epoch": 750.37,
|
| 217860 |
+
"learning_rate": 8.524070512820514e-06,
|
| 217861 |
+
"loss": 0.5679,
|
| 217862 |
+
"step": 93135
|
| 217863 |
+
},
|
| 217864 |
+
{
|
| 217865 |
+
"epoch": 750.41,
|
| 217866 |
+
"learning_rate": 8.523990384615385e-06,
|
| 217867 |
+
"loss": 1.1,
|
| 217868 |
+
"step": 93140
|
| 217869 |
+
},
|
| 217870 |
+
{
|
| 217871 |
+
"epoch": 750.45,
|
| 217872 |
+
"learning_rate": 8.523910256410257e-06,
|
| 217873 |
+
"loss": 0.3734,
|
| 217874 |
+
"step": 93145
|
| 217875 |
+
},
|
| 217876 |
+
{
|
| 217877 |
+
"epoch": 750.49,
|
| 217878 |
+
"learning_rate": 8.52383012820513e-06,
|
| 217879 |
+
"loss": 0.278,
|
| 217880 |
+
"step": 93150
|
| 217881 |
+
},
|
| 217882 |
+
{
|
| 217883 |
+
"epoch": 750.53,
|
| 217884 |
+
"learning_rate": 8.523750000000001e-06,
|
| 217885 |
+
"loss": 0.4092,
|
| 217886 |
+
"step": 93155
|
| 217887 |
+
},
|
| 217888 |
+
{
|
| 217889 |
+
"epoch": 750.57,
|
| 217890 |
+
"learning_rate": 8.523669871794873e-06,
|
| 217891 |
+
"loss": 0.6068,
|
| 217892 |
+
"step": 93160
|
| 217893 |
+
},
|
| 217894 |
+
{
|
| 217895 |
+
"epoch": 750.61,
|
| 217896 |
+
"learning_rate": 8.523589743589744e-06,
|
| 217897 |
+
"loss": 1.1341,
|
| 217898 |
+
"step": 93165
|
| 217899 |
+
},
|
| 217900 |
+
{
|
| 217901 |
+
"epoch": 750.65,
|
| 217902 |
+
"learning_rate": 8.523509615384617e-06,
|
| 217903 |
+
"loss": 0.3188,
|
| 217904 |
+
"step": 93170
|
| 217905 |
+
},
|
| 217906 |
+
{
|
| 217907 |
+
"epoch": 750.69,
|
| 217908 |
+
"learning_rate": 8.523429487179488e-06,
|
| 217909 |
+
"loss": 0.2805,
|
| 217910 |
+
"step": 93175
|
| 217911 |
+
},
|
| 217912 |
+
{
|
| 217913 |
+
"epoch": 750.73,
|
| 217914 |
+
"learning_rate": 8.52334935897436e-06,
|
| 217915 |
+
"loss": 0.3427,
|
| 217916 |
+
"step": 93180
|
| 217917 |
+
},
|
| 217918 |
+
{
|
| 217919 |
+
"epoch": 750.77,
|
| 217920 |
+
"learning_rate": 8.523269230769231e-06,
|
| 217921 |
+
"loss": 0.5412,
|
| 217922 |
+
"step": 93185
|
| 217923 |
+
},
|
| 217924 |
+
{
|
| 217925 |
+
"epoch": 750.81,
|
| 217926 |
+
"learning_rate": 8.523189102564104e-06,
|
| 217927 |
+
"loss": 1.1955,
|
| 217928 |
+
"step": 93190
|
| 217929 |
+
},
|
| 217930 |
+
{
|
| 217931 |
+
"epoch": 750.85,
|
| 217932 |
+
"learning_rate": 8.523108974358974e-06,
|
| 217933 |
+
"loss": 0.2784,
|
| 217934 |
+
"step": 93195
|
| 217935 |
+
},
|
| 217936 |
+
{
|
| 217937 |
+
"epoch": 750.89,
|
| 217938 |
+
"learning_rate": 8.523028846153847e-06,
|
| 217939 |
+
"loss": 0.295,
|
| 217940 |
+
"step": 93200
|
| 217941 |
+
},
|
| 217942 |
+
{
|
| 217943 |
+
"epoch": 750.93,
|
| 217944 |
+
"learning_rate": 8.52294871794872e-06,
|
| 217945 |
+
"loss": 0.333,
|
| 217946 |
+
"step": 93205
|
| 217947 |
+
},
|
| 217948 |
+
{
|
| 217949 |
+
"epoch": 750.97,
|
| 217950 |
+
"learning_rate": 8.52286858974359e-06,
|
| 217951 |
+
"loss": 0.5604,
|
| 217952 |
+
"step": 93210
|
| 217953 |
+
},
|
| 217954 |
+
{
|
| 217955 |
+
"epoch": 751.0,
|
| 217956 |
+
"eval_loss": 0.4374641180038452,
|
| 217957 |
+
"eval_runtime": 38.7976,
|
| 217958 |
+
"eval_samples_per_second": 21.599,
|
| 217959 |
+
"eval_steps_per_second": 0.696,
|
| 217960 |
+
"eval_wer": 0.19383097762073026,
|
| 217961 |
+
"step": 93213
|
| 217962 |
+
},
|
| 217963 |
+
{
|
| 217964 |
+
"epoch": 751.02,
|
| 217965 |
+
"learning_rate": 8.522788461538463e-06,
|
| 217966 |
+
"loss": 0.4429,
|
| 217967 |
+
"step": 93215
|
| 217968 |
+
},
|
| 217969 |
+
{
|
| 217970 |
+
"epoch": 751.06,
|
| 217971 |
+
"learning_rate": 8.522708333333334e-06,
|
| 217972 |
+
"loss": 0.277,
|
| 217973 |
+
"step": 93220
|
| 217974 |
+
},
|
| 217975 |
+
{
|
| 217976 |
+
"epoch": 751.1,
|
| 217977 |
+
"learning_rate": 8.522628205128205e-06,
|
| 217978 |
+
"loss": 0.2919,
|
| 217979 |
+
"step": 93225
|
| 217980 |
+
},
|
| 217981 |
+
{
|
| 217982 |
+
"epoch": 751.14,
|
| 217983 |
+
"learning_rate": 8.522548076923077e-06,
|
| 217984 |
+
"loss": 0.4176,
|
| 217985 |
+
"step": 93230
|
| 217986 |
+
},
|
| 217987 |
+
{
|
| 217988 |
+
"epoch": 751.18,
|
| 217989 |
+
"learning_rate": 8.52246794871795e-06,
|
| 217990 |
+
"loss": 0.6155,
|
| 217991 |
+
"step": 93235
|
| 217992 |
+
},
|
| 217993 |
+
{
|
| 217994 |
+
"epoch": 751.22,
|
| 217995 |
+
"learning_rate": 8.522387820512821e-06,
|
| 217996 |
+
"loss": 0.9501,
|
| 217997 |
+
"step": 93240
|
| 217998 |
+
},
|
| 217999 |
+
{
|
| 218000 |
+
"epoch": 751.26,
|
| 218001 |
+
"learning_rate": 8.522307692307692e-06,
|
| 218002 |
+
"loss": 0.311,
|
| 218003 |
+
"step": 93245
|
| 218004 |
+
},
|
| 218005 |
+
{
|
| 218006 |
+
"epoch": 751.3,
|
| 218007 |
+
"learning_rate": 8.522227564102565e-06,
|
| 218008 |
+
"loss": 0.2841,
|
| 218009 |
+
"step": 93250
|
| 218010 |
+
},
|
| 218011 |
+
{
|
| 218012 |
+
"epoch": 751.34,
|
| 218013 |
+
"learning_rate": 8.522147435897437e-06,
|
| 218014 |
+
"loss": 0.356,
|
| 218015 |
+
"step": 93255
|
| 218016 |
+
},
|
| 218017 |
+
{
|
| 218018 |
+
"epoch": 751.38,
|
| 218019 |
+
"learning_rate": 8.522067307692308e-06,
|
| 218020 |
+
"loss": 0.6652,
|
| 218021 |
+
"step": 93260
|
| 218022 |
+
},
|
| 218023 |
+
{
|
| 218024 |
+
"epoch": 751.42,
|
| 218025 |
+
"learning_rate": 8.52198717948718e-06,
|
| 218026 |
+
"loss": 1.0212,
|
| 218027 |
+
"step": 93265
|
| 218028 |
+
},
|
| 218029 |
+
{
|
| 218030 |
+
"epoch": 751.46,
|
| 218031 |
+
"learning_rate": 8.521907051282053e-06,
|
| 218032 |
+
"loss": 0.2716,
|
| 218033 |
+
"step": 93270
|
| 218034 |
+
},
|
| 218035 |
+
{
|
| 218036 |
+
"epoch": 751.5,
|
| 218037 |
+
"learning_rate": 8.521826923076924e-06,
|
| 218038 |
+
"loss": 0.2725,
|
| 218039 |
+
"step": 93275
|
| 218040 |
+
},
|
| 218041 |
+
{
|
| 218042 |
+
"epoch": 751.54,
|
| 218043 |
+
"learning_rate": 8.521746794871795e-06,
|
| 218044 |
+
"loss": 0.3049,
|
| 218045 |
+
"step": 93280
|
| 218046 |
+
},
|
| 218047 |
+
{
|
| 218048 |
+
"epoch": 751.58,
|
| 218049 |
+
"learning_rate": 8.521666666666667e-06,
|
| 218050 |
+
"loss": 0.6016,
|
| 218051 |
+
"step": 93285
|
| 218052 |
+
},
|
| 218053 |
+
{
|
| 218054 |
+
"epoch": 751.62,
|
| 218055 |
+
"learning_rate": 8.52158653846154e-06,
|
| 218056 |
+
"loss": 0.9861,
|
| 218057 |
+
"step": 93290
|
| 218058 |
+
},
|
| 218059 |
+
{
|
| 218060 |
+
"epoch": 751.66,
|
| 218061 |
+
"learning_rate": 8.521506410256411e-06,
|
| 218062 |
+
"loss": 0.3237,
|
| 218063 |
+
"step": 93295
|
| 218064 |
+
},
|
| 218065 |
+
{
|
| 218066 |
+
"epoch": 751.7,
|
| 218067 |
+
"learning_rate": 8.521426282051282e-06,
|
| 218068 |
+
"loss": 0.3375,
|
| 218069 |
+
"step": 93300
|
| 218070 |
+
},
|
| 218071 |
+
{
|
| 218072 |
+
"epoch": 751.74,
|
| 218073 |
+
"learning_rate": 8.521346153846155e-06,
|
| 218074 |
+
"loss": 0.3626,
|
| 218075 |
+
"step": 93305
|
| 218076 |
+
},
|
| 218077 |
+
{
|
| 218078 |
+
"epoch": 751.78,
|
| 218079 |
+
"learning_rate": 8.521266025641027e-06,
|
| 218080 |
+
"loss": 0.5895,
|
| 218081 |
+
"step": 93310
|
| 218082 |
+
},
|
| 218083 |
+
{
|
| 218084 |
+
"epoch": 751.82,
|
| 218085 |
+
"learning_rate": 8.521185897435898e-06,
|
| 218086 |
+
"loss": 1.1215,
|
| 218087 |
+
"step": 93315
|
| 218088 |
+
},
|
| 218089 |
+
{
|
| 218090 |
+
"epoch": 751.86,
|
| 218091 |
+
"learning_rate": 8.52110576923077e-06,
|
| 218092 |
+
"loss": 0.3631,
|
| 218093 |
+
"step": 93320
|
| 218094 |
+
},
|
| 218095 |
+
{
|
| 218096 |
+
"epoch": 751.9,
|
| 218097 |
+
"learning_rate": 8.521025641025643e-06,
|
| 218098 |
+
"loss": 0.3603,
|
| 218099 |
+
"step": 93325
|
| 218100 |
+
},
|
| 218101 |
+
{
|
| 218102 |
+
"epoch": 751.94,
|
| 218103 |
+
"learning_rate": 8.520945512820514e-06,
|
| 218104 |
+
"loss": 0.3875,
|
| 218105 |
+
"step": 93330
|
| 218106 |
+
},
|
| 218107 |
+
{
|
| 218108 |
+
"epoch": 751.98,
|
| 218109 |
+
"learning_rate": 8.520865384615385e-06,
|
| 218110 |
+
"loss": 0.6634,
|
| 218111 |
+
"step": 93335
|
| 218112 |
+
},
|
| 218113 |
+
{
|
| 218114 |
+
"epoch": 752.0,
|
| 218115 |
+
"eval_loss": 0.3540174067020416,
|
| 218116 |
+
"eval_runtime": 38.9251,
|
| 218117 |
+
"eval_samples_per_second": 21.529,
|
| 218118 |
+
"eval_steps_per_second": 0.694,
|
| 218119 |
+
"eval_wer": 0.1765940508069769,
|
| 218120 |
+
"step": 93337
|
| 218121 |
+
},
|
| 218122 |
+
{
|
| 218123 |
+
"epoch": 746.02,
|
| 218124 |
+
"learning_rate": 8.520785256410257e-06,
|
| 218125 |
+
"loss": 0.2982,
|
| 218126 |
+
"step": 93340
|
| 218127 |
+
},
|
| 218128 |
+
{
|
| 218129 |
+
"epoch": 746.06,
|
| 218130 |
+
"learning_rate": 8.52070512820513e-06,
|
| 218131 |
+
"loss": 0.2931,
|
| 218132 |
+
"step": 93345
|
| 218133 |
+
},
|
| 218134 |
+
{
|
| 218135 |
+
"epoch": 746.1,
|
| 218136 |
+
"learning_rate": 8.520625e-06,
|
| 218137 |
+
"loss": 0.2853,
|
| 218138 |
+
"step": 93350
|
| 218139 |
+
},
|
| 218140 |
+
{
|
| 218141 |
+
"epoch": 746.14,
|
| 218142 |
+
"learning_rate": 8.520544871794872e-06,
|
| 218143 |
+
"loss": 0.3717,
|
| 218144 |
+
"step": 93355
|
| 218145 |
+
},
|
| 218146 |
+
{
|
| 218147 |
+
"epoch": 746.18,
|
| 218148 |
+
"learning_rate": 8.520464743589746e-06,
|
| 218149 |
+
"loss": 0.8425,
|
| 218150 |
+
"step": 93360
|
| 218151 |
+
},
|
| 218152 |
+
{
|
| 218153 |
+
"epoch": 746.22,
|
| 218154 |
+
"learning_rate": 8.520384615384615e-06,
|
| 218155 |
+
"loss": 0.8917,
|
| 218156 |
+
"step": 93365
|
| 218157 |
+
},
|
| 218158 |
+
{
|
| 218159 |
+
"epoch": 746.26,
|
| 218160 |
+
"learning_rate": 8.520304487179488e-06,
|
| 218161 |
+
"loss": 0.2906,
|
| 218162 |
+
"step": 93370
|
| 218163 |
+
},
|
| 218164 |
+
{
|
| 218165 |
+
"epoch": 746.3,
|
| 218166 |
+
"learning_rate": 8.52022435897436e-06,
|
| 218167 |
+
"loss": 0.3077,
|
| 218168 |
+
"step": 93375
|
| 218169 |
+
},
|
| 218170 |
+
{
|
| 218171 |
+
"epoch": 746.34,
|
| 218172 |
+
"learning_rate": 8.520144230769231e-06,
|
| 218173 |
+
"loss": 0.3911,
|
| 218174 |
+
"step": 93380
|
| 218175 |
+
},
|
| 218176 |
+
{
|
| 218177 |
+
"epoch": 746.38,
|
| 218178 |
+
"learning_rate": 8.520064102564102e-06,
|
| 218179 |
+
"loss": 0.7335,
|
| 218180 |
+
"step": 93385
|
| 218181 |
+
},
|
| 218182 |
+
{
|
| 218183 |
+
"epoch": 746.42,
|
| 218184 |
+
"learning_rate": 8.519983974358975e-06,
|
| 218185 |
+
"loss": 0.8044,
|
| 218186 |
+
"step": 93390
|
| 218187 |
+
},
|
| 218188 |
+
{
|
| 218189 |
+
"epoch": 746.46,
|
| 218190 |
+
"learning_rate": 8.519903846153847e-06,
|
| 218191 |
+
"loss": 0.3038,
|
| 218192 |
+
"step": 93395
|
| 218193 |
+
},
|
| 218194 |
+
{
|
| 218195 |
+
"epoch": 746.5,
|
| 218196 |
+
"learning_rate": 8.519823717948718e-06,
|
| 218197 |
+
"loss": 0.2585,
|
| 218198 |
+
"step": 93400
|
| 218199 |
+
},
|
| 218200 |
+
{
|
| 218201 |
+
"epoch": 746.54,
|
| 218202 |
+
"learning_rate": 8.519743589743591e-06,
|
| 218203 |
+
"loss": 0.4126,
|
| 218204 |
+
"step": 93405
|
| 218205 |
+
},
|
| 218206 |
+
{
|
| 218207 |
+
"epoch": 746.58,
|
| 218208 |
+
"learning_rate": 8.519663461538462e-06,
|
| 218209 |
+
"loss": 0.7608,
|
| 218210 |
+
"step": 93410
|
| 218211 |
+
},
|
| 218212 |
+
{
|
| 218213 |
+
"epoch": 746.62,
|
| 218214 |
+
"learning_rate": 8.519583333333334e-06,
|
| 218215 |
+
"loss": 1.017,
|
| 218216 |
+
"step": 93415
|
| 218217 |
+
},
|
| 218218 |
+
{
|
| 218219 |
+
"epoch": 746.66,
|
| 218220 |
+
"learning_rate": 8.519503205128205e-06,
|
| 218221 |
+
"loss": 0.3271,
|
| 218222 |
+
"step": 93420
|
| 218223 |
+
},
|
| 218224 |
+
{
|
| 218225 |
+
"epoch": 746.7,
|
| 218226 |
+
"learning_rate": 8.519423076923078e-06,
|
| 218227 |
+
"loss": 0.4955,
|
| 218228 |
+
"step": 93425
|
| 218229 |
+
},
|
| 218230 |
+
{
|
| 218231 |
+
"epoch": 746.74,
|
| 218232 |
+
"learning_rate": 8.51934294871795e-06,
|
| 218233 |
+
"loss": 0.4092,
|
| 218234 |
+
"step": 93430
|
| 218235 |
+
},
|
| 218236 |
+
{
|
| 218237 |
+
"epoch": 746.78,
|
| 218238 |
+
"learning_rate": 8.519262820512821e-06,
|
| 218239 |
+
"loss": 0.723,
|
| 218240 |
+
"step": 93435
|
| 218241 |
+
},
|
| 218242 |
+
{
|
| 218243 |
+
"epoch": 746.82,
|
| 218244 |
+
"learning_rate": 8.519182692307692e-06,
|
| 218245 |
+
"loss": 0.8715,
|
| 218246 |
+
"step": 93440
|
| 218247 |
+
},
|
| 218248 |
+
{
|
| 218249 |
+
"epoch": 746.86,
|
| 218250 |
+
"learning_rate": 8.519102564102565e-06,
|
| 218251 |
+
"loss": 0.2983,
|
| 218252 |
+
"step": 93445
|
| 218253 |
+
},
|
| 218254 |
+
{
|
| 218255 |
+
"epoch": 746.9,
|
| 218256 |
+
"learning_rate": 8.519022435897437e-06,
|
| 218257 |
+
"loss": 0.2913,
|
| 218258 |
+
"step": 93450
|
| 218259 |
+
},
|
| 218260 |
+
{
|
| 218261 |
+
"epoch": 746.94,
|
| 218262 |
+
"learning_rate": 8.518942307692308e-06,
|
| 218263 |
+
"loss": 0.384,
|
| 218264 |
+
"step": 93455
|
| 218265 |
+
},
|
| 218266 |
+
{
|
| 218267 |
+
"epoch": 746.98,
|
| 218268 |
+
"learning_rate": 8.518862179487181e-06,
|
| 218269 |
+
"loss": 0.7787,
|
| 218270 |
+
"step": 93460
|
| 218271 |
+
},
|
| 218272 |
+
{
|
| 218273 |
+
"epoch": 747.0,
|
| 218274 |
+
"eval_loss": 0.4655894935131073,
|
| 218275 |
+
"eval_runtime": 39.6595,
|
| 218276 |
+
"eval_samples_per_second": 21.13,
|
| 218277 |
+
"eval_steps_per_second": 0.681,
|
| 218278 |
+
"eval_wer": 0.19279159265837773,
|
| 218279 |
+
"step": 93462
|
| 218280 |
+
},
|
| 218281 |
+
{
|
| 218282 |
+
"epoch": 753.02,
|
| 218283 |
+
"learning_rate": 8.518782051282053e-06,
|
| 218284 |
+
"loss": 0.3159,
|
| 218285 |
+
"step": 93465
|
| 218286 |
+
},
|
| 218287 |
+
{
|
| 218288 |
+
"epoch": 753.06,
|
| 218289 |
+
"learning_rate": 8.518701923076924e-06,
|
| 218290 |
+
"loss": 0.2954,
|
| 218291 |
+
"step": 93470
|
| 218292 |
+
},
|
| 218293 |
+
{
|
| 218294 |
+
"epoch": 753.1,
|
| 218295 |
+
"learning_rate": 8.518621794871795e-06,
|
| 218296 |
+
"loss": 0.32,
|
| 218297 |
+
"step": 93475
|
| 218298 |
+
},
|
| 218299 |
+
{
|
| 218300 |
+
"epoch": 753.14,
|
| 218301 |
+
"learning_rate": 8.518541666666668e-06,
|
| 218302 |
+
"loss": 0.4195,
|
| 218303 |
+
"step": 93480
|
| 218304 |
+
},
|
| 218305 |
+
{
|
| 218306 |
+
"epoch": 753.18,
|
| 218307 |
+
"learning_rate": 8.518461538461538e-06,
|
| 218308 |
+
"loss": 0.709,
|
| 218309 |
+
"step": 93485
|
| 218310 |
+
},
|
| 218311 |
+
{
|
| 218312 |
+
"epoch": 753.22,
|
| 218313 |
+
"learning_rate": 8.518381410256411e-06,
|
| 218314 |
+
"loss": 0.8771,
|
| 218315 |
+
"step": 93490
|
| 218316 |
+
},
|
| 218317 |
+
{
|
| 218318 |
+
"epoch": 753.27,
|
| 218319 |
+
"learning_rate": 8.518301282051282e-06,
|
| 218320 |
+
"loss": 0.2633,
|
| 218321 |
+
"step": 93495
|
| 218322 |
+
},
|
| 218323 |
+
{
|
| 218324 |
+
"epoch": 753.31,
|
| 218325 |
+
"learning_rate": 8.518221153846154e-06,
|
| 218326 |
+
"loss": 0.2711,
|
| 218327 |
+
"step": 93500
|
| 218328 |
+
},
|
| 218329 |
+
{
|
| 218330 |
+
"epoch": 753.35,
|
| 218331 |
+
"learning_rate": 8.518141025641027e-06,
|
| 218332 |
+
"loss": 0.4551,
|
| 218333 |
+
"step": 93505
|
| 218334 |
+
},
|
| 218335 |
+
{
|
| 218336 |
+
"epoch": 753.39,
|
| 218337 |
+
"learning_rate": 8.518060897435898e-06,
|
| 218338 |
+
"loss": 0.7144,
|
| 218339 |
+
"step": 93510
|
| 218340 |
+
},
|
| 218341 |
+
{
|
| 218342 |
+
"epoch": 753.43,
|
| 218343 |
+
"learning_rate": 8.51798076923077e-06,
|
| 218344 |
+
"loss": 0.8337,
|
| 218345 |
+
"step": 93515
|
| 218346 |
+
},
|
| 218347 |
+
{
|
| 218348 |
+
"epoch": 753.47,
|
| 218349 |
+
"learning_rate": 8.517900641025641e-06,
|
| 218350 |
+
"loss": 0.2779,
|
| 218351 |
+
"step": 93520
|
| 218352 |
+
},
|
| 218353 |
+
{
|
| 218354 |
+
"epoch": 753.51,
|
| 218355 |
+
"learning_rate": 8.517820512820514e-06,
|
| 218356 |
+
"loss": 0.4137,
|
| 218357 |
+
"step": 93525
|
| 218358 |
+
},
|
| 218359 |
+
{
|
| 218360 |
+
"epoch": 753.55,
|
| 218361 |
+
"learning_rate": 8.517740384615385e-06,
|
| 218362 |
+
"loss": 0.3697,
|
| 218363 |
+
"step": 93530
|
| 218364 |
+
},
|
| 218365 |
+
{
|
| 218366 |
+
"epoch": 753.59,
|
| 218367 |
+
"learning_rate": 8.517660256410257e-06,
|
| 218368 |
+
"loss": 0.7943,
|
| 218369 |
+
"step": 93535
|
| 218370 |
+
},
|
| 218371 |
+
{
|
| 218372 |
+
"epoch": 753.63,
|
| 218373 |
+
"learning_rate": 8.517580128205128e-06,
|
| 218374 |
+
"loss": 0.8019,
|
| 218375 |
+
"step": 93540
|
| 218376 |
+
},
|
| 218377 |
+
{
|
| 218378 |
+
"epoch": 753.67,
|
| 218379 |
+
"learning_rate": 8.517500000000001e-06,
|
| 218380 |
+
"loss": 0.3453,
|
| 218381 |
+
"step": 93545
|
| 218382 |
+
},
|
| 218383 |
+
{
|
| 218384 |
+
"epoch": 753.71,
|
| 218385 |
+
"learning_rate": 8.517419871794872e-06,
|
| 218386 |
+
"loss": 0.3126,
|
| 218387 |
+
"step": 93550
|
| 218388 |
+
},
|
| 218389 |
+
{
|
| 218390 |
+
"epoch": 753.75,
|
| 218391 |
+
"learning_rate": 8.517339743589744e-06,
|
| 218392 |
+
"loss": 0.3679,
|
| 218393 |
+
"step": 93555
|
| 218394 |
+
},
|
| 218395 |
+
{
|
| 218396 |
+
"epoch": 753.79,
|
| 218397 |
+
"learning_rate": 8.517259615384617e-06,
|
| 218398 |
+
"loss": 0.7963,
|
| 218399 |
+
"step": 93560
|
| 218400 |
+
},
|
| 218401 |
+
{
|
| 218402 |
+
"epoch": 753.83,
|
| 218403 |
+
"learning_rate": 8.517179487179488e-06,
|
| 218404 |
+
"loss": 0.8713,
|
| 218405 |
+
"step": 93565
|
| 218406 |
+
},
|
| 218407 |
+
{
|
| 218408 |
+
"epoch": 753.87,
|
| 218409 |
+
"learning_rate": 8.51709935897436e-06,
|
| 218410 |
+
"loss": 0.2662,
|
| 218411 |
+
"step": 93570
|
| 218412 |
+
},
|
| 218413 |
+
{
|
| 218414 |
+
"epoch": 753.91,
|
| 218415 |
+
"learning_rate": 8.517019230769231e-06,
|
| 218416 |
+
"loss": 0.306,
|
| 218417 |
+
"step": 93575
|
| 218418 |
+
},
|
| 218419 |
+
{
|
| 218420 |
+
"epoch": 753.95,
|
| 218421 |
+
"learning_rate": 8.516939102564104e-06,
|
| 218422 |
+
"loss": 0.3898,
|
| 218423 |
+
"step": 93580
|
| 218424 |
+
},
|
| 218425 |
+
{
|
| 218426 |
+
"epoch": 753.99,
|
| 218427 |
+
"learning_rate": 8.516858974358975e-06,
|
| 218428 |
+
"loss": 0.7886,
|
| 218429 |
+
"step": 93585
|
| 218430 |
+
},
|
| 218431 |
+
{
|
| 218432 |
+
"epoch": 754.0,
|
| 218433 |
+
"eval_loss": 0.33248934149742126,
|
| 218434 |
+
"eval_runtime": 40.6199,
|
| 218435 |
+
"eval_samples_per_second": 20.655,
|
| 218436 |
+
"eval_steps_per_second": 0.665,
|
| 218437 |
+
"eval_wer": 0.19042131796903133,
|
| 218438 |
+
"step": 93586
|
| 218439 |
+
},
|
| 218440 |
+
{
|
| 218441 |
+
"epoch": 748.03,
|
| 218442 |
+
"learning_rate": 8.516778846153847e-06,
|
| 218443 |
+
"loss": 0.3288,
|
| 218444 |
+
"step": 93590
|
| 218445 |
+
},
|
| 218446 |
+
{
|
| 218447 |
+
"epoch": 748.07,
|
| 218448 |
+
"learning_rate": 8.516698717948718e-06,
|
| 218449 |
+
"loss": 0.2822,
|
| 218450 |
+
"step": 93595
|
| 218451 |
+
},
|
| 218452 |
+
{
|
| 218453 |
+
"epoch": 748.11,
|
| 218454 |
+
"learning_rate": 8.516618589743591e-06,
|
| 218455 |
+
"loss": 0.3252,
|
| 218456 |
+
"step": 93600
|
| 218457 |
+
},
|
| 218458 |
+
{
|
| 218459 |
+
"epoch": 748.15,
|
| 218460 |
+
"learning_rate": 8.516538461538462e-06,
|
| 218461 |
+
"loss": 0.3961,
|
| 218462 |
+
"step": 93605
|
| 218463 |
+
},
|
| 218464 |
+
{
|
| 218465 |
+
"epoch": 748.19,
|
| 218466 |
+
"learning_rate": 8.516458333333334e-06,
|
| 218467 |
+
"loss": 0.9088,
|
| 218468 |
+
"step": 93610
|
| 218469 |
+
},
|
| 218470 |
+
{
|
| 218471 |
+
"epoch": 748.23,
|
| 218472 |
+
"learning_rate": 8.516378205128207e-06,
|
| 218473 |
+
"loss": 0.7068,
|
| 218474 |
+
"step": 93615
|
| 218475 |
+
},
|
| 218476 |
+
{
|
| 218477 |
+
"epoch": 748.27,
|
| 218478 |
+
"learning_rate": 8.516298076923078e-06,
|
| 218479 |
+
"loss": 0.2668,
|
| 218480 |
+
"step": 93620
|
| 218481 |
+
},
|
| 218482 |
+
{
|
| 218483 |
+
"epoch": 748.31,
|
| 218484 |
+
"learning_rate": 8.51621794871795e-06,
|
| 218485 |
+
"loss": 0.298,
|
| 218486 |
+
"step": 93625
|
| 218487 |
+
},
|
| 218488 |
+
{
|
| 218489 |
+
"epoch": 748.35,
|
| 218490 |
+
"learning_rate": 8.516137820512821e-06,
|
| 218491 |
+
"loss": 0.415,
|
| 218492 |
+
"step": 93630
|
| 218493 |
+
},
|
| 218494 |
+
{
|
| 218495 |
+
"epoch": 748.39,
|
| 218496 |
+
"learning_rate": 8.516057692307694e-06,
|
| 218497 |
+
"loss": 0.9226,
|
| 218498 |
+
"step": 93635
|
| 218499 |
+
},
|
| 218500 |
+
{
|
| 218501 |
+
"epoch": 748.43,
|
| 218502 |
+
"learning_rate": 8.515977564102564e-06,
|
| 218503 |
+
"loss": 0.69,
|
| 218504 |
+
"step": 93640
|
| 218505 |
+
},
|
| 218506 |
+
{
|
| 218507 |
+
"epoch": 748.47,
|
| 218508 |
+
"learning_rate": 8.515897435897437e-06,
|
| 218509 |
+
"loss": 0.2476,
|
| 218510 |
+
"step": 93645
|
| 218511 |
+
},
|
| 218512 |
+
{
|
| 218513 |
+
"epoch": 748.51,
|
| 218514 |
+
"learning_rate": 8.515817307692308e-06,
|
| 218515 |
+
"loss": 0.325,
|
| 218516 |
+
"step": 93650
|
| 218517 |
+
},
|
| 218518 |
+
{
|
| 218519 |
+
"epoch": 748.55,
|
| 218520 |
+
"learning_rate": 8.51573717948718e-06,
|
| 218521 |
+
"loss": 0.4583,
|
| 218522 |
+
"step": 93655
|
| 218523 |
+
},
|
| 218524 |
+
{
|
| 218525 |
+
"epoch": 748.59,
|
| 218526 |
+
"learning_rate": 8.515657051282052e-06,
|
| 218527 |
+
"loss": 0.8614,
|
| 218528 |
+
"step": 93660
|
| 218529 |
+
},
|
| 218530 |
+
{
|
| 218531 |
+
"epoch": 748.63,
|
| 218532 |
+
"learning_rate": 8.515576923076924e-06,
|
| 218533 |
+
"loss": 0.6535,
|
| 218534 |
+
"step": 93665
|
| 218535 |
+
},
|
| 218536 |
+
{
|
| 218537 |
+
"epoch": 748.67,
|
| 218538 |
+
"learning_rate": 8.515496794871795e-06,
|
| 218539 |
+
"loss": 0.3172,
|
| 218540 |
+
"step": 93670
|
| 218541 |
+
},
|
| 218542 |
+
{
|
| 218543 |
+
"epoch": 748.71,
|
| 218544 |
+
"learning_rate": 8.515416666666667e-06,
|
| 218545 |
+
"loss": 0.314,
|
| 218546 |
+
"step": 93675
|
| 218547 |
+
},
|
| 218548 |
+
{
|
| 218549 |
+
"epoch": 748.75,
|
| 218550 |
+
"learning_rate": 8.51533653846154e-06,
|
| 218551 |
+
"loss": 0.4656,
|
| 218552 |
+
"step": 93680
|
| 218553 |
+
},
|
| 218554 |
+
{
|
| 218555 |
+
"epoch": 748.79,
|
| 218556 |
+
"learning_rate": 8.515256410256411e-06,
|
| 218557 |
+
"loss": 0.9987,
|
| 218558 |
+
"step": 93685
|
| 218559 |
+
},
|
| 218560 |
+
{
|
| 218561 |
+
"epoch": 748.83,
|
| 218562 |
+
"learning_rate": 8.515176282051282e-06,
|
| 218563 |
+
"loss": 0.6576,
|
| 218564 |
+
"step": 93690
|
| 218565 |
+
},
|
| 218566 |
+
{
|
| 218567 |
+
"epoch": 748.87,
|
| 218568 |
+
"learning_rate": 8.515096153846154e-06,
|
| 218569 |
+
"loss": 0.3404,
|
| 218570 |
+
"step": 93695
|
| 218571 |
+
},
|
| 218572 |
+
{
|
| 218573 |
+
"epoch": 748.91,
|
| 218574 |
+
"learning_rate": 8.515016025641027e-06,
|
| 218575 |
+
"loss": 0.2821,
|
| 218576 |
+
"step": 93700
|
| 218577 |
+
},
|
| 218578 |
+
{
|
| 218579 |
+
"epoch": 748.95,
|
| 218580 |
+
"learning_rate": 8.514935897435898e-06,
|
| 218581 |
+
"loss": 0.3785,
|
| 218582 |
+
"step": 93705
|
| 218583 |
+
},
|
| 218584 |
+
{
|
| 218585 |
+
"epoch": 748.99,
|
| 218586 |
+
"learning_rate": 8.51485576923077e-06,
|
| 218587 |
+
"loss": 0.966,
|
| 218588 |
+
"step": 93710
|
| 218589 |
+
},
|
| 218590 |
+
{
|
| 218591 |
+
"epoch": 749.0,
|
| 218592 |
+
"eval_loss": 0.4866238534450531,
|
| 218593 |
+
"eval_runtime": 38.5067,
|
| 218594 |
+
"eval_samples_per_second": 21.788,
|
| 218595 |
+
"eval_steps_per_second": 0.701,
|
| 218596 |
+
"eval_wer": 0.19120631341600902,
|
| 218597 |
+
"step": 93711
|
| 218598 |
+
},
|
| 218599 |
+
{
|
| 218600 |
+
"epoch": 749.03,
|
| 218601 |
+
"learning_rate": 8.514775641025642e-06,
|
| 218602 |
+
"loss": 0.369,
|
| 218603 |
+
"step": 93715
|
| 218604 |
+
},
|
| 218605 |
+
{
|
| 218606 |
+
"epoch": 749.07,
|
| 218607 |
+
"learning_rate": 8.514695512820514e-06,
|
| 218608 |
+
"loss": 0.3605,
|
| 218609 |
+
"step": 93720
|
| 218610 |
+
},
|
| 218611 |
+
{
|
| 218612 |
+
"epoch": 749.11,
|
| 218613 |
+
"learning_rate": 8.514615384615385e-06,
|
| 218614 |
+
"loss": 0.2692,
|
| 218615 |
+
"step": 93725
|
| 218616 |
+
},
|
| 218617 |
+
{
|
| 218618 |
+
"epoch": 749.15,
|
| 218619 |
+
"learning_rate": 8.514535256410257e-06,
|
| 218620 |
+
"loss": 0.4142,
|
| 218621 |
+
"step": 93730
|
| 218622 |
+
},
|
| 218623 |
+
{
|
| 218624 |
+
"epoch": 749.19,
|
| 218625 |
+
"learning_rate": 8.51445512820513e-06,
|
| 218626 |
+
"loss": 0.9357,
|
| 218627 |
+
"step": 93735
|
| 218628 |
+
},
|
| 218629 |
+
{
|
| 218630 |
+
"epoch": 749.23,
|
| 218631 |
+
"learning_rate": 8.514375000000001e-06,
|
| 218632 |
+
"loss": 0.6897,
|
| 218633 |
+
"step": 93740
|
| 218634 |
+
},
|
| 218635 |
+
{
|
| 218636 |
+
"epoch": 749.27,
|
| 218637 |
+
"learning_rate": 8.514294871794872e-06,
|
| 218638 |
+
"loss": 0.3877,
|
| 218639 |
+
"step": 93745
|
| 218640 |
+
},
|
| 218641 |
+
{
|
| 218642 |
+
"epoch": 749.31,
|
| 218643 |
+
"learning_rate": 8.514214743589744e-06,
|
| 218644 |
+
"loss": 0.3677,
|
| 218645 |
+
"step": 93750
|
| 218646 |
+
},
|
| 218647 |
+
{
|
| 218648 |
+
"epoch": 749.35,
|
| 218649 |
+
"learning_rate": 8.514134615384617e-06,
|
| 218650 |
+
"loss": 0.4088,
|
| 218651 |
+
"step": 93755
|
| 218652 |
+
},
|
| 218653 |
+
{
|
| 218654 |
+
"epoch": 749.39,
|
| 218655 |
+
"learning_rate": 8.514054487179488e-06,
|
| 218656 |
+
"loss": 0.8744,
|
| 218657 |
+
"step": 93760
|
| 218658 |
+
},
|
| 218659 |
+
{
|
| 218660 |
+
"epoch": 749.43,
|
| 218661 |
+
"learning_rate": 8.51397435897436e-06,
|
| 218662 |
+
"loss": 0.651,
|
| 218663 |
+
"step": 93765
|
| 218664 |
+
},
|
| 218665 |
+
{
|
| 218666 |
+
"epoch": 749.47,
|
| 218667 |
+
"learning_rate": 8.513894230769232e-06,
|
| 218668 |
+
"loss": 0.2776,
|
| 218669 |
+
"step": 93770
|
| 218670 |
+
},
|
| 218671 |
+
{
|
| 218672 |
+
"epoch": 749.51,
|
| 218673 |
+
"learning_rate": 8.513814102564104e-06,
|
| 218674 |
+
"loss": 0.4864,
|
| 218675 |
+
"step": 93775
|
| 218676 |
+
},
|
| 218677 |
+
{
|
| 218678 |
+
"epoch": 749.55,
|
| 218679 |
+
"learning_rate": 8.513733974358975e-06,
|
| 218680 |
+
"loss": 0.4179,
|
| 218681 |
+
"step": 93780
|
| 218682 |
+
},
|
| 218683 |
+
{
|
| 218684 |
+
"epoch": 749.59,
|
| 218685 |
+
"learning_rate": 8.513653846153847e-06,
|
| 218686 |
+
"loss": 0.9494,
|
| 218687 |
+
"step": 93785
|
| 218688 |
+
},
|
| 218689 |
+
{
|
| 218690 |
+
"epoch": 749.63,
|
| 218691 |
+
"learning_rate": 8.51357371794872e-06,
|
| 218692 |
+
"loss": 0.6507,
|
| 218693 |
+
"step": 93790
|
| 218694 |
+
},
|
| 218695 |
+
{
|
| 218696 |
+
"epoch": 749.67,
|
| 218697 |
+
"learning_rate": 8.51349358974359e-06,
|
| 218698 |
+
"loss": 0.3145,
|
| 218699 |
+
"step": 93795
|
| 218700 |
+
},
|
| 218701 |
+
{
|
| 218702 |
+
"epoch": 749.71,
|
| 218703 |
+
"learning_rate": 8.513413461538462e-06,
|
| 218704 |
+
"loss": 0.3237,
|
| 218705 |
+
"step": 93800
|
| 218706 |
+
},
|
| 218707 |
+
{
|
| 218708 |
+
"epoch": 749.75,
|
| 218709 |
+
"learning_rate": 8.513333333333335e-06,
|
| 218710 |
+
"loss": 0.397,
|
| 218711 |
+
"step": 93805
|
| 218712 |
+
},
|
| 218713 |
+
{
|
| 218714 |
+
"epoch": 749.79,
|
| 218715 |
+
"learning_rate": 8.513253205128205e-06,
|
| 218716 |
+
"loss": 0.9229,
|
| 218717 |
+
"step": 93810
|
| 218718 |
+
},
|
| 218719 |
+
{
|
| 218720 |
+
"epoch": 749.83,
|
| 218721 |
+
"learning_rate": 8.513173076923078e-06,
|
| 218722 |
+
"loss": 0.6785,
|
| 218723 |
+
"step": 93815
|
| 218724 |
+
},
|
| 218725 |
+
{
|
| 218726 |
+
"epoch": 749.87,
|
| 218727 |
+
"learning_rate": 8.51309294871795e-06,
|
| 218728 |
+
"loss": 0.3045,
|
| 218729 |
+
"step": 93820
|
| 218730 |
+
},
|
| 218731 |
+
{
|
| 218732 |
+
"epoch": 749.91,
|
| 218733 |
+
"learning_rate": 8.51301282051282e-06,
|
| 218734 |
+
"loss": 0.362,
|
| 218735 |
+
"step": 93825
|
| 218736 |
+
},
|
| 218737 |
+
{
|
| 218738 |
+
"epoch": 749.95,
|
| 218739 |
+
"learning_rate": 8.512932692307692e-06,
|
| 218740 |
+
"loss": 0.4119,
|
| 218741 |
+
"step": 93830
|
| 218742 |
+
},
|
| 218743 |
+
{
|
| 218744 |
+
"epoch": 749.99,
|
| 218745 |
+
"learning_rate": 8.512852564102565e-06,
|
| 218746 |
+
"loss": 1.0564,
|
| 218747 |
+
"step": 93835
|
| 218748 |
+
},
|
| 218749 |
+
{
|
| 218750 |
+
"epoch": 750.0,
|
| 218751 |
+
"eval_loss": 0.3657403886318207,
|
| 218752 |
+
"eval_runtime": 40.3739,
|
| 218753 |
+
"eval_samples_per_second": 20.756,
|
| 218754 |
+
"eval_steps_per_second": 0.669,
|
| 218755 |
+
"eval_wer": 0.18745001090671126,
|
| 218756 |
+
"step": 93836
|
| 218757 |
+
},
|
| 218758 |
+
{
|
| 218759 |
+
"epoch": 756.03,
|
| 218760 |
+
"learning_rate": 8.512772435897437e-06,
|
| 218761 |
+
"loss": 0.2971,
|
| 218762 |
+
"step": 93840
|
| 218763 |
+
},
|
| 218764 |
+
{
|
| 218765 |
+
"epoch": 756.07,
|
| 218766 |
+
"learning_rate": 8.512692307692308e-06,
|
| 218767 |
+
"loss": 0.3112,
|
| 218768 |
+
"step": 93845
|
| 218769 |
+
},
|
| 218770 |
+
{
|
| 218771 |
+
"epoch": 756.11,
|
| 218772 |
+
"learning_rate": 8.51261217948718e-06,
|
| 218773 |
+
"loss": 0.323,
|
| 218774 |
+
"step": 93850
|
| 218775 |
+
},
|
| 218776 |
+
{
|
| 218777 |
+
"epoch": 756.15,
|
| 218778 |
+
"learning_rate": 8.512532051282052e-06,
|
| 218779 |
+
"loss": 0.3942,
|
| 218780 |
+
"step": 93855
|
| 218781 |
+
},
|
| 218782 |
+
{
|
| 218783 |
+
"epoch": 756.19,
|
| 218784 |
+
"learning_rate": 8.512451923076924e-06,
|
| 218785 |
+
"loss": 0.9728,
|
| 218786 |
+
"step": 93860
|
| 218787 |
+
},
|
| 218788 |
+
{
|
| 218789 |
+
"epoch": 756.23,
|
| 218790 |
+
"learning_rate": 8.512371794871795e-06,
|
| 218791 |
+
"loss": 0.6976,
|
| 218792 |
+
"step": 93865
|
| 218793 |
+
},
|
| 218794 |
+
{
|
| 218795 |
+
"epoch": 756.27,
|
| 218796 |
+
"learning_rate": 8.512291666666668e-06,
|
| 218797 |
+
"loss": 0.3086,
|
| 218798 |
+
"step": 93870
|
| 218799 |
+
},
|
| 218800 |
+
{
|
| 218801 |
+
"epoch": 756.31,
|
| 218802 |
+
"learning_rate": 8.51221153846154e-06,
|
| 218803 |
+
"loss": 0.2799,
|
| 218804 |
+
"step": 93875
|
| 218805 |
+
},
|
| 218806 |
+
{
|
| 218807 |
+
"epoch": 756.35,
|
| 218808 |
+
"learning_rate": 8.512131410256411e-06,
|
| 218809 |
+
"loss": 0.3701,
|
| 218810 |
+
"step": 93880
|
| 218811 |
+
},
|
| 218812 |
+
{
|
| 218813 |
+
"epoch": 756.39,
|
| 218814 |
+
"learning_rate": 8.512051282051282e-06,
|
| 218815 |
+
"loss": 0.9532,
|
| 218816 |
+
"step": 93885
|
| 218817 |
+
},
|
| 218818 |
+
{
|
| 218819 |
+
"epoch": 756.43,
|
| 218820 |
+
"learning_rate": 8.511971153846155e-06,
|
| 218821 |
+
"loss": 0.7129,
|
| 218822 |
+
"step": 93890
|
| 218823 |
+
},
|
| 218824 |
+
{
|
| 218825 |
+
"epoch": 756.47,
|
| 218826 |
+
"learning_rate": 8.511891025641027e-06,
|
| 218827 |
+
"loss": 0.308,
|
| 218828 |
+
"step": 93895
|
| 218829 |
+
},
|
| 218830 |
+
{
|
| 218831 |
+
"epoch": 756.51,
|
| 218832 |
+
"learning_rate": 8.511810897435898e-06,
|
| 218833 |
+
"loss": 0.322,
|
| 218834 |
+
"step": 93900
|
| 218835 |
+
},
|
| 218836 |
+
{
|
| 218837 |
+
"epoch": 756.55,
|
| 218838 |
+
"learning_rate": 8.511730769230771e-06,
|
| 218839 |
+
"loss": 0.4047,
|
| 218840 |
+
"step": 93905
|
| 218841 |
+
},
|
| 218842 |
+
{
|
| 218843 |
+
"epoch": 756.59,
|
| 218844 |
+
"learning_rate": 8.511650641025642e-06,
|
| 218845 |
+
"loss": 0.8819,
|
| 218846 |
+
"step": 93910
|
| 218847 |
+
},
|
| 218848 |
+
{
|
| 218849 |
+
"epoch": 756.63,
|
| 218850 |
+
"learning_rate": 8.511570512820514e-06,
|
| 218851 |
+
"loss": 0.7582,
|
| 218852 |
+
"step": 93915
|
| 218853 |
+
},
|
| 218854 |
+
{
|
| 218855 |
+
"epoch": 756.67,
|
| 218856 |
+
"learning_rate": 8.511490384615385e-06,
|
| 218857 |
+
"loss": 0.286,
|
| 218858 |
+
"step": 93920
|
| 218859 |
+
},
|
| 218860 |
+
{
|
| 218861 |
+
"epoch": 756.71,
|
| 218862 |
+
"learning_rate": 8.511410256410258e-06,
|
| 218863 |
+
"loss": 0.326,
|
| 218864 |
+
"step": 93925
|
| 218865 |
+
},
|
| 218866 |
+
{
|
| 218867 |
+
"epoch": 756.76,
|
| 218868 |
+
"learning_rate": 8.511330128205128e-06,
|
| 218869 |
+
"loss": 0.4234,
|
| 218870 |
+
"step": 93930
|
| 218871 |
+
},
|
| 218872 |
+
{
|
| 218873 |
+
"epoch": 756.8,
|
| 218874 |
+
"learning_rate": 8.511250000000001e-06,
|
| 218875 |
+
"loss": 0.8456,
|
| 218876 |
+
"step": 93935
|
| 218877 |
+
},
|
| 218878 |
+
{
|
| 218879 |
+
"epoch": 756.84,
|
| 218880 |
+
"learning_rate": 8.511169871794872e-06,
|
| 218881 |
+
"loss": 0.6125,
|
| 218882 |
+
"step": 93940
|
| 218883 |
+
},
|
| 218884 |
+
{
|
| 218885 |
+
"epoch": 756.88,
|
| 218886 |
+
"learning_rate": 8.511089743589744e-06,
|
| 218887 |
+
"loss": 0.2815,
|
| 218888 |
+
"step": 93945
|
| 218889 |
+
},
|
| 218890 |
+
{
|
| 218891 |
+
"epoch": 756.92,
|
| 218892 |
+
"learning_rate": 8.511009615384615e-06,
|
| 218893 |
+
"loss": 0.3681,
|
| 218894 |
+
"step": 93950
|
| 218895 |
+
},
|
| 218896 |
+
{
|
| 218897 |
+
"epoch": 756.96,
|
| 218898 |
+
"learning_rate": 8.510929487179488e-06,
|
| 218899 |
+
"loss": 0.4144,
|
| 218900 |
+
"step": 93955
|
| 218901 |
+
},
|
| 218902 |
+
{
|
| 218903 |
+
"epoch": 757.0,
|
| 218904 |
+
"learning_rate": 8.51084935897436e-06,
|
| 218905 |
+
"loss": 1.381,
|
| 218906 |
+
"step": 93960
|
| 218907 |
+
},
|
| 218908 |
+
{
|
| 218909 |
+
"epoch": 757.0,
|
| 218910 |
+
"eval_loss": 0.3931798040866852,
|
| 218911 |
+
"eval_runtime": 39.9111,
|
| 218912 |
+
"eval_samples_per_second": 20.997,
|
| 218913 |
+
"eval_steps_per_second": 0.677,
|
| 218914 |
+
"eval_wer": 0.18971618401097712,
|
| 218915 |
+
"step": 93960
|
| 218916 |
+
},
|
| 218917 |
+
{
|
| 218918 |
+
"epoch": 751.04,
|
| 218919 |
+
"learning_rate": 8.51076923076923e-06,
|
| 218920 |
+
"loss": 0.3278,
|
| 218921 |
+
"step": 93965
|
| 218922 |
+
},
|
| 218923 |
+
{
|
| 218924 |
+
"epoch": 751.08,
|
| 218925 |
+
"learning_rate": 8.510689102564104e-06,
|
| 218926 |
+
"loss": 0.2895,
|
| 218927 |
+
"step": 93970
|
| 218928 |
+
},
|
| 218929 |
+
{
|
| 218930 |
+
"epoch": 751.12,
|
| 218931 |
+
"learning_rate": 8.510608974358975e-06,
|
| 218932 |
+
"loss": 0.3708,
|
| 218933 |
+
"step": 93975
|
| 218934 |
+
},
|
| 218935 |
+
{
|
| 218936 |
+
"epoch": 751.16,
|
| 218937 |
+
"learning_rate": 8.510528846153846e-06,
|
| 218938 |
+
"loss": 0.5214,
|
| 218939 |
+
"step": 93980
|
| 218940 |
+
},
|
| 218941 |
+
{
|
| 218942 |
+
"epoch": 751.2,
|
| 218943 |
+
"learning_rate": 8.510448717948718e-06,
|
| 218944 |
+
"loss": 1.45,
|
| 218945 |
+
"step": 93985
|
| 218946 |
+
},
|
| 218947 |
+
{
|
| 218948 |
+
"epoch": 751.24,
|
| 218949 |
+
"learning_rate": 8.510368589743591e-06,
|
| 218950 |
+
"loss": 0.2901,
|
| 218951 |
+
"step": 93990
|
| 218952 |
+
},
|
| 218953 |
+
{
|
| 218954 |
+
"epoch": 751.28,
|
| 218955 |
+
"learning_rate": 8.510288461538462e-06,
|
| 218956 |
+
"loss": 0.2749,
|
| 218957 |
+
"step": 93995
|
| 218958 |
+
},
|
| 218959 |
+
{
|
| 218960 |
+
"epoch": 751.32,
|
| 218961 |
+
"learning_rate": 8.510208333333334e-06,
|
| 218962 |
+
"loss": 0.3866,
|
| 218963 |
+
"step": 94000
|
| 218964 |
+
},
|
| 218965 |
+
{
|
| 218966 |
+
"epoch": 751.36,
|
| 218967 |
+
"learning_rate": 8.510128205128207e-06,
|
| 218968 |
+
"loss": 0.5277,
|
| 218969 |
+
"step": 94005
|
| 218970 |
+
},
|
| 218971 |
+
{
|
| 218972 |
+
"epoch": 751.4,
|
| 218973 |
+
"learning_rate": 8.510048076923078e-06,
|
| 218974 |
+
"loss": 1.0501,
|
| 218975 |
+
"step": 94010
|
| 218976 |
+
},
|
| 218977 |
+
{
|
| 218978 |
+
"epoch": 751.44,
|
| 218979 |
+
"learning_rate": 8.50996794871795e-06,
|
| 218980 |
+
"loss": 0.3593,
|
| 218981 |
+
"step": 94015
|
| 218982 |
+
},
|
| 218983 |
+
{
|
| 218984 |
+
"epoch": 751.48,
|
| 218985 |
+
"learning_rate": 8.50988782051282e-06,
|
| 218986 |
+
"loss": 0.2661,
|
| 218987 |
+
"step": 94020
|
| 218988 |
+
},
|
| 218989 |
+
{
|
| 218990 |
+
"epoch": 751.52,
|
| 218991 |
+
"learning_rate": 8.509807692307694e-06,
|
| 218992 |
+
"loss": 0.3336,
|
| 218993 |
+
"step": 94025
|
| 218994 |
+
},
|
| 218995 |
+
{
|
| 218996 |
+
"epoch": 751.56,
|
| 218997 |
+
"learning_rate": 8.509727564102565e-06,
|
| 218998 |
+
"loss": 0.4038,
|
| 218999 |
+
"step": 94030
|
| 219000 |
+
},
|
| 219001 |
+
{
|
| 219002 |
+
"epoch": 751.6,
|
| 219003 |
+
"learning_rate": 8.509647435897437e-06,
|
| 219004 |
+
"loss": 1.0587,
|
| 219005 |
+
"step": 94035
|
| 219006 |
+
},
|
| 219007 |
+
{
|
| 219008 |
+
"epoch": 751.64,
|
| 219009 |
+
"learning_rate": 8.509567307692308e-06,
|
| 219010 |
+
"loss": 0.3351,
|
| 219011 |
+
"step": 94040
|
| 219012 |
+
},
|
| 219013 |
+
{
|
| 219014 |
+
"epoch": 751.68,
|
| 219015 |
+
"learning_rate": 8.509487179487181e-06,
|
| 219016 |
+
"loss": 0.2673,
|
| 219017 |
+
"step": 94045
|
| 219018 |
+
},
|
| 219019 |
+
{
|
| 219020 |
+
"epoch": 751.72,
|
| 219021 |
+
"learning_rate": 8.509407051282052e-06,
|
| 219022 |
+
"loss": 0.3,
|
| 219023 |
+
"step": 94050
|
| 219024 |
+
},
|
| 219025 |
+
{
|
| 219026 |
+
"epoch": 751.76,
|
| 219027 |
+
"learning_rate": 8.509326923076924e-06,
|
| 219028 |
+
"loss": 0.4632,
|
| 219029 |
+
"step": 94055
|
| 219030 |
+
},
|
| 219031 |
+
{
|
| 219032 |
+
"epoch": 751.8,
|
| 219033 |
+
"learning_rate": 8.509246794871797e-06,
|
| 219034 |
+
"loss": 1.0868,
|
| 219035 |
+
"step": 94060
|
| 219036 |
+
},
|
| 219037 |
+
{
|
| 219038 |
+
"epoch": 751.84,
|
| 219039 |
+
"learning_rate": 8.509166666666668e-06,
|
| 219040 |
+
"loss": 0.3317,
|
| 219041 |
+
"step": 94065
|
| 219042 |
+
},
|
| 219043 |
+
{
|
| 219044 |
+
"epoch": 751.88,
|
| 219045 |
+
"learning_rate": 8.50908653846154e-06,
|
| 219046 |
+
"loss": 0.2943,
|
| 219047 |
+
"step": 94070
|
| 219048 |
+
},
|
| 219049 |
+
{
|
| 219050 |
+
"epoch": 751.92,
|
| 219051 |
+
"learning_rate": 8.50900641025641e-06,
|
| 219052 |
+
"loss": 0.3133,
|
| 219053 |
+
"step": 94075
|
| 219054 |
+
},
|
| 219055 |
+
{
|
| 219056 |
+
"epoch": 751.96,
|
| 219057 |
+
"learning_rate": 8.508926282051284e-06,
|
| 219058 |
+
"loss": 0.498,
|
| 219059 |
+
"step": 94080
|
| 219060 |
+
},
|
| 219061 |
+
{
|
| 219062 |
+
"epoch": 752.0,
|
| 219063 |
+
"learning_rate": 8.508846153846153e-06,
|
| 219064 |
+
"loss": 1.3055,
|
| 219065 |
+
"step": 94085
|
| 219066 |
+
},
|
| 219067 |
+
{
|
| 219068 |
+
"epoch": 752.0,
|
| 219069 |
+
"eval_loss": 0.3784423768520355,
|
| 219070 |
+
"eval_runtime": 39.658,
|
| 219071 |
+
"eval_samples_per_second": 21.131,
|
| 219072 |
+
"eval_steps_per_second": 0.681,
|
| 219073 |
+
"eval_wer": 0.1819756115159824,
|
| 219074 |
+
"step": 94085
|
| 219075 |
+
},
|
| 219076 |
+
{
|
| 219077 |
+
"epoch": 758.04,
|
| 219078 |
+
"learning_rate": 8.508766025641027e-06,
|
| 219079 |
+
"loss": 0.3594,
|
| 219080 |
+
"step": 94090
|
| 219081 |
+
},
|
| 219082 |
+
{
|
| 219083 |
+
"epoch": 758.08,
|
| 219084 |
+
"learning_rate": 8.508685897435898e-06,
|
| 219085 |
+
"loss": 0.3195,
|
| 219086 |
+
"step": 94095
|
| 219087 |
+
},
|
| 219088 |
+
{
|
| 219089 |
+
"epoch": 758.12,
|
| 219090 |
+
"learning_rate": 8.50860576923077e-06,
|
| 219091 |
+
"loss": 0.3556,
|
| 219092 |
+
"step": 94100
|
| 219093 |
+
},
|
| 219094 |
+
{
|
| 219095 |
+
"epoch": 758.16,
|
| 219096 |
+
"learning_rate": 8.508525641025642e-06,
|
| 219097 |
+
"loss": 0.4796,
|
| 219098 |
+
"step": 94105
|
| 219099 |
+
},
|
| 219100 |
+
{
|
| 219101 |
+
"epoch": 758.2,
|
| 219102 |
+
"learning_rate": 8.508445512820514e-06,
|
| 219103 |
+
"loss": 1.2387,
|
| 219104 |
+
"step": 94110
|
| 219105 |
+
},
|
| 219106 |
+
{
|
| 219107 |
+
"epoch": 758.24,
|
| 219108 |
+
"learning_rate": 8.508365384615385e-06,
|
| 219109 |
+
"loss": 0.3018,
|
| 219110 |
+
"step": 94115
|
| 219111 |
+
},
|
| 219112 |
+
{
|
| 219113 |
+
"epoch": 758.28,
|
| 219114 |
+
"learning_rate": 8.508285256410256e-06,
|
| 219115 |
+
"loss": 0.2822,
|
| 219116 |
+
"step": 94120
|
| 219117 |
+
},
|
| 219118 |
+
{
|
| 219119 |
+
"epoch": 758.32,
|
| 219120 |
+
"learning_rate": 8.50820512820513e-06,
|
| 219121 |
+
"loss": 0.352,
|
| 219122 |
+
"step": 94125
|
| 219123 |
+
},
|
| 219124 |
+
{
|
| 219125 |
+
"epoch": 758.36,
|
| 219126 |
+
"learning_rate": 8.508125e-06,
|
| 219127 |
+
"loss": 0.5246,
|
| 219128 |
+
"step": 94130
|
| 219129 |
+
},
|
| 219130 |
+
{
|
| 219131 |
+
"epoch": 758.4,
|
| 219132 |
+
"learning_rate": 8.508044871794872e-06,
|
| 219133 |
+
"loss": 1.2258,
|
| 219134 |
+
"step": 94135
|
| 219135 |
+
},
|
| 219136 |
+
{
|
| 219137 |
+
"epoch": 758.44,
|
| 219138 |
+
"learning_rate": 8.507964743589744e-06,
|
| 219139 |
+
"loss": 0.3109,
|
| 219140 |
+
"step": 94140
|
| 219141 |
+
},
|
| 219142 |
+
{
|
| 219143 |
+
"epoch": 758.48,
|
| 219144 |
+
"learning_rate": 8.507884615384617e-06,
|
| 219145 |
+
"loss": 0.2892,
|
| 219146 |
+
"step": 94145
|
| 219147 |
+
},
|
| 219148 |
+
{
|
| 219149 |
+
"epoch": 758.52,
|
| 219150 |
+
"learning_rate": 8.507804487179488e-06,
|
| 219151 |
+
"loss": 0.3321,
|
| 219152 |
+
"step": 94150
|
| 219153 |
+
},
|
| 219154 |
+
{
|
| 219155 |
+
"epoch": 758.56,
|
| 219156 |
+
"learning_rate": 8.50772435897436e-06,
|
| 219157 |
+
"loss": 0.468,
|
| 219158 |
+
"step": 94155
|
| 219159 |
+
},
|
| 219160 |
+
{
|
| 219161 |
+
"epoch": 758.6,
|
| 219162 |
+
"learning_rate": 8.507644230769232e-06,
|
| 219163 |
+
"loss": 1.1824,
|
| 219164 |
+
"step": 94160
|
| 219165 |
+
},
|
| 219166 |
+
{
|
| 219167 |
+
"epoch": 758.64,
|
| 219168 |
+
"learning_rate": 8.507564102564104e-06,
|
| 219169 |
+
"loss": 0.3597,
|
| 219170 |
+
"step": 94165
|
| 219171 |
+
},
|
| 219172 |
+
{
|
| 219173 |
+
"epoch": 758.68,
|
| 219174 |
+
"learning_rate": 8.507483974358975e-06,
|
| 219175 |
+
"loss": 0.331,
|
| 219176 |
+
"step": 94170
|
| 219177 |
+
},
|
| 219178 |
+
{
|
| 219179 |
+
"epoch": 758.72,
|
| 219180 |
+
"learning_rate": 8.507403846153846e-06,
|
| 219181 |
+
"loss": 0.3436,
|
| 219182 |
+
"step": 94175
|
| 219183 |
+
},
|
| 219184 |
+
{
|
| 219185 |
+
"epoch": 758.76,
|
| 219186 |
+
"learning_rate": 8.50732371794872e-06,
|
| 219187 |
+
"loss": 0.5242,
|
| 219188 |
+
"step": 94180
|
| 219189 |
+
},
|
| 219190 |
+
{
|
| 219191 |
+
"epoch": 758.8,
|
| 219192 |
+
"learning_rate": 8.50724358974359e-06,
|
| 219193 |
+
"loss": 1.3293,
|
| 219194 |
+
"step": 94185
|
| 219195 |
+
},
|
| 219196 |
+
{
|
| 219197 |
+
"epoch": 758.84,
|
| 219198 |
+
"learning_rate": 8.507163461538462e-06,
|
| 219199 |
+
"loss": 0.3482,
|
| 219200 |
+
"step": 94190
|
| 219201 |
+
},
|
| 219202 |
+
{
|
| 219203 |
+
"epoch": 758.88,
|
| 219204 |
+
"learning_rate": 8.507083333333334e-06,
|
| 219205 |
+
"loss": 0.28,
|
| 219206 |
+
"step": 94195
|
| 219207 |
+
},
|
| 219208 |
+
{
|
| 219209 |
+
"epoch": 758.92,
|
| 219210 |
+
"learning_rate": 8.507003205128207e-06,
|
| 219211 |
+
"loss": 0.287,
|
| 219212 |
+
"step": 94200
|
| 219213 |
+
},
|
| 219214 |
+
{
|
| 219215 |
+
"epoch": 758.96,
|
| 219216 |
+
"learning_rate": 8.506923076923078e-06,
|
| 219217 |
+
"loss": 0.5444,
|
| 219218 |
+
"step": 94205
|
| 219219 |
+
},
|
| 219220 |
+
{
|
| 219221 |
+
"epoch": 759.0,
|
| 219222 |
+
"eval_loss": 0.43794122338294983,
|
| 219223 |
+
"eval_runtime": 38.5307,
|
| 219224 |
+
"eval_samples_per_second": 21.749,
|
| 219225 |
+
"eval_steps_per_second": 0.701,
|
| 219226 |
+
"eval_wer": 0.1961866902237927,
|
| 219227 |
+
"step": 94209
|
| 219228 |
+
},
|
| 219229 |
+
{
|
| 219230 |
+
"epoch": 759.01,
|
| 219231 |
+
"learning_rate": 8.50684294871795e-06,
|
| 219232 |
+
"loss": 0.5366,
|
| 219233 |
+
"step": 94210
|
| 219234 |
+
},
|
| 219235 |
+
{
|
| 219236 |
+
"epoch": 759.05,
|
| 219237 |
+
"learning_rate": 8.506762820512822e-06,
|
| 219238 |
+
"loss": 0.3192,
|
| 219239 |
+
"step": 94215
|
| 219240 |
+
},
|
| 219241 |
+
{
|
| 219242 |
+
"epoch": 759.09,
|
| 219243 |
+
"learning_rate": 8.506682692307692e-06,
|
| 219244 |
+
"loss": 0.2532,
|
| 219245 |
+
"step": 94220
|
| 219246 |
+
},
|
| 219247 |
+
{
|
| 219248 |
+
"epoch": 759.13,
|
| 219249 |
+
"learning_rate": 8.506602564102565e-06,
|
| 219250 |
+
"loss": 0.3294,
|
| 219251 |
+
"step": 94225
|
| 219252 |
+
},
|
| 219253 |
+
{
|
| 219254 |
+
"epoch": 759.17,
|
| 219255 |
+
"learning_rate": 8.506522435897436e-06,
|
| 219256 |
+
"loss": 0.5267,
|
| 219257 |
+
"step": 94230
|
| 219258 |
+
},
|
| 219259 |
+
{
|
| 219260 |
+
"epoch": 759.21,
|
| 219261 |
+
"learning_rate": 8.506442307692308e-06,
|
| 219262 |
+
"loss": 1.1953,
|
| 219263 |
+
"step": 94235
|
| 219264 |
+
},
|
| 219265 |
+
{
|
| 219266 |
+
"epoch": 759.25,
|
| 219267 |
+
"learning_rate": 8.50636217948718e-06,
|
| 219268 |
+
"loss": 0.33,
|
| 219269 |
+
"step": 94240
|
| 219270 |
+
},
|
| 219271 |
+
{
|
| 219272 |
+
"epoch": 759.29,
|
| 219273 |
+
"learning_rate": 8.506282051282052e-06,
|
| 219274 |
+
"loss": 0.3258,
|
| 219275 |
+
"step": 94245
|
| 219276 |
+
},
|
| 219277 |
+
{
|
| 219278 |
+
"epoch": 759.33,
|
| 219279 |
+
"learning_rate": 8.506201923076924e-06,
|
| 219280 |
+
"loss": 0.4281,
|
| 219281 |
+
"step": 94250
|
| 219282 |
+
},
|
| 219283 |
+
{
|
| 219284 |
+
"epoch": 759.37,
|
| 219285 |
+
"learning_rate": 8.506121794871795e-06,
|
| 219286 |
+
"loss": 0.5508,
|
| 219287 |
+
"step": 94255
|
| 219288 |
+
},
|
| 219289 |
+
{
|
| 219290 |
+
"epoch": 759.41,
|
| 219291 |
+
"learning_rate": 8.506041666666668e-06,
|
| 219292 |
+
"loss": 1.1474,
|
| 219293 |
+
"step": 94260
|
| 219294 |
+
},
|
| 219295 |
+
{
|
| 219296 |
+
"epoch": 759.45,
|
| 219297 |
+
"learning_rate": 8.50596153846154e-06,
|
| 219298 |
+
"loss": 0.3027,
|
| 219299 |
+
"step": 94265
|
| 219300 |
+
},
|
| 219301 |
+
{
|
| 219302 |
+
"epoch": 759.49,
|
| 219303 |
+
"learning_rate": 8.50588141025641e-06,
|
| 219304 |
+
"loss": 0.3026,
|
| 219305 |
+
"step": 94270
|
| 219306 |
+
},
|
| 219307 |
+
{
|
| 219308 |
+
"epoch": 759.53,
|
| 219309 |
+
"learning_rate": 8.505801282051282e-06,
|
| 219310 |
+
"loss": 0.3823,
|
| 219311 |
+
"step": 94275
|
| 219312 |
+
},
|
| 219313 |
+
{
|
| 219314 |
+
"epoch": 759.57,
|
| 219315 |
+
"learning_rate": 8.505721153846155e-06,
|
| 219316 |
+
"loss": 0.5083,
|
| 219317 |
+
"step": 94280
|
| 219318 |
+
},
|
| 219319 |
+
{
|
| 219320 |
+
"epoch": 759.61,
|
| 219321 |
+
"learning_rate": 8.505641025641026e-06,
|
| 219322 |
+
"loss": 1.4258,
|
| 219323 |
+
"step": 94285
|
| 219324 |
+
},
|
| 219325 |
+
{
|
| 219326 |
+
"epoch": 759.65,
|
| 219327 |
+
"learning_rate": 8.505560897435898e-06,
|
| 219328 |
+
"loss": 0.3108,
|
| 219329 |
+
"step": 94290
|
| 219330 |
+
},
|
| 219331 |
+
{
|
| 219332 |
+
"epoch": 759.69,
|
| 219333 |
+
"learning_rate": 8.50548076923077e-06,
|
| 219334 |
+
"loss": 0.3102,
|
| 219335 |
+
"step": 94295
|
| 219336 |
+
},
|
| 219337 |
+
{
|
| 219338 |
+
"epoch": 759.73,
|
| 219339 |
+
"learning_rate": 8.505400641025642e-06,
|
| 219340 |
+
"loss": 0.3525,
|
| 219341 |
+
"step": 94300
|
| 219342 |
+
},
|
| 219343 |
+
{
|
| 219344 |
+
"epoch": 759.77,
|
| 219345 |
+
"learning_rate": 8.505320512820514e-06,
|
| 219346 |
+
"loss": 0.5332,
|
| 219347 |
+
"step": 94305
|
| 219348 |
+
},
|
| 219349 |
+
{
|
| 219350 |
+
"epoch": 759.81,
|
| 219351 |
+
"learning_rate": 8.505240384615385e-06,
|
| 219352 |
+
"loss": 1.0223,
|
| 219353 |
+
"step": 94310
|
| 219354 |
+
},
|
| 219355 |
+
{
|
| 219356 |
+
"epoch": 759.85,
|
| 219357 |
+
"learning_rate": 8.505160256410258e-06,
|
| 219358 |
+
"loss": 0.296,
|
| 219359 |
+
"step": 94315
|
| 219360 |
+
},
|
| 219361 |
+
{
|
| 219362 |
+
"epoch": 759.9,
|
| 219363 |
+
"learning_rate": 8.50508012820513e-06,
|
| 219364 |
+
"loss": 0.3009,
|
| 219365 |
+
"step": 94320
|
| 219366 |
+
},
|
| 219367 |
+
{
|
| 219368 |
+
"epoch": 759.94,
|
| 219369 |
+
"learning_rate": 8.505e-06,
|
| 219370 |
+
"loss": 0.3876,
|
| 219371 |
+
"step": 94325
|
| 219372 |
+
},
|
| 219373 |
+
{
|
| 219374 |
+
"epoch": 759.98,
|
| 219375 |
+
"learning_rate": 8.504919871794872e-06,
|
| 219376 |
+
"loss": 0.6779,
|
| 219377 |
+
"step": 94330
|
| 219378 |
+
},
|
| 219379 |
+
{
|
| 219380 |
+
"epoch": 760.0,
|
| 219381 |
+
"eval_loss": 0.358523428440094,
|
| 219382 |
+
"eval_runtime": 40.0339,
|
| 219383 |
+
"eval_samples_per_second": 20.932,
|
| 219384 |
+
"eval_steps_per_second": 0.674,
|
| 219385 |
+
"eval_wer": 0.17565318086415285,
|
| 219386 |
+
"step": 94333
|
| 219387 |
+
},
|
| 219388 |
+
{
|
| 219389 |
+
"epoch": 760.02,
|
| 219390 |
+
"learning_rate": 8.50485576923077e-06,
|
| 219391 |
+
"loss": 0.496,
|
| 219392 |
+
"step": 94335
|
| 219393 |
+
},
|
| 219394 |
+
{
|
| 219395 |
+
"epoch": 760.06,
|
| 219396 |
+
"learning_rate": 8.504775641025643e-06,
|
| 219397 |
+
"loss": 0.2538,
|
| 219398 |
+
"step": 94340
|
| 219399 |
+
},
|
| 219400 |
+
{
|
| 219401 |
+
"epoch": 760.1,
|
| 219402 |
+
"learning_rate": 8.504695512820514e-06,
|
| 219403 |
+
"loss": 0.288,
|
| 219404 |
+
"step": 94345
|
| 219405 |
+
},
|
| 219406 |
+
{
|
| 219407 |
+
"epoch": 760.14,
|
| 219408 |
+
"learning_rate": 8.504615384615385e-06,
|
| 219409 |
+
"loss": 0.317,
|
| 219410 |
+
"step": 94350
|
| 219411 |
+
},
|
| 219412 |
+
{
|
| 219413 |
+
"epoch": 760.18,
|
| 219414 |
+
"learning_rate": 8.504535256410257e-06,
|
| 219415 |
+
"loss": 0.6992,
|
| 219416 |
+
"step": 94355
|
| 219417 |
+
},
|
| 219418 |
+
{
|
| 219419 |
+
"epoch": 760.22,
|
| 219420 |
+
"learning_rate": 8.50445512820513e-06,
|
| 219421 |
+
"loss": 1.0059,
|
| 219422 |
+
"step": 94360
|
| 219423 |
+
},
|
| 219424 |
+
{
|
| 219425 |
+
"epoch": 760.26,
|
| 219426 |
+
"learning_rate": 8.504375000000001e-06,
|
| 219427 |
+
"loss": 0.3093,
|
| 219428 |
+
"step": 94365
|
| 219429 |
+
},
|
| 219430 |
+
{
|
| 219431 |
+
"epoch": 760.3,
|
| 219432 |
+
"learning_rate": 8.504294871794872e-06,
|
| 219433 |
+
"loss": 0.2887,
|
| 219434 |
+
"step": 94370
|
| 219435 |
+
},
|
| 219436 |
+
{
|
| 219437 |
+
"epoch": 760.34,
|
| 219438 |
+
"learning_rate": 8.504214743589744e-06,
|
| 219439 |
+
"loss": 0.3078,
|
| 219440 |
+
"step": 94375
|
| 219441 |
+
},
|
| 219442 |
+
{
|
| 219443 |
+
"epoch": 760.38,
|
| 219444 |
+
"learning_rate": 8.504134615384617e-06,
|
| 219445 |
+
"loss": 0.6396,
|
| 219446 |
+
"step": 94380
|
| 219447 |
+
},
|
| 219448 |
+
{
|
| 219449 |
+
"epoch": 760.42,
|
| 219450 |
+
"learning_rate": 8.504054487179488e-06,
|
| 219451 |
+
"loss": 0.9829,
|
| 219452 |
+
"step": 94385
|
| 219453 |
+
},
|
| 219454 |
+
{
|
| 219455 |
+
"epoch": 760.46,
|
| 219456 |
+
"learning_rate": 8.50397435897436e-06,
|
| 219457 |
+
"loss": 0.345,
|
| 219458 |
+
"step": 94390
|
| 219459 |
+
},
|
| 219460 |
+
{
|
| 219461 |
+
"epoch": 760.5,
|
| 219462 |
+
"learning_rate": 8.503894230769233e-06,
|
| 219463 |
+
"loss": 0.3304,
|
| 219464 |
+
"step": 94395
|
| 219465 |
+
},
|
| 219466 |
+
{
|
| 219467 |
+
"epoch": 760.54,
|
| 219468 |
+
"learning_rate": 8.503814102564102e-06,
|
| 219469 |
+
"loss": 0.3178,
|
| 219470 |
+
"step": 94400
|
| 219471 |
+
},
|
| 219472 |
+
{
|
| 219473 |
+
"epoch": 760.58,
|
| 219474 |
+
"learning_rate": 8.503733974358975e-06,
|
| 219475 |
+
"loss": 0.548,
|
| 219476 |
+
"step": 94405
|
| 219477 |
+
},
|
| 219478 |
+
{
|
| 219479 |
+
"epoch": 760.62,
|
| 219480 |
+
"learning_rate": 8.503653846153847e-06,
|
| 219481 |
+
"loss": 1.029,
|
| 219482 |
+
"step": 94410
|
| 219483 |
+
},
|
| 219484 |
+
{
|
| 219485 |
+
"epoch": 760.66,
|
| 219486 |
+
"learning_rate": 8.503573717948718e-06,
|
| 219487 |
+
"loss": 0.2847,
|
| 219488 |
+
"step": 94415
|
| 219489 |
+
},
|
| 219490 |
+
{
|
| 219491 |
+
"epoch": 760.7,
|
| 219492 |
+
"learning_rate": 8.50349358974359e-06,
|
| 219493 |
+
"loss": 0.3181,
|
| 219494 |
+
"step": 94420
|
| 219495 |
+
},
|
| 219496 |
+
{
|
| 219497 |
+
"epoch": 760.74,
|
| 219498 |
+
"learning_rate": 8.503413461538462e-06,
|
| 219499 |
+
"loss": 0.318,
|
| 219500 |
+
"step": 94425
|
| 219501 |
+
},
|
| 219502 |
+
{
|
| 219503 |
+
"epoch": 760.78,
|
| 219504 |
+
"learning_rate": 8.503333333333334e-06,
|
| 219505 |
+
"loss": 0.6047,
|
| 219506 |
+
"step": 94430
|
| 219507 |
+
},
|
| 219508 |
+
{
|
| 219509 |
+
"epoch": 760.82,
|
| 219510 |
+
"learning_rate": 8.503253205128205e-06,
|
| 219511 |
+
"loss": 1.0722,
|
| 219512 |
+
"step": 94435
|
| 219513 |
+
},
|
| 219514 |
+
{
|
| 219515 |
+
"epoch": 760.86,
|
| 219516 |
+
"learning_rate": 8.503173076923078e-06,
|
| 219517 |
+
"loss": 0.3217,
|
| 219518 |
+
"step": 94440
|
| 219519 |
+
},
|
| 219520 |
+
{
|
| 219521 |
+
"epoch": 760.9,
|
| 219522 |
+
"learning_rate": 8.50309294871795e-06,
|
| 219523 |
+
"loss": 0.2679,
|
| 219524 |
+
"step": 94445
|
| 219525 |
+
},
|
| 219526 |
+
{
|
| 219527 |
+
"epoch": 760.94,
|
| 219528 |
+
"learning_rate": 8.503012820512821e-06,
|
| 219529 |
+
"loss": 0.3798,
|
| 219530 |
+
"step": 94450
|
| 219531 |
+
},
|
| 219532 |
+
{
|
| 219533 |
+
"epoch": 760.98,
|
| 219534 |
+
"learning_rate": 8.502932692307692e-06,
|
| 219535 |
+
"loss": 0.7887,
|
| 219536 |
+
"step": 94455
|
| 219537 |
+
},
|
| 219538 |
+
{
|
| 219539 |
+
"epoch": 761.0,
|
| 219540 |
+
"eval_loss": 0.4737759232521057,
|
| 219541 |
+
"eval_runtime": 39.6244,
|
| 219542 |
+
"eval_samples_per_second": 21.149,
|
| 219543 |
+
"eval_steps_per_second": 0.681,
|
| 219544 |
+
"eval_wer": 0.19094138543516873,
|
| 219545 |
+
"step": 94457
|
| 219546 |
+
},
|
| 219547 |
+
{
|
| 219548 |
+
"epoch": 761.02,
|
| 219549 |
+
"learning_rate": 8.502852564102565e-06,
|
| 219550 |
+
"loss": 0.3958,
|
| 219551 |
+
"step": 94460
|
| 219552 |
+
},
|
| 219553 |
+
{
|
| 219554 |
+
"epoch": 761.06,
|
| 219555 |
+
"learning_rate": 8.502772435897437e-06,
|
| 219556 |
+
"loss": 0.3011,
|
| 219557 |
+
"step": 94465
|
| 219558 |
+
},
|
| 219559 |
+
{
|
| 219560 |
+
"epoch": 761.1,
|
| 219561 |
+
"learning_rate": 8.502692307692308e-06,
|
| 219562 |
+
"loss": 0.3545,
|
| 219563 |
+
"step": 94470
|
| 219564 |
+
},
|
| 219565 |
+
{
|
| 219566 |
+
"epoch": 761.14,
|
| 219567 |
+
"learning_rate": 8.50261217948718e-06,
|
| 219568 |
+
"loss": 0.4277,
|
| 219569 |
+
"step": 94475
|
| 219570 |
+
},
|
| 219571 |
+
{
|
| 219572 |
+
"epoch": 761.18,
|
| 219573 |
+
"learning_rate": 8.502532051282052e-06,
|
| 219574 |
+
"loss": 0.8837,
|
| 219575 |
+
"step": 94480
|
| 219576 |
+
},
|
| 219577 |
+
{
|
| 219578 |
+
"epoch": 761.22,
|
| 219579 |
+
"learning_rate": 8.502451923076924e-06,
|
| 219580 |
+
"loss": 1.0198,
|
| 219581 |
+
"step": 94485
|
| 219582 |
+
},
|
| 219583 |
+
{
|
| 219584 |
+
"epoch": 761.27,
|
| 219585 |
+
"learning_rate": 8.502371794871795e-06,
|
| 219586 |
+
"loss": 0.3615,
|
| 219587 |
+
"step": 94490
|
| 219588 |
+
},
|
| 219589 |
+
{
|
| 219590 |
+
"epoch": 761.31,
|
| 219591 |
+
"learning_rate": 8.502291666666668e-06,
|
| 219592 |
+
"loss": 0.2866,
|
| 219593 |
+
"step": 94495
|
| 219594 |
+
},
|
| 219595 |
+
{
|
| 219596 |
+
"epoch": 761.35,
|
| 219597 |
+
"learning_rate": 8.50221153846154e-06,
|
| 219598 |
+
"loss": 0.3173,
|
| 219599 |
+
"step": 94500
|
| 219600 |
+
},
|
| 219601 |
+
{
|
| 219602 |
+
"epoch": 761.39,
|
| 219603 |
+
"learning_rate": 8.502131410256411e-06,
|
| 219604 |
+
"loss": 0.7403,
|
| 219605 |
+
"step": 94505
|
| 219606 |
+
},
|
| 219607 |
+
{
|
| 219608 |
+
"epoch": 761.43,
|
| 219609 |
+
"learning_rate": 8.502051282051282e-06,
|
| 219610 |
+
"loss": 0.9529,
|
| 219611 |
+
"step": 94510
|
| 219612 |
+
},
|
| 219613 |
+
{
|
| 219614 |
+
"epoch": 761.47,
|
| 219615 |
+
"learning_rate": 8.501971153846155e-06,
|
| 219616 |
+
"loss": 0.3188,
|
| 219617 |
+
"step": 94515
|
| 219618 |
+
},
|
| 219619 |
+
{
|
| 219620 |
+
"epoch": 761.51,
|
| 219621 |
+
"learning_rate": 8.501891025641025e-06,
|
| 219622 |
+
"loss": 0.2888,
|
| 219623 |
+
"step": 94520
|
| 219624 |
+
},
|
| 219625 |
+
{
|
| 219626 |
+
"epoch": 761.55,
|
| 219627 |
+
"learning_rate": 8.501810897435898e-06,
|
| 219628 |
+
"loss": 0.3427,
|
| 219629 |
+
"step": 94525
|
| 219630 |
+
},
|
| 219631 |
+
{
|
| 219632 |
+
"epoch": 761.59,
|
| 219633 |
+
"learning_rate": 8.501730769230771e-06,
|
| 219634 |
+
"loss": 0.7738,
|
| 219635 |
+
"step": 94530
|
| 219636 |
+
},
|
| 219637 |
+
{
|
| 219638 |
+
"epoch": 761.63,
|
| 219639 |
+
"learning_rate": 8.50165064102564e-06,
|
| 219640 |
+
"loss": 0.9023,
|
| 219641 |
+
"step": 94535
|
| 219642 |
+
},
|
| 219643 |
+
{
|
| 219644 |
+
"epoch": 761.67,
|
| 219645 |
+
"learning_rate": 8.501570512820514e-06,
|
| 219646 |
+
"loss": 0.2613,
|
| 219647 |
+
"step": 94540
|
| 219648 |
+
},
|
| 219649 |
+
{
|
| 219650 |
+
"epoch": 761.71,
|
| 219651 |
+
"learning_rate": 8.501490384615385e-06,
|
| 219652 |
+
"loss": 0.2663,
|
| 219653 |
+
"step": 94545
|
| 219654 |
+
},
|
| 219655 |
+
{
|
| 219656 |
+
"epoch": 761.75,
|
| 219657 |
+
"learning_rate": 8.501410256410257e-06,
|
| 219658 |
+
"loss": 0.3923,
|
| 219659 |
+
"step": 94550
|
| 219660 |
+
},
|
| 219661 |
+
{
|
| 219662 |
+
"epoch": 761.79,
|
| 219663 |
+
"learning_rate": 8.501330128205128e-06,
|
| 219664 |
+
"loss": 0.8863,
|
| 219665 |
+
"step": 94555
|
| 219666 |
+
},
|
| 219667 |
+
{
|
| 219668 |
+
"epoch": 761.83,
|
| 219669 |
+
"learning_rate": 8.501250000000001e-06,
|
| 219670 |
+
"loss": 0.82,
|
| 219671 |
+
"step": 94560
|
| 219672 |
+
},
|
| 219673 |
+
{
|
| 219674 |
+
"epoch": 761.87,
|
| 219675 |
+
"learning_rate": 8.501169871794872e-06,
|
| 219676 |
+
"loss": 0.3776,
|
| 219677 |
+
"step": 94565
|
| 219678 |
+
},
|
| 219679 |
+
{
|
| 219680 |
+
"epoch": 761.91,
|
| 219681 |
+
"learning_rate": 8.501089743589744e-06,
|
| 219682 |
+
"loss": 0.322,
|
| 219683 |
+
"step": 94570
|
| 219684 |
+
},
|
| 219685 |
+
{
|
| 219686 |
+
"epoch": 761.95,
|
| 219687 |
+
"learning_rate": 8.501009615384615e-06,
|
| 219688 |
+
"loss": 0.3873,
|
| 219689 |
+
"step": 94575
|
| 219690 |
+
},
|
| 219691 |
+
{
|
| 219692 |
+
"epoch": 761.99,
|
| 219693 |
+
"learning_rate": 8.500929487179488e-06,
|
| 219694 |
+
"loss": 0.9905,
|
| 219695 |
+
"step": 94580
|
| 219696 |
+
},
|
| 219697 |
+
{
|
| 219698 |
+
"epoch": 762.0,
|
| 219699 |
+
"eval_loss": 0.3278201222419739,
|
| 219700 |
+
"eval_runtime": 40.3116,
|
| 219701 |
+
"eval_samples_per_second": 20.813,
|
| 219702 |
+
"eval_steps_per_second": 0.67,
|
| 219703 |
+
"eval_wer": 0.1897731364782139,
|
| 219704 |
+
"step": 94581
|
| 219705 |
}
|
| 219706 |
],
|
| 219707 |
+
"max_steps": 620000,
|
| 219708 |
"num_train_epochs": 5000,
|
| 219709 |
+
"total_flos": 2.6615641399079443e+20,
|
| 219710 |
"trial_name": null,
|
| 219711 |
"trial_params": null
|
| 219712 |
}
|
model-bin/finetune/base/{checkpoint-93960 β checkpoint-94581}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1629959918.4620056/events.out.tfevents.1629959918.8e89bd551565.924.91
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:808102867fe2c0dff126f27e6358c5f69bddfaa2267f9642fd4df1668233452a
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629960372.9279814/events.out.tfevents.1629960372.8e89bd551565.924.93
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9bb638604ed6dcc35a829ad879f437fdbc325ca62fed3c25440f9a324fdbf289
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629960821.1668708/events.out.tfevents.1629960821.8e89bd551565.924.95
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a55fc98cb78ae17d951ee5425b162c7e0fc49537dbee6d808b38ddafa497ae58
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629961261.3675253/events.out.tfevents.1629961261.8e89bd551565.924.97
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:721a3622e0c5aec94e968985176f3cb308d26153741ec1a80852147a9c598c9d
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1629961689.1069834/events.out.tfevents.1629961689.8e89bd551565.924.99
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e1ffb0eead90a03cd9fbe26ae6b80878680dd9fa01566133f3cb5932f4d4b170
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1629959918.8e89bd551565.924.90
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:506efca73eee16fc0611fab57834bc1e9af8d921ba3523ac642e8ca1415d8444
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629960372.8e89bd551565.924.92
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fe1ed39d0d5eaff9aa309a9460307d219164978408ad23d046810f1de0b1fa2d
|
| 3 |
+
size 8462
|
model-bin/finetune/base/log/events.out.tfevents.1629960819.8e89bd551565.924.94
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c2864d24e2601ec9eb3e204336e9df575ab12149d958834f7bc22bfc362ebcbc
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629961261.8e89bd551565.924.96
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0e72176c8c74b72c3a162a21a10dd42c06b9cef75fb424125de1e4477b8a91af
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1629961689.8e89bd551565.924.98
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9977143a20a7dbdaad9264094aa46ae3f15861b4d33b3e1e86d3eae1f8107fe8
|
| 3 |
+
size 8622
|