"auto-commit"
Browse files- model-bin/finetune/base/{checkpoint-145602 β checkpoint-146596}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-145602 β checkpoint-146596}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-145602 β checkpoint-146596}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-146472 β checkpoint-146596}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-146472 β checkpoint-146596}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-146472 β checkpoint-146596}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-145602 β checkpoint-146596}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-146472 β checkpoint-146596}/trainer_state.json +164 -5
- model-bin/finetune/base/{checkpoint-145602 β checkpoint-146596}/training_args.bin +0 -0
- model-bin/finetune/base/{checkpoint-146472 β checkpoint-147094}/config.json +0 -0
- model-bin/finetune/base/{checkpoint-146472 β checkpoint-147094}/optimizer.pt +1 -1
- model-bin/finetune/base/{checkpoint-146472 β checkpoint-147094}/preprocessor_config.json +0 -0
- model-bin/finetune/base/{checkpoint-145602 β checkpoint-147094}/pytorch_model.bin +1 -1
- model-bin/finetune/base/{checkpoint-145602 β checkpoint-147094}/rng_state.pth +1 -1
- model-bin/finetune/base/{checkpoint-145602 β checkpoint-147094}/scaler.pt +1 -1
- model-bin/finetune/base/{checkpoint-146472 β checkpoint-147094}/scheduler.pt +1 -1
- model-bin/finetune/base/{checkpoint-145602 β checkpoint-147094}/trainer_state.json +1902 -6
- model-bin/finetune/base/{checkpoint-146472 β checkpoint-147094}/training_args.bin +0 -0
- model-bin/finetune/base/log/1630230484.4148061/events.out.tfevents.1630230484.cc93b136ebf5.1086.119 +3 -0
- model-bin/finetune/base/log/1630230917.0829825/events.out.tfevents.1630230917.cc93b136ebf5.1086.121 +3 -0
- model-bin/finetune/base/log/1630231362.9249244/events.out.tfevents.1630231362.cc93b136ebf5.1086.123 +3 -0
- model-bin/finetune/base/log/1630231799.7950127/events.out.tfevents.1630231799.cc93b136ebf5.1086.125 +3 -0
- model-bin/finetune/base/log/1630232234.2438745/events.out.tfevents.1630232234.cc93b136ebf5.1086.127 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630230484.cc93b136ebf5.1086.118 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630230917.cc93b136ebf5.1086.120 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630231362.cc93b136ebf5.1086.122 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630231799.cc93b136ebf5.1086.124 +3 -0
- model-bin/finetune/base/log/events.out.tfevents.1630232234.cc93b136ebf5.1086.126 +3 -0
model-bin/finetune/base/{checkpoint-145602 β checkpoint-146596}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-145602 β checkpoint-146596}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a0ea9552f899d2d8f535fc434267063e77290df22b3ea0d489d5f3bcb93f46d0
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-145602 β checkpoint-146596}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-146472 β checkpoint-146596}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8dc25b2405ed70ef3243f1c3b26f7665ff79883dac75f72b56452985fad0a79b
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-146472 β checkpoint-146596}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9340a78af637a8c0ce5c4b0fa3f2905a42f8cd1ccd9b627b31325f5eaab00977
|
| 3 |
size 14503
|
model-bin/finetune/base/{checkpoint-146472 β checkpoint-146596}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:998b6be1a485973ded56fe9124d5ea9cfa3bd59e3499608d8b1f8a68c1d7c92c
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-145602 β checkpoint-146596}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:70ffda1a735fe44bfbcdb5fdd2fcf7dbc1ddead2a83c5e723492366f2b5337c5
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-146472 β checkpoint-146596}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
-
"best_metric": 0.
|
| 3 |
-
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -285723,11 +285723,170 @@
|
|
| 285723 |
"eval_steps_per_second": 0.645,
|
| 285724 |
"eval_wer": 0.1758169934640523,
|
| 285725 |
"step": 146472
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 285726 |
}
|
| 285727 |
],
|
| 285728 |
"max_steps": 620000,
|
| 285729 |
"num_train_epochs": 5000,
|
| 285730 |
-
"total_flos": 4.
|
| 285731 |
"trial_name": null,
|
| 285732 |
"trial_params": null
|
| 285733 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"best_metric": 0.1668011442822563,
|
| 3 |
+
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
| 4 |
+
"epoch": 1181.995983935743,
|
| 5 |
+
"global_step": 146596,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 285723 |
"eval_steps_per_second": 0.645,
|
| 285724 |
"eval_wer": 0.1758169934640523,
|
| 285725 |
"step": 146472
|
| 285726 |
+
},
|
| 285727 |
+
{
|
| 285728 |
+
"epoch": 1181.02,
|
| 285729 |
+
"learning_rate": 7.65075928917609e-06,
|
| 285730 |
+
"loss": 0.306,
|
| 285731 |
+
"step": 146475
|
| 285732 |
+
},
|
| 285733 |
+
{
|
| 285734 |
+
"epoch": 1181.06,
|
| 285735 |
+
"learning_rate": 7.650678513731826e-06,
|
| 285736 |
+
"loss": 0.3149,
|
| 285737 |
+
"step": 146480
|
| 285738 |
+
},
|
| 285739 |
+
{
|
| 285740 |
+
"epoch": 1181.1,
|
| 285741 |
+
"learning_rate": 7.65059773828756e-06,
|
| 285742 |
+
"loss": 0.2999,
|
| 285743 |
+
"step": 146485
|
| 285744 |
+
},
|
| 285745 |
+
{
|
| 285746 |
+
"epoch": 1181.14,
|
| 285747 |
+
"learning_rate": 7.650516962843296e-06,
|
| 285748 |
+
"loss": 0.3347,
|
| 285749 |
+
"step": 146490
|
| 285750 |
+
},
|
| 285751 |
+
{
|
| 285752 |
+
"epoch": 1181.18,
|
| 285753 |
+
"learning_rate": 7.650436187399032e-06,
|
| 285754 |
+
"loss": 0.6613,
|
| 285755 |
+
"step": 146495
|
| 285756 |
+
},
|
| 285757 |
+
{
|
| 285758 |
+
"epoch": 1181.22,
|
| 285759 |
+
"learning_rate": 7.650355411954766e-06,
|
| 285760 |
+
"loss": 0.8353,
|
| 285761 |
+
"step": 146500
|
| 285762 |
+
},
|
| 285763 |
+
{
|
| 285764 |
+
"epoch": 1181.27,
|
| 285765 |
+
"learning_rate": 7.650274636510502e-06,
|
| 285766 |
+
"loss": 0.3035,
|
| 285767 |
+
"step": 146505
|
| 285768 |
+
},
|
| 285769 |
+
{
|
| 285770 |
+
"epoch": 1181.31,
|
| 285771 |
+
"learning_rate": 7.650193861066236e-06,
|
| 285772 |
+
"loss": 0.3257,
|
| 285773 |
+
"step": 146510
|
| 285774 |
+
},
|
| 285775 |
+
{
|
| 285776 |
+
"epoch": 1181.35,
|
| 285777 |
+
"learning_rate": 7.650113085621972e-06,
|
| 285778 |
+
"loss": 0.4288,
|
| 285779 |
+
"step": 146515
|
| 285780 |
+
},
|
| 285781 |
+
{
|
| 285782 |
+
"epoch": 1181.39,
|
| 285783 |
+
"learning_rate": 7.650032310177706e-06,
|
| 285784 |
+
"loss": 0.6413,
|
| 285785 |
+
"step": 146520
|
| 285786 |
+
},
|
| 285787 |
+
{
|
| 285788 |
+
"epoch": 1181.43,
|
| 285789 |
+
"learning_rate": 7.649951534733442e-06,
|
| 285790 |
+
"loss": 0.8146,
|
| 285791 |
+
"step": 146525
|
| 285792 |
+
},
|
| 285793 |
+
{
|
| 285794 |
+
"epoch": 1181.47,
|
| 285795 |
+
"learning_rate": 7.649870759289176e-06,
|
| 285796 |
+
"loss": 0.2767,
|
| 285797 |
+
"step": 146530
|
| 285798 |
+
},
|
| 285799 |
+
{
|
| 285800 |
+
"epoch": 1181.51,
|
| 285801 |
+
"learning_rate": 7.649789983844912e-06,
|
| 285802 |
+
"loss": 0.2579,
|
| 285803 |
+
"step": 146535
|
| 285804 |
+
},
|
| 285805 |
+
{
|
| 285806 |
+
"epoch": 1181.55,
|
| 285807 |
+
"learning_rate": 7.649709208400646e-06,
|
| 285808 |
+
"loss": 0.3754,
|
| 285809 |
+
"step": 146540
|
| 285810 |
+
},
|
| 285811 |
+
{
|
| 285812 |
+
"epoch": 1181.59,
|
| 285813 |
+
"learning_rate": 7.649628432956382e-06,
|
| 285814 |
+
"loss": 0.7602,
|
| 285815 |
+
"step": 146545
|
| 285816 |
+
},
|
| 285817 |
+
{
|
| 285818 |
+
"epoch": 1181.63,
|
| 285819 |
+
"learning_rate": 7.649547657512118e-06,
|
| 285820 |
+
"loss": 0.879,
|
| 285821 |
+
"step": 146550
|
| 285822 |
+
},
|
| 285823 |
+
{
|
| 285824 |
+
"epoch": 1181.67,
|
| 285825 |
+
"learning_rate": 7.649466882067852e-06,
|
| 285826 |
+
"loss": 0.2737,
|
| 285827 |
+
"step": 146555
|
| 285828 |
+
},
|
| 285829 |
+
{
|
| 285830 |
+
"epoch": 1181.71,
|
| 285831 |
+
"learning_rate": 7.649386106623588e-06,
|
| 285832 |
+
"loss": 0.2831,
|
| 285833 |
+
"step": 146560
|
| 285834 |
+
},
|
| 285835 |
+
{
|
| 285836 |
+
"epoch": 1181.75,
|
| 285837 |
+
"learning_rate": 7.649305331179322e-06,
|
| 285838 |
+
"loss": 0.3641,
|
| 285839 |
+
"step": 146565
|
| 285840 |
+
},
|
| 285841 |
+
{
|
| 285842 |
+
"epoch": 1181.79,
|
| 285843 |
+
"learning_rate": 7.649224555735058e-06,
|
| 285844 |
+
"loss": 0.7179,
|
| 285845 |
+
"step": 146570
|
| 285846 |
+
},
|
| 285847 |
+
{
|
| 285848 |
+
"epoch": 1181.83,
|
| 285849 |
+
"learning_rate": 7.649143780290792e-06,
|
| 285850 |
+
"loss": 0.8389,
|
| 285851 |
+
"step": 146575
|
| 285852 |
+
},
|
| 285853 |
+
{
|
| 285854 |
+
"epoch": 1181.87,
|
| 285855 |
+
"learning_rate": 7.649063004846527e-06,
|
| 285856 |
+
"loss": 0.2527,
|
| 285857 |
+
"step": 146580
|
| 285858 |
+
},
|
| 285859 |
+
{
|
| 285860 |
+
"epoch": 1181.91,
|
| 285861 |
+
"learning_rate": 7.648982229402262e-06,
|
| 285862 |
+
"loss": 0.2547,
|
| 285863 |
+
"step": 146585
|
| 285864 |
+
},
|
| 285865 |
+
{
|
| 285866 |
+
"epoch": 1181.95,
|
| 285867 |
+
"learning_rate": 7.648901453957997e-06,
|
| 285868 |
+
"loss": 0.3927,
|
| 285869 |
+
"step": 146590
|
| 285870 |
+
},
|
| 285871 |
+
{
|
| 285872 |
+
"epoch": 1181.99,
|
| 285873 |
+
"learning_rate": 7.648820678513732e-06,
|
| 285874 |
+
"loss": 0.8367,
|
| 285875 |
+
"step": 146595
|
| 285876 |
+
},
|
| 285877 |
+
{
|
| 285878 |
+
"epoch": 1182.0,
|
| 285879 |
+
"eval_loss": 0.34015288949012756,
|
| 285880 |
+
"eval_runtime": 41.2224,
|
| 285881 |
+
"eval_samples_per_second": 20.304,
|
| 285882 |
+
"eval_steps_per_second": 0.655,
|
| 285883 |
+
"eval_wer": 0.1668011442822563,
|
| 285884 |
+
"step": 146596
|
| 285885 |
}
|
| 285886 |
],
|
| 285887 |
"max_steps": 620000,
|
| 285888 |
"num_train_epochs": 5000,
|
| 285889 |
+
"total_flos": 4.125294303144466e+20,
|
| 285890 |
"trial_name": null,
|
| 285891 |
"trial_params": null
|
| 285892 |
}
|
model-bin/finetune/base/{checkpoint-145602 β checkpoint-146596}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-146472 β checkpoint-147094}/config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-146472 β checkpoint-147094}/optimizer.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 722165393
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0bd46150aabe97ab3943950ceac270eaf7ba5a905ee8fc857f765d37689b06b8
|
| 3 |
size 722165393
|
model-bin/finetune/base/{checkpoint-146472 β checkpoint-147094}/preprocessor_config.json
RENAMED
|
File without changes
|
model-bin/finetune/base/{checkpoint-145602 β checkpoint-147094}/pytorch_model.bin
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 377909911
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e642ff64d498eff54009c654ed787928c5cbb4d47b0eda3d26f1a951e1f4ec32
|
| 3 |
size 377909911
|
model-bin/finetune/base/{checkpoint-145602 β checkpoint-147094}/rng_state.pth
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 14503
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b6232982667952bbc19502a4f0ef987e311dcdf6d628a958e5239d2cb0215ece
|
| 3 |
size 14503
|
model-bin/finetune/base/{checkpoint-145602 β checkpoint-147094}/scaler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 559
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c47fa02eeeae7ce0eb2caa75c3dd36cc9207f4a30770a12e539672a58d4fadb9
|
| 3 |
size 559
|
model-bin/finetune/base/{checkpoint-146472 β checkpoint-147094}/scheduler.pt
RENAMED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 623
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2e7f4fe6d1acaa457416128487b7a022a2db612c7ffc4ebe77644069b686eb9a
|
| 3 |
size 623
|
model-bin/finetune/base/{checkpoint-145602 β checkpoint-147094}/trainer_state.json
RENAMED
|
@@ -1,8 +1,8 @@
|
|
| 1 |
{
|
| 2 |
-
"best_metric": 0.
|
| 3 |
-
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-
|
| 4 |
-
"epoch":
|
| 5 |
-
"global_step":
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
@@ -284616,11 +284616,1907 @@
|
|
| 284616 |
"eval_steps_per_second": 0.687,
|
| 284617 |
"eval_wer": 0.16716790141568252,
|
| 284618 |
"step": 145602
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 284619 |
}
|
| 284620 |
],
|
| 284621 |
-
"max_steps":
|
| 284622 |
"num_train_epochs": 5000,
|
| 284623 |
-
"total_flos": 4.
|
| 284624 |
"trial_name": null,
|
| 284625 |
"trial_params": null
|
| 284626 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"best_metric": 0.1668011442822563,
|
| 3 |
+
"best_model_checkpoint": "./model-bin/finetune/base/checkpoint-146596",
|
| 4 |
+
"epoch": 1185.995983935743,
|
| 5 |
+
"global_step": 147094,
|
| 6 |
"is_hyper_param_search": false,
|
| 7 |
"is_local_process_zero": true,
|
| 8 |
"is_world_process_zero": true,
|
|
|
|
| 284616 |
"eval_steps_per_second": 0.687,
|
| 284617 |
"eval_wer": 0.16716790141568252,
|
| 284618 |
"step": 145602
|
| 284619 |
+
},
|
| 284620 |
+
{
|
| 284621 |
+
"epoch": 1164.02,
|
| 284622 |
+
"learning_rate": 7.66481421647819e-06,
|
| 284623 |
+
"loss": 0.3273,
|
| 284624 |
+
"step": 145605
|
| 284625 |
+
},
|
| 284626 |
+
{
|
| 284627 |
+
"epoch": 1164.06,
|
| 284628 |
+
"learning_rate": 7.664733441033926e-06,
|
| 284629 |
+
"loss": 0.2867,
|
| 284630 |
+
"step": 145610
|
| 284631 |
+
},
|
| 284632 |
+
{
|
| 284633 |
+
"epoch": 1164.1,
|
| 284634 |
+
"learning_rate": 7.664652665589662e-06,
|
| 284635 |
+
"loss": 0.3066,
|
| 284636 |
+
"step": 145615
|
| 284637 |
+
},
|
| 284638 |
+
{
|
| 284639 |
+
"epoch": 1164.14,
|
| 284640 |
+
"learning_rate": 7.664571890145396e-06,
|
| 284641 |
+
"loss": 0.3405,
|
| 284642 |
+
"step": 145620
|
| 284643 |
+
},
|
| 284644 |
+
{
|
| 284645 |
+
"epoch": 1164.18,
|
| 284646 |
+
"learning_rate": 7.664491114701132e-06,
|
| 284647 |
+
"loss": 0.7475,
|
| 284648 |
+
"step": 145625
|
| 284649 |
+
},
|
| 284650 |
+
{
|
| 284651 |
+
"epoch": 1164.22,
|
| 284652 |
+
"learning_rate": 7.664410339256866e-06,
|
| 284653 |
+
"loss": 0.9464,
|
| 284654 |
+
"step": 145630
|
| 284655 |
+
},
|
| 284656 |
+
{
|
| 284657 |
+
"epoch": 1164.26,
|
| 284658 |
+
"learning_rate": 7.664329563812602e-06,
|
| 284659 |
+
"loss": 0.3219,
|
| 284660 |
+
"step": 145635
|
| 284661 |
+
},
|
| 284662 |
+
{
|
| 284663 |
+
"epoch": 1164.3,
|
| 284664 |
+
"learning_rate": 7.664248788368336e-06,
|
| 284665 |
+
"loss": 0.3095,
|
| 284666 |
+
"step": 145640
|
| 284667 |
+
},
|
| 284668 |
+
{
|
| 284669 |
+
"epoch": 1164.34,
|
| 284670 |
+
"learning_rate": 7.664168012924072e-06,
|
| 284671 |
+
"loss": 0.3189,
|
| 284672 |
+
"step": 145645
|
| 284673 |
+
},
|
| 284674 |
+
{
|
| 284675 |
+
"epoch": 1164.38,
|
| 284676 |
+
"learning_rate": 7.664087237479806e-06,
|
| 284677 |
+
"loss": 0.7207,
|
| 284678 |
+
"step": 145650
|
| 284679 |
+
},
|
| 284680 |
+
{
|
| 284681 |
+
"epoch": 1164.42,
|
| 284682 |
+
"learning_rate": 7.664006462035542e-06,
|
| 284683 |
+
"loss": 0.8142,
|
| 284684 |
+
"step": 145655
|
| 284685 |
+
},
|
| 284686 |
+
{
|
| 284687 |
+
"epoch": 1164.46,
|
| 284688 |
+
"learning_rate": 7.663925686591276e-06,
|
| 284689 |
+
"loss": 0.2481,
|
| 284690 |
+
"step": 145660
|
| 284691 |
+
},
|
| 284692 |
+
{
|
| 284693 |
+
"epoch": 1164.5,
|
| 284694 |
+
"learning_rate": 7.663844911147012e-06,
|
| 284695 |
+
"loss": 0.2636,
|
| 284696 |
+
"step": 145665
|
| 284697 |
+
},
|
| 284698 |
+
{
|
| 284699 |
+
"epoch": 1164.54,
|
| 284700 |
+
"learning_rate": 7.663764135702746e-06,
|
| 284701 |
+
"loss": 0.3821,
|
| 284702 |
+
"step": 145670
|
| 284703 |
+
},
|
| 284704 |
+
{
|
| 284705 |
+
"epoch": 1164.58,
|
| 284706 |
+
"learning_rate": 7.663683360258482e-06,
|
| 284707 |
+
"loss": 0.7099,
|
| 284708 |
+
"step": 145675
|
| 284709 |
+
},
|
| 284710 |
+
{
|
| 284711 |
+
"epoch": 1164.62,
|
| 284712 |
+
"learning_rate": 7.663602584814218e-06,
|
| 284713 |
+
"loss": 0.7693,
|
| 284714 |
+
"step": 145680
|
| 284715 |
+
},
|
| 284716 |
+
{
|
| 284717 |
+
"epoch": 1164.66,
|
| 284718 |
+
"learning_rate": 7.663521809369952e-06,
|
| 284719 |
+
"loss": 0.3184,
|
| 284720 |
+
"step": 145685
|
| 284721 |
+
},
|
| 284722 |
+
{
|
| 284723 |
+
"epoch": 1164.7,
|
| 284724 |
+
"learning_rate": 7.663441033925688e-06,
|
| 284725 |
+
"loss": 0.2638,
|
| 284726 |
+
"step": 145690
|
| 284727 |
+
},
|
| 284728 |
+
{
|
| 284729 |
+
"epoch": 1164.74,
|
| 284730 |
+
"learning_rate": 7.663360258481422e-06,
|
| 284731 |
+
"loss": 0.4138,
|
| 284732 |
+
"step": 145695
|
| 284733 |
+
},
|
| 284734 |
+
{
|
| 284735 |
+
"epoch": 1164.78,
|
| 284736 |
+
"learning_rate": 7.663279483037158e-06,
|
| 284737 |
+
"loss": 0.7006,
|
| 284738 |
+
"step": 145700
|
| 284739 |
+
},
|
| 284740 |
+
{
|
| 284741 |
+
"epoch": 1164.82,
|
| 284742 |
+
"learning_rate": 7.663198707592892e-06,
|
| 284743 |
+
"loss": 0.9005,
|
| 284744 |
+
"step": 145705
|
| 284745 |
+
},
|
| 284746 |
+
{
|
| 284747 |
+
"epoch": 1164.86,
|
| 284748 |
+
"learning_rate": 7.663117932148628e-06,
|
| 284749 |
+
"loss": 0.2489,
|
| 284750 |
+
"step": 145710
|
| 284751 |
+
},
|
| 284752 |
+
{
|
| 284753 |
+
"epoch": 1164.9,
|
| 284754 |
+
"learning_rate": 7.663037156704362e-06,
|
| 284755 |
+
"loss": 0.2708,
|
| 284756 |
+
"step": 145715
|
| 284757 |
+
},
|
| 284758 |
+
{
|
| 284759 |
+
"epoch": 1164.94,
|
| 284760 |
+
"learning_rate": 7.662956381260098e-06,
|
| 284761 |
+
"loss": 0.3742,
|
| 284762 |
+
"step": 145720
|
| 284763 |
+
},
|
| 284764 |
+
{
|
| 284765 |
+
"epoch": 1164.98,
|
| 284766 |
+
"learning_rate": 7.662875605815832e-06,
|
| 284767 |
+
"loss": 0.7858,
|
| 284768 |
+
"step": 145725
|
| 284769 |
+
},
|
| 284770 |
+
{
|
| 284771 |
+
"epoch": 1165.0,
|
| 284772 |
+
"eval_loss": 0.38368138670921326,
|
| 284773 |
+
"eval_runtime": 40.6993,
|
| 284774 |
+
"eval_samples_per_second": 20.565,
|
| 284775 |
+
"eval_steps_per_second": 0.663,
|
| 284776 |
+
"eval_wer": 0.1798013485101138,
|
| 284777 |
+
"step": 145727
|
| 284778 |
+
},
|
| 284779 |
+
{
|
| 284780 |
+
"epoch": 1165.02,
|
| 284781 |
+
"learning_rate": 7.662794830371568e-06,
|
| 284782 |
+
"loss": 0.4956,
|
| 284783 |
+
"step": 145730
|
| 284784 |
+
},
|
| 284785 |
+
{
|
| 284786 |
+
"epoch": 1165.06,
|
| 284787 |
+
"learning_rate": 7.662714054927302e-06,
|
| 284788 |
+
"loss": 0.2444,
|
| 284789 |
+
"step": 145735
|
| 284790 |
+
},
|
| 284791 |
+
{
|
| 284792 |
+
"epoch": 1165.1,
|
| 284793 |
+
"learning_rate": 7.662633279483038e-06,
|
| 284794 |
+
"loss": 0.2835,
|
| 284795 |
+
"step": 145740
|
| 284796 |
+
},
|
| 284797 |
+
{
|
| 284798 |
+
"epoch": 1165.14,
|
| 284799 |
+
"learning_rate": 7.662552504038772e-06,
|
| 284800 |
+
"loss": 0.4172,
|
| 284801 |
+
"step": 145745
|
| 284802 |
+
},
|
| 284803 |
+
{
|
| 284804 |
+
"epoch": 1165.18,
|
| 284805 |
+
"learning_rate": 7.662471728594508e-06,
|
| 284806 |
+
"loss": 0.656,
|
| 284807 |
+
"step": 145750
|
| 284808 |
+
},
|
| 284809 |
+
{
|
| 284810 |
+
"epoch": 1165.22,
|
| 284811 |
+
"learning_rate": 7.662390953150244e-06,
|
| 284812 |
+
"loss": 0.8101,
|
| 284813 |
+
"step": 145755
|
| 284814 |
+
},
|
| 284815 |
+
{
|
| 284816 |
+
"epoch": 1165.26,
|
| 284817 |
+
"learning_rate": 7.662310177705978e-06,
|
| 284818 |
+
"loss": 0.2991,
|
| 284819 |
+
"step": 145760
|
| 284820 |
+
},
|
| 284821 |
+
{
|
| 284822 |
+
"epoch": 1165.3,
|
| 284823 |
+
"learning_rate": 7.662229402261714e-06,
|
| 284824 |
+
"loss": 0.2851,
|
| 284825 |
+
"step": 145765
|
| 284826 |
+
},
|
| 284827 |
+
{
|
| 284828 |
+
"epoch": 1165.34,
|
| 284829 |
+
"learning_rate": 7.662148626817448e-06,
|
| 284830 |
+
"loss": 0.3243,
|
| 284831 |
+
"step": 145770
|
| 284832 |
+
},
|
| 284833 |
+
{
|
| 284834 |
+
"epoch": 1165.38,
|
| 284835 |
+
"learning_rate": 7.662067851373183e-06,
|
| 284836 |
+
"loss": 0.7754,
|
| 284837 |
+
"step": 145775
|
| 284838 |
+
},
|
| 284839 |
+
{
|
| 284840 |
+
"epoch": 1165.42,
|
| 284841 |
+
"learning_rate": 7.661987075928918e-06,
|
| 284842 |
+
"loss": 0.9141,
|
| 284843 |
+
"step": 145780
|
| 284844 |
+
},
|
| 284845 |
+
{
|
| 284846 |
+
"epoch": 1165.46,
|
| 284847 |
+
"learning_rate": 7.661906300484653e-06,
|
| 284848 |
+
"loss": 0.2886,
|
| 284849 |
+
"step": 145785
|
| 284850 |
+
},
|
| 284851 |
+
{
|
| 284852 |
+
"epoch": 1165.5,
|
| 284853 |
+
"learning_rate": 7.661825525040388e-06,
|
| 284854 |
+
"loss": 0.305,
|
| 284855 |
+
"step": 145790
|
| 284856 |
+
},
|
| 284857 |
+
{
|
| 284858 |
+
"epoch": 1165.54,
|
| 284859 |
+
"learning_rate": 7.661744749596123e-06,
|
| 284860 |
+
"loss": 0.5382,
|
| 284861 |
+
"step": 145795
|
| 284862 |
+
},
|
| 284863 |
+
{
|
| 284864 |
+
"epoch": 1165.58,
|
| 284865 |
+
"learning_rate": 7.661663974151858e-06,
|
| 284866 |
+
"loss": 0.7651,
|
| 284867 |
+
"step": 145800
|
| 284868 |
+
},
|
| 284869 |
+
{
|
| 284870 |
+
"epoch": 1165.62,
|
| 284871 |
+
"learning_rate": 7.661583198707593e-06,
|
| 284872 |
+
"loss": 0.8388,
|
| 284873 |
+
"step": 145805
|
| 284874 |
+
},
|
| 284875 |
+
{
|
| 284876 |
+
"epoch": 1165.66,
|
| 284877 |
+
"learning_rate": 7.661502423263328e-06,
|
| 284878 |
+
"loss": 0.2597,
|
| 284879 |
+
"step": 145810
|
| 284880 |
+
},
|
| 284881 |
+
{
|
| 284882 |
+
"epoch": 1165.7,
|
| 284883 |
+
"learning_rate": 7.661421647819063e-06,
|
| 284884 |
+
"loss": 0.3229,
|
| 284885 |
+
"step": 145815
|
| 284886 |
+
},
|
| 284887 |
+
{
|
| 284888 |
+
"epoch": 1165.74,
|
| 284889 |
+
"learning_rate": 7.6613408723748e-06,
|
| 284890 |
+
"loss": 0.376,
|
| 284891 |
+
"step": 145820
|
| 284892 |
+
},
|
| 284893 |
+
{
|
| 284894 |
+
"epoch": 1165.78,
|
| 284895 |
+
"learning_rate": 7.661260096930533e-06,
|
| 284896 |
+
"loss": 0.6114,
|
| 284897 |
+
"step": 145825
|
| 284898 |
+
},
|
| 284899 |
+
{
|
| 284900 |
+
"epoch": 1165.82,
|
| 284901 |
+
"learning_rate": 7.66117932148627e-06,
|
| 284902 |
+
"loss": 0.8276,
|
| 284903 |
+
"step": 145830
|
| 284904 |
+
},
|
| 284905 |
+
{
|
| 284906 |
+
"epoch": 1165.86,
|
| 284907 |
+
"learning_rate": 7.661098546042003e-06,
|
| 284908 |
+
"loss": 0.2959,
|
| 284909 |
+
"step": 145835
|
| 284910 |
+
},
|
| 284911 |
+
{
|
| 284912 |
+
"epoch": 1165.9,
|
| 284913 |
+
"learning_rate": 7.66101777059774e-06,
|
| 284914 |
+
"loss": 0.2697,
|
| 284915 |
+
"step": 145840
|
| 284916 |
+
},
|
| 284917 |
+
{
|
| 284918 |
+
"epoch": 1165.94,
|
| 284919 |
+
"learning_rate": 7.660936995153473e-06,
|
| 284920 |
+
"loss": 0.3293,
|
| 284921 |
+
"step": 145845
|
| 284922 |
+
},
|
| 284923 |
+
{
|
| 284924 |
+
"epoch": 1165.98,
|
| 284925 |
+
"learning_rate": 7.660856219709209e-06,
|
| 284926 |
+
"loss": 0.7433,
|
| 284927 |
+
"step": 145850
|
| 284928 |
+
},
|
| 284929 |
+
{
|
| 284930 |
+
"epoch": 1166.0,
|
| 284931 |
+
"eval_loss": 0.33688074350357056,
|
| 284932 |
+
"eval_runtime": 41.2402,
|
| 284933 |
+
"eval_samples_per_second": 20.296,
|
| 284934 |
+
"eval_steps_per_second": 0.655,
|
| 284935 |
+
"eval_wer": 0.17012599228024178,
|
| 284936 |
+
"step": 145852
|
| 284937 |
+
},
|
| 284938 |
+
{
|
| 284939 |
+
"epoch": 1176.02,
|
| 284940 |
+
"learning_rate": 7.660775444264943e-06,
|
| 284941 |
+
"loss": 0.4127,
|
| 284942 |
+
"step": 145855
|
| 284943 |
+
},
|
| 284944 |
+
{
|
| 284945 |
+
"epoch": 1176.06,
|
| 284946 |
+
"learning_rate": 7.660694668820679e-06,
|
| 284947 |
+
"loss": 0.2818,
|
| 284948 |
+
"step": 145860
|
| 284949 |
+
},
|
| 284950 |
+
{
|
| 284951 |
+
"epoch": 1176.1,
|
| 284952 |
+
"learning_rate": 7.660613893376413e-06,
|
| 284953 |
+
"loss": 0.2966,
|
| 284954 |
+
"step": 145865
|
| 284955 |
+
},
|
| 284956 |
+
{
|
| 284957 |
+
"epoch": 1176.15,
|
| 284958 |
+
"learning_rate": 7.660533117932149e-06,
|
| 284959 |
+
"loss": 0.3262,
|
| 284960 |
+
"step": 145870
|
| 284961 |
+
},
|
| 284962 |
+
{
|
| 284963 |
+
"epoch": 1176.19,
|
| 284964 |
+
"learning_rate": 7.660452342487883e-06,
|
| 284965 |
+
"loss": 0.744,
|
| 284966 |
+
"step": 145875
|
| 284967 |
+
},
|
| 284968 |
+
{
|
| 284969 |
+
"epoch": 1176.23,
|
| 284970 |
+
"learning_rate": 7.660371567043619e-06,
|
| 284971 |
+
"loss": 0.8761,
|
| 284972 |
+
"step": 145880
|
| 284973 |
+
},
|
| 284974 |
+
{
|
| 284975 |
+
"epoch": 1176.27,
|
| 284976 |
+
"learning_rate": 7.660290791599355e-06,
|
| 284977 |
+
"loss": 0.2513,
|
| 284978 |
+
"step": 145885
|
| 284979 |
+
},
|
| 284980 |
+
{
|
| 284981 |
+
"epoch": 1176.31,
|
| 284982 |
+
"learning_rate": 7.660210016155089e-06,
|
| 284983 |
+
"loss": 0.2938,
|
| 284984 |
+
"step": 145890
|
| 284985 |
+
},
|
| 284986 |
+
{
|
| 284987 |
+
"epoch": 1176.35,
|
| 284988 |
+
"learning_rate": 7.660129240710825e-06,
|
| 284989 |
+
"loss": 0.3237,
|
| 284990 |
+
"step": 145895
|
| 284991 |
+
},
|
| 284992 |
+
{
|
| 284993 |
+
"epoch": 1176.39,
|
| 284994 |
+
"learning_rate": 7.660048465266559e-06,
|
| 284995 |
+
"loss": 0.7375,
|
| 284996 |
+
"step": 145900
|
| 284997 |
+
},
|
| 284998 |
+
{
|
| 284999 |
+
"epoch": 1176.43,
|
| 285000 |
+
"learning_rate": 7.659967689822295e-06,
|
| 285001 |
+
"loss": 0.8132,
|
| 285002 |
+
"step": 145905
|
| 285003 |
+
},
|
| 285004 |
+
{
|
| 285005 |
+
"epoch": 1176.47,
|
| 285006 |
+
"learning_rate": 7.659886914378029e-06,
|
| 285007 |
+
"loss": 0.3004,
|
| 285008 |
+
"step": 145910
|
| 285009 |
+
},
|
| 285010 |
+
{
|
| 285011 |
+
"epoch": 1176.51,
|
| 285012 |
+
"learning_rate": 7.659806138933765e-06,
|
| 285013 |
+
"loss": 0.2726,
|
| 285014 |
+
"step": 145915
|
| 285015 |
+
},
|
| 285016 |
+
{
|
| 285017 |
+
"epoch": 1176.55,
|
| 285018 |
+
"learning_rate": 7.659725363489499e-06,
|
| 285019 |
+
"loss": 0.3235,
|
| 285020 |
+
"step": 145920
|
| 285021 |
+
},
|
| 285022 |
+
{
|
| 285023 |
+
"epoch": 1176.59,
|
| 285024 |
+
"learning_rate": 7.659644588045235e-06,
|
| 285025 |
+
"loss": 0.7467,
|
| 285026 |
+
"step": 145925
|
| 285027 |
+
},
|
| 285028 |
+
{
|
| 285029 |
+
"epoch": 1176.63,
|
| 285030 |
+
"learning_rate": 7.659563812600969e-06,
|
| 285031 |
+
"loss": 0.7942,
|
| 285032 |
+
"step": 145930
|
| 285033 |
+
},
|
| 285034 |
+
{
|
| 285035 |
+
"epoch": 1176.67,
|
| 285036 |
+
"learning_rate": 7.659483037156705e-06,
|
| 285037 |
+
"loss": 0.2774,
|
| 285038 |
+
"step": 145935
|
| 285039 |
+
},
|
| 285040 |
+
{
|
| 285041 |
+
"epoch": 1176.71,
|
| 285042 |
+
"learning_rate": 7.659402261712439e-06,
|
| 285043 |
+
"loss": 0.2752,
|
| 285044 |
+
"step": 145940
|
| 285045 |
+
},
|
| 285046 |
+
{
|
| 285047 |
+
"epoch": 1176.75,
|
| 285048 |
+
"learning_rate": 7.659321486268175e-06,
|
| 285049 |
+
"loss": 0.3578,
|
| 285050 |
+
"step": 145945
|
| 285051 |
+
},
|
| 285052 |
+
{
|
| 285053 |
+
"epoch": 1176.79,
|
| 285054 |
+
"learning_rate": 7.659240710823909e-06,
|
| 285055 |
+
"loss": 0.6986,
|
| 285056 |
+
"step": 145950
|
| 285057 |
+
},
|
| 285058 |
+
{
|
| 285059 |
+
"epoch": 1176.83,
|
| 285060 |
+
"learning_rate": 7.659159935379645e-06,
|
| 285061 |
+
"loss": 0.7739,
|
| 285062 |
+
"step": 145955
|
| 285063 |
+
},
|
| 285064 |
+
{
|
| 285065 |
+
"epoch": 1176.87,
|
| 285066 |
+
"learning_rate": 7.65907915993538e-06,
|
| 285067 |
+
"loss": 0.2683,
|
| 285068 |
+
"step": 145960
|
| 285069 |
+
},
|
| 285070 |
+
{
|
| 285071 |
+
"epoch": 1176.91,
|
| 285072 |
+
"learning_rate": 7.658998384491115e-06,
|
| 285073 |
+
"loss": 0.3132,
|
| 285074 |
+
"step": 145965
|
| 285075 |
+
},
|
| 285076 |
+
{
|
| 285077 |
+
"epoch": 1176.95,
|
| 285078 |
+
"learning_rate": 7.65891760904685e-06,
|
| 285079 |
+
"loss": 0.373,
|
| 285080 |
+
"step": 145970
|
| 285081 |
+
},
|
| 285082 |
+
{
|
| 285083 |
+
"epoch": 1176.99,
|
| 285084 |
+
"learning_rate": 7.658836833602585e-06,
|
| 285085 |
+
"loss": 1.049,
|
| 285086 |
+
"step": 145975
|
| 285087 |
+
},
|
| 285088 |
+
{
|
| 285089 |
+
"epoch": 1177.0,
|
| 285090 |
+
"eval_loss": 0.3793235719203949,
|
| 285091 |
+
"eval_runtime": 40.3575,
|
| 285092 |
+
"eval_samples_per_second": 20.74,
|
| 285093 |
+
"eval_steps_per_second": 0.669,
|
| 285094 |
+
"eval_wer": 0.17174677608440797,
|
| 285095 |
+
"step": 145976
|
| 285096 |
+
},
|
| 285097 |
+
{
|
| 285098 |
+
"epoch": 1177.03,
|
| 285099 |
+
"learning_rate": 7.65875605815832e-06,
|
| 285100 |
+
"loss": 0.3211,
|
| 285101 |
+
"step": 145980
|
| 285102 |
+
},
|
| 285103 |
+
{
|
| 285104 |
+
"epoch": 1177.07,
|
| 285105 |
+
"learning_rate": 7.658675282714055e-06,
|
| 285106 |
+
"loss": 0.2587,
|
| 285107 |
+
"step": 145985
|
| 285108 |
+
},
|
| 285109 |
+
{
|
| 285110 |
+
"epoch": 1177.11,
|
| 285111 |
+
"learning_rate": 7.65859450726979e-06,
|
| 285112 |
+
"loss": 0.3043,
|
| 285113 |
+
"step": 145990
|
| 285114 |
+
},
|
| 285115 |
+
{
|
| 285116 |
+
"epoch": 1177.15,
|
| 285117 |
+
"learning_rate": 7.658513731825525e-06,
|
| 285118 |
+
"loss": 0.374,
|
| 285119 |
+
"step": 145995
|
| 285120 |
+
},
|
| 285121 |
+
{
|
| 285122 |
+
"epoch": 1177.19,
|
| 285123 |
+
"learning_rate": 7.65843295638126e-06,
|
| 285124 |
+
"loss": 0.9153,
|
| 285125 |
+
"step": 146000
|
| 285126 |
+
},
|
| 285127 |
+
{
|
| 285128 |
+
"epoch": 1177.23,
|
| 285129 |
+
"learning_rate": 7.658352180936995e-06,
|
| 285130 |
+
"loss": 0.7194,
|
| 285131 |
+
"step": 146005
|
| 285132 |
+
},
|
| 285133 |
+
{
|
| 285134 |
+
"epoch": 1177.27,
|
| 285135 |
+
"learning_rate": 7.65827140549273e-06,
|
| 285136 |
+
"loss": 0.2684,
|
| 285137 |
+
"step": 146010
|
| 285138 |
+
},
|
| 285139 |
+
{
|
| 285140 |
+
"epoch": 1177.31,
|
| 285141 |
+
"learning_rate": 7.658190630048465e-06,
|
| 285142 |
+
"loss": 0.2866,
|
| 285143 |
+
"step": 146015
|
| 285144 |
+
},
|
| 285145 |
+
{
|
| 285146 |
+
"epoch": 1177.35,
|
| 285147 |
+
"learning_rate": 7.6581098546042e-06,
|
| 285148 |
+
"loss": 0.3972,
|
| 285149 |
+
"step": 146020
|
| 285150 |
+
},
|
| 285151 |
+
{
|
| 285152 |
+
"epoch": 1177.39,
|
| 285153 |
+
"learning_rate": 7.658029079159936e-06,
|
| 285154 |
+
"loss": 1.0221,
|
| 285155 |
+
"step": 146025
|
| 285156 |
+
},
|
| 285157 |
+
{
|
| 285158 |
+
"epoch": 1177.43,
|
| 285159 |
+
"learning_rate": 7.65794830371567e-06,
|
| 285160 |
+
"loss": 0.5779,
|
| 285161 |
+
"step": 146030
|
| 285162 |
+
},
|
| 285163 |
+
{
|
| 285164 |
+
"epoch": 1177.47,
|
| 285165 |
+
"learning_rate": 7.657867528271406e-06,
|
| 285166 |
+
"loss": 0.3161,
|
| 285167 |
+
"step": 146035
|
| 285168 |
+
},
|
| 285169 |
+
{
|
| 285170 |
+
"epoch": 1177.51,
|
| 285171 |
+
"learning_rate": 7.65778675282714e-06,
|
| 285172 |
+
"loss": 0.2923,
|
| 285173 |
+
"step": 146040
|
| 285174 |
+
},
|
| 285175 |
+
{
|
| 285176 |
+
"epoch": 1177.55,
|
| 285177 |
+
"learning_rate": 7.657705977382876e-06,
|
| 285178 |
+
"loss": 0.4056,
|
| 285179 |
+
"step": 146045
|
| 285180 |
+
},
|
| 285181 |
+
{
|
| 285182 |
+
"epoch": 1177.59,
|
| 285183 |
+
"learning_rate": 7.65762520193861e-06,
|
| 285184 |
+
"loss": 0.8064,
|
| 285185 |
+
"step": 146050
|
| 285186 |
+
},
|
| 285187 |
+
{
|
| 285188 |
+
"epoch": 1177.63,
|
| 285189 |
+
"learning_rate": 7.657544426494346e-06,
|
| 285190 |
+
"loss": 0.6026,
|
| 285191 |
+
"step": 146055
|
| 285192 |
+
},
|
| 285193 |
+
{
|
| 285194 |
+
"epoch": 1177.67,
|
| 285195 |
+
"learning_rate": 7.65746365105008e-06,
|
| 285196 |
+
"loss": 0.2731,
|
| 285197 |
+
"step": 146060
|
| 285198 |
+
},
|
| 285199 |
+
{
|
| 285200 |
+
"epoch": 1177.71,
|
| 285201 |
+
"learning_rate": 7.657382875605816e-06,
|
| 285202 |
+
"loss": 0.2761,
|
| 285203 |
+
"step": 146065
|
| 285204 |
+
},
|
| 285205 |
+
{
|
| 285206 |
+
"epoch": 1177.76,
|
| 285207 |
+
"learning_rate": 7.65730210016155e-06,
|
| 285208 |
+
"loss": 0.3563,
|
| 285209 |
+
"step": 146070
|
| 285210 |
+
},
|
| 285211 |
+
{
|
| 285212 |
+
"epoch": 1177.8,
|
| 285213 |
+
"learning_rate": 7.657221324717286e-06,
|
| 285214 |
+
"loss": 0.8102,
|
| 285215 |
+
"step": 146075
|
| 285216 |
+
},
|
| 285217 |
+
{
|
| 285218 |
+
"epoch": 1177.84,
|
| 285219 |
+
"learning_rate": 7.65714054927302e-06,
|
| 285220 |
+
"loss": 0.5813,
|
| 285221 |
+
"step": 146080
|
| 285222 |
+
},
|
| 285223 |
+
{
|
| 285224 |
+
"epoch": 1177.88,
|
| 285225 |
+
"learning_rate": 7.657059773828756e-06,
|
| 285226 |
+
"loss": 0.3342,
|
| 285227 |
+
"step": 146085
|
| 285228 |
+
},
|
| 285229 |
+
{
|
| 285230 |
+
"epoch": 1177.92,
|
| 285231 |
+
"learning_rate": 7.656978998384492e-06,
|
| 285232 |
+
"loss": 0.3355,
|
| 285233 |
+
"step": 146090
|
| 285234 |
+
},
|
| 285235 |
+
{
|
| 285236 |
+
"epoch": 1177.96,
|
| 285237 |
+
"learning_rate": 7.656898222940226e-06,
|
| 285238 |
+
"loss": 0.4923,
|
| 285239 |
+
"step": 146095
|
| 285240 |
+
},
|
| 285241 |
+
{
|
| 285242 |
+
"epoch": 1178.0,
|
| 285243 |
+
"learning_rate": 7.656817447495962e-06,
|
| 285244 |
+
"loss": 1.2601,
|
| 285245 |
+
"step": 146100
|
| 285246 |
+
},
|
| 285247 |
+
{
|
| 285248 |
+
"epoch": 1178.0,
|
| 285249 |
+
"eval_loss": 0.4012053608894348,
|
| 285250 |
+
"eval_runtime": 40.577,
|
| 285251 |
+
"eval_samples_per_second": 20.627,
|
| 285252 |
+
"eval_steps_per_second": 0.665,
|
| 285253 |
+
"eval_wer": 0.1786023678891135,
|
| 285254 |
+
"step": 146100
|
| 285255 |
+
},
|
| 285256 |
+
{
|
| 285257 |
+
"epoch": 1178.04,
|
| 285258 |
+
"learning_rate": 7.656736672051696e-06,
|
| 285259 |
+
"loss": 0.2932,
|
| 285260 |
+
"step": 146105
|
| 285261 |
+
},
|
| 285262 |
+
{
|
| 285263 |
+
"epoch": 1178.08,
|
| 285264 |
+
"learning_rate": 7.656655896607432e-06,
|
| 285265 |
+
"loss": 0.2603,
|
| 285266 |
+
"step": 146110
|
| 285267 |
+
},
|
| 285268 |
+
{
|
| 285269 |
+
"epoch": 1178.12,
|
| 285270 |
+
"learning_rate": 7.656575121163166e-06,
|
| 285271 |
+
"loss": 0.2993,
|
| 285272 |
+
"step": 146115
|
| 285273 |
+
},
|
| 285274 |
+
{
|
| 285275 |
+
"epoch": 1178.16,
|
| 285276 |
+
"learning_rate": 7.656494345718902e-06,
|
| 285277 |
+
"loss": 0.4776,
|
| 285278 |
+
"step": 146120
|
| 285279 |
+
},
|
| 285280 |
+
{
|
| 285281 |
+
"epoch": 1178.2,
|
| 285282 |
+
"learning_rate": 7.656413570274636e-06,
|
| 285283 |
+
"loss": 1.1354,
|
| 285284 |
+
"step": 146125
|
| 285285 |
+
},
|
| 285286 |
+
{
|
| 285287 |
+
"epoch": 1178.24,
|
| 285288 |
+
"learning_rate": 7.656332794830372e-06,
|
| 285289 |
+
"loss": 0.2897,
|
| 285290 |
+
"step": 146130
|
| 285291 |
+
},
|
| 285292 |
+
{
|
| 285293 |
+
"epoch": 1178.28,
|
| 285294 |
+
"learning_rate": 7.656252019386106e-06,
|
| 285295 |
+
"loss": 0.2256,
|
| 285296 |
+
"step": 146135
|
| 285297 |
+
},
|
| 285298 |
+
{
|
| 285299 |
+
"epoch": 1178.32,
|
| 285300 |
+
"learning_rate": 7.656171243941842e-06,
|
| 285301 |
+
"loss": 0.2845,
|
| 285302 |
+
"step": 146140
|
| 285303 |
+
},
|
| 285304 |
+
{
|
| 285305 |
+
"epoch": 1178.36,
|
| 285306 |
+
"learning_rate": 7.656090468497578e-06,
|
| 285307 |
+
"loss": 0.4063,
|
| 285308 |
+
"step": 146145
|
| 285309 |
+
},
|
| 285310 |
+
{
|
| 285311 |
+
"epoch": 1178.4,
|
| 285312 |
+
"learning_rate": 7.656009693053312e-06,
|
| 285313 |
+
"loss": 1.1799,
|
| 285314 |
+
"step": 146150
|
| 285315 |
+
},
|
| 285316 |
+
{
|
| 285317 |
+
"epoch": 1178.44,
|
| 285318 |
+
"learning_rate": 7.655928917609048e-06,
|
| 285319 |
+
"loss": 0.3141,
|
| 285320 |
+
"step": 146155
|
| 285321 |
+
},
|
| 285322 |
+
{
|
| 285323 |
+
"epoch": 1178.48,
|
| 285324 |
+
"learning_rate": 7.655848142164782e-06,
|
| 285325 |
+
"loss": 0.2586,
|
| 285326 |
+
"step": 146160
|
| 285327 |
+
},
|
| 285328 |
+
{
|
| 285329 |
+
"epoch": 1178.52,
|
| 285330 |
+
"learning_rate": 7.655767366720518e-06,
|
| 285331 |
+
"loss": 0.3502,
|
| 285332 |
+
"step": 146165
|
| 285333 |
+
},
|
| 285334 |
+
{
|
| 285335 |
+
"epoch": 1178.56,
|
| 285336 |
+
"learning_rate": 7.655686591276252e-06,
|
| 285337 |
+
"loss": 0.4228,
|
| 285338 |
+
"step": 146170
|
| 285339 |
+
},
|
| 285340 |
+
{
|
| 285341 |
+
"epoch": 1178.6,
|
| 285342 |
+
"learning_rate": 7.655605815831988e-06,
|
| 285343 |
+
"loss": 1.2601,
|
| 285344 |
+
"step": 146175
|
| 285345 |
+
},
|
| 285346 |
+
{
|
| 285347 |
+
"epoch": 1178.64,
|
| 285348 |
+
"learning_rate": 7.655525040387722e-06,
|
| 285349 |
+
"loss": 0.2679,
|
| 285350 |
+
"step": 146180
|
| 285351 |
+
},
|
| 285352 |
+
{
|
| 285353 |
+
"epoch": 1178.68,
|
| 285354 |
+
"learning_rate": 7.655444264943458e-06,
|
| 285355 |
+
"loss": 0.2357,
|
| 285356 |
+
"step": 146185
|
| 285357 |
+
},
|
| 285358 |
+
{
|
| 285359 |
+
"epoch": 1178.72,
|
| 285360 |
+
"learning_rate": 7.655363489499192e-06,
|
| 285361 |
+
"loss": 0.3045,
|
| 285362 |
+
"step": 146190
|
| 285363 |
+
},
|
| 285364 |
+
{
|
| 285365 |
+
"epoch": 1178.76,
|
| 285366 |
+
"learning_rate": 7.655282714054928e-06,
|
| 285367 |
+
"loss": 0.5178,
|
| 285368 |
+
"step": 146195
|
| 285369 |
+
},
|
| 285370 |
+
{
|
| 285371 |
+
"epoch": 1178.8,
|
| 285372 |
+
"learning_rate": 7.655201938610663e-06,
|
| 285373 |
+
"loss": 1.2614,
|
| 285374 |
+
"step": 146200
|
| 285375 |
+
},
|
| 285376 |
+
{
|
| 285377 |
+
"epoch": 1178.84,
|
| 285378 |
+
"learning_rate": 7.655121163166397e-06,
|
| 285379 |
+
"loss": 0.3759,
|
| 285380 |
+
"step": 146205
|
| 285381 |
+
},
|
| 285382 |
+
{
|
| 285383 |
+
"epoch": 1178.88,
|
| 285384 |
+
"learning_rate": 7.655040387722133e-06,
|
| 285385 |
+
"loss": 0.2394,
|
| 285386 |
+
"step": 146210
|
| 285387 |
+
},
|
| 285388 |
+
{
|
| 285389 |
+
"epoch": 1178.92,
|
| 285390 |
+
"learning_rate": 7.654959612277867e-06,
|
| 285391 |
+
"loss": 0.275,
|
| 285392 |
+
"step": 146215
|
| 285393 |
+
},
|
| 285394 |
+
{
|
| 285395 |
+
"epoch": 1178.96,
|
| 285396 |
+
"learning_rate": 7.654878836833603e-06,
|
| 285397 |
+
"loss": 0.5791,
|
| 285398 |
+
"step": 146220
|
| 285399 |
+
},
|
| 285400 |
+
{
|
| 285401 |
+
"epoch": 1179.0,
|
| 285402 |
+
"eval_loss": 0.358010470867157,
|
| 285403 |
+
"eval_runtime": 40.5201,
|
| 285404 |
+
"eval_samples_per_second": 20.656,
|
| 285405 |
+
"eval_steps_per_second": 0.666,
|
| 285406 |
+
"eval_wer": 0.176530911221799,
|
| 285407 |
+
"step": 146224
|
| 285408 |
+
},
|
| 285409 |
+
{
|
| 285410 |
+
"epoch": 1179.01,
|
| 285411 |
+
"learning_rate": 7.654798061389337e-06,
|
| 285412 |
+
"loss": 0.5013,
|
| 285413 |
+
"step": 146225
|
| 285414 |
+
},
|
| 285415 |
+
{
|
| 285416 |
+
"epoch": 1179.05,
|
| 285417 |
+
"learning_rate": 7.654717285945073e-06,
|
| 285418 |
+
"loss": 0.2654,
|
| 285419 |
+
"step": 146230
|
| 285420 |
+
},
|
| 285421 |
+
{
|
| 285422 |
+
"epoch": 1179.09,
|
| 285423 |
+
"learning_rate": 7.654636510500807e-06,
|
| 285424 |
+
"loss": 0.2647,
|
| 285425 |
+
"step": 146235
|
| 285426 |
+
},
|
| 285427 |
+
{
|
| 285428 |
+
"epoch": 1179.13,
|
| 285429 |
+
"learning_rate": 7.654555735056543e-06,
|
| 285430 |
+
"loss": 0.3248,
|
| 285431 |
+
"step": 146240
|
| 285432 |
+
},
|
| 285433 |
+
{
|
| 285434 |
+
"epoch": 1179.17,
|
| 285435 |
+
"learning_rate": 7.654474959612277e-06,
|
| 285436 |
+
"loss": 0.4694,
|
| 285437 |
+
"step": 146245
|
| 285438 |
+
},
|
| 285439 |
+
{
|
| 285440 |
+
"epoch": 1179.21,
|
| 285441 |
+
"learning_rate": 7.654394184168013e-06,
|
| 285442 |
+
"loss": 1.0148,
|
| 285443 |
+
"step": 146250
|
| 285444 |
+
},
|
| 285445 |
+
{
|
| 285446 |
+
"epoch": 1179.25,
|
| 285447 |
+
"learning_rate": 7.654313408723747e-06,
|
| 285448 |
+
"loss": 0.2833,
|
| 285449 |
+
"step": 146255
|
| 285450 |
+
},
|
| 285451 |
+
{
|
| 285452 |
+
"epoch": 1179.29,
|
| 285453 |
+
"learning_rate": 7.654232633279483e-06,
|
| 285454 |
+
"loss": 0.2983,
|
| 285455 |
+
"step": 146260
|
| 285456 |
+
},
|
| 285457 |
+
{
|
| 285458 |
+
"epoch": 1179.33,
|
| 285459 |
+
"learning_rate": 7.654151857835219e-06,
|
| 285460 |
+
"loss": 0.3522,
|
| 285461 |
+
"step": 146265
|
| 285462 |
+
},
|
| 285463 |
+
{
|
| 285464 |
+
"epoch": 1179.37,
|
| 285465 |
+
"learning_rate": 7.654071082390953e-06,
|
| 285466 |
+
"loss": 0.5271,
|
| 285467 |
+
"step": 146270
|
| 285468 |
+
},
|
| 285469 |
+
{
|
| 285470 |
+
"epoch": 1179.41,
|
| 285471 |
+
"learning_rate": 7.653990306946689e-06,
|
| 285472 |
+
"loss": 1.1747,
|
| 285473 |
+
"step": 146275
|
| 285474 |
+
},
|
| 285475 |
+
{
|
| 285476 |
+
"epoch": 1179.45,
|
| 285477 |
+
"learning_rate": 7.653909531502423e-06,
|
| 285478 |
+
"loss": 0.2796,
|
| 285479 |
+
"step": 146280
|
| 285480 |
+
},
|
| 285481 |
+
{
|
| 285482 |
+
"epoch": 1179.49,
|
| 285483 |
+
"learning_rate": 7.653828756058159e-06,
|
| 285484 |
+
"loss": 0.2518,
|
| 285485 |
+
"step": 146285
|
| 285486 |
+
},
|
| 285487 |
+
{
|
| 285488 |
+
"epoch": 1179.53,
|
| 285489 |
+
"learning_rate": 7.653747980613893e-06,
|
| 285490 |
+
"loss": 0.307,
|
| 285491 |
+
"step": 146290
|
| 285492 |
+
},
|
| 285493 |
+
{
|
| 285494 |
+
"epoch": 1179.57,
|
| 285495 |
+
"learning_rate": 7.653667205169629e-06,
|
| 285496 |
+
"loss": 0.4902,
|
| 285497 |
+
"step": 146295
|
| 285498 |
+
},
|
| 285499 |
+
{
|
| 285500 |
+
"epoch": 1179.61,
|
| 285501 |
+
"learning_rate": 7.653586429725363e-06,
|
| 285502 |
+
"loss": 1.0019,
|
| 285503 |
+
"step": 146300
|
| 285504 |
+
},
|
| 285505 |
+
{
|
| 285506 |
+
"epoch": 1179.65,
|
| 285507 |
+
"learning_rate": 7.653505654281099e-06,
|
| 285508 |
+
"loss": 0.278,
|
| 285509 |
+
"step": 146305
|
| 285510 |
+
},
|
| 285511 |
+
{
|
| 285512 |
+
"epoch": 1179.69,
|
| 285513 |
+
"learning_rate": 7.653424878836833e-06,
|
| 285514 |
+
"loss": 0.3384,
|
| 285515 |
+
"step": 146310
|
| 285516 |
+
},
|
| 285517 |
+
{
|
| 285518 |
+
"epoch": 1179.73,
|
| 285519 |
+
"learning_rate": 7.653344103392569e-06,
|
| 285520 |
+
"loss": 0.3353,
|
| 285521 |
+
"step": 146315
|
| 285522 |
+
},
|
| 285523 |
+
{
|
| 285524 |
+
"epoch": 1179.77,
|
| 285525 |
+
"learning_rate": 7.653263327948305e-06,
|
| 285526 |
+
"loss": 0.5559,
|
| 285527 |
+
"step": 146320
|
| 285528 |
+
},
|
| 285529 |
+
{
|
| 285530 |
+
"epoch": 1179.81,
|
| 285531 |
+
"learning_rate": 7.653182552504039e-06,
|
| 285532 |
+
"loss": 1.4385,
|
| 285533 |
+
"step": 146325
|
| 285534 |
+
},
|
| 285535 |
+
{
|
| 285536 |
+
"epoch": 1179.85,
|
| 285537 |
+
"learning_rate": 7.653101777059775e-06,
|
| 285538 |
+
"loss": 0.2435,
|
| 285539 |
+
"step": 146330
|
| 285540 |
+
},
|
| 285541 |
+
{
|
| 285542 |
+
"epoch": 1179.9,
|
| 285543 |
+
"learning_rate": 7.653021001615509e-06,
|
| 285544 |
+
"loss": 0.2579,
|
| 285545 |
+
"step": 146335
|
| 285546 |
+
},
|
| 285547 |
+
{
|
| 285548 |
+
"epoch": 1179.94,
|
| 285549 |
+
"learning_rate": 7.652940226171245e-06,
|
| 285550 |
+
"loss": 0.345,
|
| 285551 |
+
"step": 146340
|
| 285552 |
+
},
|
| 285553 |
+
{
|
| 285554 |
+
"epoch": 1179.98,
|
| 285555 |
+
"learning_rate": 7.652859450726979e-06,
|
| 285556 |
+
"loss": 0.5621,
|
| 285557 |
+
"step": 146345
|
| 285558 |
+
},
|
| 285559 |
+
{
|
| 285560 |
+
"epoch": 1180.0,
|
| 285561 |
+
"eval_loss": 0.449453204870224,
|
| 285562 |
+
"eval_runtime": 40.0278,
|
| 285563 |
+
"eval_samples_per_second": 20.91,
|
| 285564 |
+
"eval_steps_per_second": 0.675,
|
| 285565 |
+
"eval_wer": 0.18728573557907288,
|
| 285566 |
+
"step": 146348
|
| 285567 |
+
},
|
| 285568 |
+
{
|
| 285569 |
+
"epoch": 1180.02,
|
| 285570 |
+
"learning_rate": 7.652778675282715e-06,
|
| 285571 |
+
"loss": 0.2882,
|
| 285572 |
+
"step": 146350
|
| 285573 |
+
},
|
| 285574 |
+
{
|
| 285575 |
+
"epoch": 1180.06,
|
| 285576 |
+
"learning_rate": 7.652697899838449e-06,
|
| 285577 |
+
"loss": 0.2908,
|
| 285578 |
+
"step": 146355
|
| 285579 |
+
},
|
| 285580 |
+
{
|
| 285581 |
+
"epoch": 1180.1,
|
| 285582 |
+
"learning_rate": 7.652617124394185e-06,
|
| 285583 |
+
"loss": 0.2317,
|
| 285584 |
+
"step": 146360
|
| 285585 |
+
},
|
| 285586 |
+
{
|
| 285587 |
+
"epoch": 1180.14,
|
| 285588 |
+
"learning_rate": 7.652536348949919e-06,
|
| 285589 |
+
"loss": 0.3416,
|
| 285590 |
+
"step": 146365
|
| 285591 |
+
},
|
| 285592 |
+
{
|
| 285593 |
+
"epoch": 1180.18,
|
| 285594 |
+
"learning_rate": 7.652455573505655e-06,
|
| 285595 |
+
"loss": 0.6812,
|
| 285596 |
+
"step": 146370
|
| 285597 |
+
},
|
| 285598 |
+
{
|
| 285599 |
+
"epoch": 1180.22,
|
| 285600 |
+
"learning_rate": 7.65237479806139e-06,
|
| 285601 |
+
"loss": 0.8339,
|
| 285602 |
+
"step": 146375
|
| 285603 |
+
},
|
| 285604 |
+
{
|
| 285605 |
+
"epoch": 1180.26,
|
| 285606 |
+
"learning_rate": 7.652294022617125e-06,
|
| 285607 |
+
"loss": 0.431,
|
| 285608 |
+
"step": 146380
|
| 285609 |
+
},
|
| 285610 |
+
{
|
| 285611 |
+
"epoch": 1180.3,
|
| 285612 |
+
"learning_rate": 7.65221324717286e-06,
|
| 285613 |
+
"loss": 0.3809,
|
| 285614 |
+
"step": 146385
|
| 285615 |
+
},
|
| 285616 |
+
{
|
| 285617 |
+
"epoch": 1180.34,
|
| 285618 |
+
"learning_rate": 7.652132471728595e-06,
|
| 285619 |
+
"loss": 0.3004,
|
| 285620 |
+
"step": 146390
|
| 285621 |
+
},
|
| 285622 |
+
{
|
| 285623 |
+
"epoch": 1180.38,
|
| 285624 |
+
"learning_rate": 7.65205169628433e-06,
|
| 285625 |
+
"loss": 0.578,
|
| 285626 |
+
"step": 146395
|
| 285627 |
+
},
|
| 285628 |
+
{
|
| 285629 |
+
"epoch": 1180.42,
|
| 285630 |
+
"learning_rate": 7.651970920840065e-06,
|
| 285631 |
+
"loss": 0.9797,
|
| 285632 |
+
"step": 146400
|
| 285633 |
+
},
|
| 285634 |
+
{
|
| 285635 |
+
"epoch": 1180.46,
|
| 285636 |
+
"learning_rate": 7.6518901453958e-06,
|
| 285637 |
+
"loss": 0.2545,
|
| 285638 |
+
"step": 146405
|
| 285639 |
+
},
|
| 285640 |
+
{
|
| 285641 |
+
"epoch": 1180.5,
|
| 285642 |
+
"learning_rate": 7.651809369951535e-06,
|
| 285643 |
+
"loss": 0.2621,
|
| 285644 |
+
"step": 146410
|
| 285645 |
+
},
|
| 285646 |
+
{
|
| 285647 |
+
"epoch": 1180.54,
|
| 285648 |
+
"learning_rate": 7.65172859450727e-06,
|
| 285649 |
+
"loss": 0.3193,
|
| 285650 |
+
"step": 146415
|
| 285651 |
+
},
|
| 285652 |
+
{
|
| 285653 |
+
"epoch": 1180.58,
|
| 285654 |
+
"learning_rate": 7.651647819063005e-06,
|
| 285655 |
+
"loss": 0.6539,
|
| 285656 |
+
"step": 146420
|
| 285657 |
+
},
|
| 285658 |
+
{
|
| 285659 |
+
"epoch": 1180.62,
|
| 285660 |
+
"learning_rate": 7.65156704361874e-06,
|
| 285661 |
+
"loss": 1.0436,
|
| 285662 |
+
"step": 146425
|
| 285663 |
+
},
|
| 285664 |
+
{
|
| 285665 |
+
"epoch": 1180.66,
|
| 285666 |
+
"learning_rate": 7.651486268174474e-06,
|
| 285667 |
+
"loss": 0.2563,
|
| 285668 |
+
"step": 146430
|
| 285669 |
+
},
|
| 285670 |
+
{
|
| 285671 |
+
"epoch": 1180.7,
|
| 285672 |
+
"learning_rate": 7.65140549273021e-06,
|
| 285673 |
+
"loss": 0.3142,
|
| 285674 |
+
"step": 146435
|
| 285675 |
+
},
|
| 285676 |
+
{
|
| 285677 |
+
"epoch": 1180.74,
|
| 285678 |
+
"learning_rate": 7.651324717285946e-06,
|
| 285679 |
+
"loss": 0.3735,
|
| 285680 |
+
"step": 146440
|
| 285681 |
+
},
|
| 285682 |
+
{
|
| 285683 |
+
"epoch": 1180.78,
|
| 285684 |
+
"learning_rate": 7.65124394184168e-06,
|
| 285685 |
+
"loss": 0.6211,
|
| 285686 |
+
"step": 146445
|
| 285687 |
+
},
|
| 285688 |
+
{
|
| 285689 |
+
"epoch": 1180.82,
|
| 285690 |
+
"learning_rate": 7.651163166397416e-06,
|
| 285691 |
+
"loss": 1.0659,
|
| 285692 |
+
"step": 146450
|
| 285693 |
+
},
|
| 285694 |
+
{
|
| 285695 |
+
"epoch": 1180.86,
|
| 285696 |
+
"learning_rate": 7.65108239095315e-06,
|
| 285697 |
+
"loss": 0.3007,
|
| 285698 |
+
"step": 146455
|
| 285699 |
+
},
|
| 285700 |
+
{
|
| 285701 |
+
"epoch": 1180.9,
|
| 285702 |
+
"learning_rate": 7.651001615508886e-06,
|
| 285703 |
+
"loss": 0.2761,
|
| 285704 |
+
"step": 146460
|
| 285705 |
+
},
|
| 285706 |
+
{
|
| 285707 |
+
"epoch": 1180.94,
|
| 285708 |
+
"learning_rate": 7.65092084006462e-06,
|
| 285709 |
+
"loss": 0.3487,
|
| 285710 |
+
"step": 146465
|
| 285711 |
+
},
|
| 285712 |
+
{
|
| 285713 |
+
"epoch": 1180.98,
|
| 285714 |
+
"learning_rate": 7.650840064620356e-06,
|
| 285715 |
+
"loss": 0.7013,
|
| 285716 |
+
"step": 146470
|
| 285717 |
+
},
|
| 285718 |
+
{
|
| 285719 |
+
"epoch": 1181.0,
|
| 285720 |
+
"eval_loss": 0.34985774755477905,
|
| 285721 |
+
"eval_runtime": 41.8541,
|
| 285722 |
+
"eval_samples_per_second": 19.998,
|
| 285723 |
+
"eval_steps_per_second": 0.645,
|
| 285724 |
+
"eval_wer": 0.1758169934640523,
|
| 285725 |
+
"step": 146472
|
| 285726 |
+
},
|
| 285727 |
+
{
|
| 285728 |
+
"epoch": 1181.02,
|
| 285729 |
+
"learning_rate": 7.65075928917609e-06,
|
| 285730 |
+
"loss": 0.306,
|
| 285731 |
+
"step": 146475
|
| 285732 |
+
},
|
| 285733 |
+
{
|
| 285734 |
+
"epoch": 1181.06,
|
| 285735 |
+
"learning_rate": 7.650678513731826e-06,
|
| 285736 |
+
"loss": 0.3149,
|
| 285737 |
+
"step": 146480
|
| 285738 |
+
},
|
| 285739 |
+
{
|
| 285740 |
+
"epoch": 1181.1,
|
| 285741 |
+
"learning_rate": 7.65059773828756e-06,
|
| 285742 |
+
"loss": 0.2999,
|
| 285743 |
+
"step": 146485
|
| 285744 |
+
},
|
| 285745 |
+
{
|
| 285746 |
+
"epoch": 1181.14,
|
| 285747 |
+
"learning_rate": 7.650516962843296e-06,
|
| 285748 |
+
"loss": 0.3347,
|
| 285749 |
+
"step": 146490
|
| 285750 |
+
},
|
| 285751 |
+
{
|
| 285752 |
+
"epoch": 1181.18,
|
| 285753 |
+
"learning_rate": 7.650436187399032e-06,
|
| 285754 |
+
"loss": 0.6613,
|
| 285755 |
+
"step": 146495
|
| 285756 |
+
},
|
| 285757 |
+
{
|
| 285758 |
+
"epoch": 1181.22,
|
| 285759 |
+
"learning_rate": 7.650355411954766e-06,
|
| 285760 |
+
"loss": 0.8353,
|
| 285761 |
+
"step": 146500
|
| 285762 |
+
},
|
| 285763 |
+
{
|
| 285764 |
+
"epoch": 1181.27,
|
| 285765 |
+
"learning_rate": 7.650274636510502e-06,
|
| 285766 |
+
"loss": 0.3035,
|
| 285767 |
+
"step": 146505
|
| 285768 |
+
},
|
| 285769 |
+
{
|
| 285770 |
+
"epoch": 1181.31,
|
| 285771 |
+
"learning_rate": 7.650193861066236e-06,
|
| 285772 |
+
"loss": 0.3257,
|
| 285773 |
+
"step": 146510
|
| 285774 |
+
},
|
| 285775 |
+
{
|
| 285776 |
+
"epoch": 1181.35,
|
| 285777 |
+
"learning_rate": 7.650113085621972e-06,
|
| 285778 |
+
"loss": 0.4288,
|
| 285779 |
+
"step": 146515
|
| 285780 |
+
},
|
| 285781 |
+
{
|
| 285782 |
+
"epoch": 1181.39,
|
| 285783 |
+
"learning_rate": 7.650032310177706e-06,
|
| 285784 |
+
"loss": 0.6413,
|
| 285785 |
+
"step": 146520
|
| 285786 |
+
},
|
| 285787 |
+
{
|
| 285788 |
+
"epoch": 1181.43,
|
| 285789 |
+
"learning_rate": 7.649951534733442e-06,
|
| 285790 |
+
"loss": 0.8146,
|
| 285791 |
+
"step": 146525
|
| 285792 |
+
},
|
| 285793 |
+
{
|
| 285794 |
+
"epoch": 1181.47,
|
| 285795 |
+
"learning_rate": 7.649870759289176e-06,
|
| 285796 |
+
"loss": 0.2767,
|
| 285797 |
+
"step": 146530
|
| 285798 |
+
},
|
| 285799 |
+
{
|
| 285800 |
+
"epoch": 1181.51,
|
| 285801 |
+
"learning_rate": 7.649789983844912e-06,
|
| 285802 |
+
"loss": 0.2579,
|
| 285803 |
+
"step": 146535
|
| 285804 |
+
},
|
| 285805 |
+
{
|
| 285806 |
+
"epoch": 1181.55,
|
| 285807 |
+
"learning_rate": 7.649709208400646e-06,
|
| 285808 |
+
"loss": 0.3754,
|
| 285809 |
+
"step": 146540
|
| 285810 |
+
},
|
| 285811 |
+
{
|
| 285812 |
+
"epoch": 1181.59,
|
| 285813 |
+
"learning_rate": 7.649628432956382e-06,
|
| 285814 |
+
"loss": 0.7602,
|
| 285815 |
+
"step": 146545
|
| 285816 |
+
},
|
| 285817 |
+
{
|
| 285818 |
+
"epoch": 1181.63,
|
| 285819 |
+
"learning_rate": 7.649547657512118e-06,
|
| 285820 |
+
"loss": 0.879,
|
| 285821 |
+
"step": 146550
|
| 285822 |
+
},
|
| 285823 |
+
{
|
| 285824 |
+
"epoch": 1181.67,
|
| 285825 |
+
"learning_rate": 7.649466882067852e-06,
|
| 285826 |
+
"loss": 0.2737,
|
| 285827 |
+
"step": 146555
|
| 285828 |
+
},
|
| 285829 |
+
{
|
| 285830 |
+
"epoch": 1181.71,
|
| 285831 |
+
"learning_rate": 7.649386106623588e-06,
|
| 285832 |
+
"loss": 0.2831,
|
| 285833 |
+
"step": 146560
|
| 285834 |
+
},
|
| 285835 |
+
{
|
| 285836 |
+
"epoch": 1181.75,
|
| 285837 |
+
"learning_rate": 7.649305331179322e-06,
|
| 285838 |
+
"loss": 0.3641,
|
| 285839 |
+
"step": 146565
|
| 285840 |
+
},
|
| 285841 |
+
{
|
| 285842 |
+
"epoch": 1181.79,
|
| 285843 |
+
"learning_rate": 7.649224555735058e-06,
|
| 285844 |
+
"loss": 0.7179,
|
| 285845 |
+
"step": 146570
|
| 285846 |
+
},
|
| 285847 |
+
{
|
| 285848 |
+
"epoch": 1181.83,
|
| 285849 |
+
"learning_rate": 7.649143780290792e-06,
|
| 285850 |
+
"loss": 0.8389,
|
| 285851 |
+
"step": 146575
|
| 285852 |
+
},
|
| 285853 |
+
{
|
| 285854 |
+
"epoch": 1181.87,
|
| 285855 |
+
"learning_rate": 7.649063004846527e-06,
|
| 285856 |
+
"loss": 0.2527,
|
| 285857 |
+
"step": 146580
|
| 285858 |
+
},
|
| 285859 |
+
{
|
| 285860 |
+
"epoch": 1181.91,
|
| 285861 |
+
"learning_rate": 7.648982229402262e-06,
|
| 285862 |
+
"loss": 0.2547,
|
| 285863 |
+
"step": 146585
|
| 285864 |
+
},
|
| 285865 |
+
{
|
| 285866 |
+
"epoch": 1181.95,
|
| 285867 |
+
"learning_rate": 7.648901453957997e-06,
|
| 285868 |
+
"loss": 0.3927,
|
| 285869 |
+
"step": 146590
|
| 285870 |
+
},
|
| 285871 |
+
{
|
| 285872 |
+
"epoch": 1181.99,
|
| 285873 |
+
"learning_rate": 7.648820678513732e-06,
|
| 285874 |
+
"loss": 0.8367,
|
| 285875 |
+
"step": 146595
|
| 285876 |
+
},
|
| 285877 |
+
{
|
| 285878 |
+
"epoch": 1182.0,
|
| 285879 |
+
"eval_loss": 0.34015288949012756,
|
| 285880 |
+
"eval_runtime": 41.2224,
|
| 285881 |
+
"eval_samples_per_second": 20.304,
|
| 285882 |
+
"eval_steps_per_second": 0.655,
|
| 285883 |
+
"eval_wer": 0.1668011442822563,
|
| 285884 |
+
"step": 146596
|
| 285885 |
+
},
|
| 285886 |
+
{
|
| 285887 |
+
"epoch": 1172.03,
|
| 285888 |
+
"learning_rate": 7.648739903069467e-06,
|
| 285889 |
+
"loss": 0.3097,
|
| 285890 |
+
"step": 146600
|
| 285891 |
+
},
|
| 285892 |
+
{
|
| 285893 |
+
"epoch": 1172.07,
|
| 285894 |
+
"learning_rate": 7.648659127625202e-06,
|
| 285895 |
+
"loss": 0.2985,
|
| 285896 |
+
"step": 146605
|
| 285897 |
+
},
|
| 285898 |
+
{
|
| 285899 |
+
"epoch": 1172.11,
|
| 285900 |
+
"learning_rate": 7.648578352180937e-06,
|
| 285901 |
+
"loss": 0.2698,
|
| 285902 |
+
"step": 146610
|
| 285903 |
+
},
|
| 285904 |
+
{
|
| 285905 |
+
"epoch": 1172.15,
|
| 285906 |
+
"learning_rate": 7.648497576736673e-06,
|
| 285907 |
+
"loss": 0.3778,
|
| 285908 |
+
"step": 146615
|
| 285909 |
+
},
|
| 285910 |
+
{
|
| 285911 |
+
"epoch": 1172.19,
|
| 285912 |
+
"learning_rate": 7.648416801292407e-06,
|
| 285913 |
+
"loss": 1.0694,
|
| 285914 |
+
"step": 146620
|
| 285915 |
+
},
|
| 285916 |
+
{
|
| 285917 |
+
"epoch": 1172.23,
|
| 285918 |
+
"learning_rate": 7.648336025848143e-06,
|
| 285919 |
+
"loss": 0.5919,
|
| 285920 |
+
"step": 146625
|
| 285921 |
+
},
|
| 285922 |
+
{
|
| 285923 |
+
"epoch": 1172.27,
|
| 285924 |
+
"learning_rate": 7.648255250403877e-06,
|
| 285925 |
+
"loss": 0.2701,
|
| 285926 |
+
"step": 146630
|
| 285927 |
+
},
|
| 285928 |
+
{
|
| 285929 |
+
"epoch": 1172.31,
|
| 285930 |
+
"learning_rate": 7.648174474959613e-06,
|
| 285931 |
+
"loss": 0.2998,
|
| 285932 |
+
"step": 146635
|
| 285933 |
+
},
|
| 285934 |
+
{
|
| 285935 |
+
"epoch": 1172.35,
|
| 285936 |
+
"learning_rate": 7.648093699515347e-06,
|
| 285937 |
+
"loss": 0.3737,
|
| 285938 |
+
"step": 146640
|
| 285939 |
+
},
|
| 285940 |
+
{
|
| 285941 |
+
"epoch": 1172.39,
|
| 285942 |
+
"learning_rate": 7.648012924071083e-06,
|
| 285943 |
+
"loss": 1.0077,
|
| 285944 |
+
"step": 146645
|
| 285945 |
+
},
|
| 285946 |
+
{
|
| 285947 |
+
"epoch": 1172.43,
|
| 285948 |
+
"learning_rate": 7.647932148626817e-06,
|
| 285949 |
+
"loss": 0.6127,
|
| 285950 |
+
"step": 146650
|
| 285951 |
+
},
|
| 285952 |
+
{
|
| 285953 |
+
"epoch": 1172.47,
|
| 285954 |
+
"learning_rate": 7.647851373182553e-06,
|
| 285955 |
+
"loss": 0.3308,
|
| 285956 |
+
"step": 146655
|
| 285957 |
+
},
|
| 285958 |
+
{
|
| 285959 |
+
"epoch": 1172.51,
|
| 285960 |
+
"learning_rate": 7.647770597738287e-06,
|
| 285961 |
+
"loss": 0.312,
|
| 285962 |
+
"step": 146660
|
| 285963 |
+
},
|
| 285964 |
+
{
|
| 285965 |
+
"epoch": 1172.55,
|
| 285966 |
+
"learning_rate": 7.647689822294023e-06,
|
| 285967 |
+
"loss": 0.4248,
|
| 285968 |
+
"step": 146665
|
| 285969 |
+
},
|
| 285970 |
+
{
|
| 285971 |
+
"epoch": 1172.59,
|
| 285972 |
+
"learning_rate": 7.647609046849759e-06,
|
| 285973 |
+
"loss": 0.8503,
|
| 285974 |
+
"step": 146670
|
| 285975 |
+
},
|
| 285976 |
+
{
|
| 285977 |
+
"epoch": 1172.63,
|
| 285978 |
+
"learning_rate": 7.647528271405493e-06,
|
| 285979 |
+
"loss": 0.7404,
|
| 285980 |
+
"step": 146675
|
| 285981 |
+
},
|
| 285982 |
+
{
|
| 285983 |
+
"epoch": 1172.67,
|
| 285984 |
+
"learning_rate": 7.647447495961229e-06,
|
| 285985 |
+
"loss": 0.3356,
|
| 285986 |
+
"step": 146680
|
| 285987 |
+
},
|
| 285988 |
+
{
|
| 285989 |
+
"epoch": 1172.71,
|
| 285990 |
+
"learning_rate": 7.647366720516963e-06,
|
| 285991 |
+
"loss": 0.2776,
|
| 285992 |
+
"step": 146685
|
| 285993 |
+
},
|
| 285994 |
+
{
|
| 285995 |
+
"epoch": 1172.75,
|
| 285996 |
+
"learning_rate": 7.647285945072699e-06,
|
| 285997 |
+
"loss": 0.4439,
|
| 285998 |
+
"step": 146690
|
| 285999 |
+
},
|
| 286000 |
+
{
|
| 286001 |
+
"epoch": 1172.79,
|
| 286002 |
+
"learning_rate": 7.647205169628433e-06,
|
| 286003 |
+
"loss": 0.957,
|
| 286004 |
+
"step": 146695
|
| 286005 |
+
},
|
| 286006 |
+
{
|
| 286007 |
+
"epoch": 1172.83,
|
| 286008 |
+
"learning_rate": 7.647124394184169e-06,
|
| 286009 |
+
"loss": 0.5475,
|
| 286010 |
+
"step": 146700
|
| 286011 |
+
},
|
| 286012 |
+
{
|
| 286013 |
+
"epoch": 1172.87,
|
| 286014 |
+
"learning_rate": 7.647043618739903e-06,
|
| 286015 |
+
"loss": 0.2893,
|
| 286016 |
+
"step": 146705
|
| 286017 |
+
},
|
| 286018 |
+
{
|
| 286019 |
+
"epoch": 1172.91,
|
| 286020 |
+
"learning_rate": 7.646962843295639e-06,
|
| 286021 |
+
"loss": 0.2689,
|
| 286022 |
+
"step": 146710
|
| 286023 |
+
},
|
| 286024 |
+
{
|
| 286025 |
+
"epoch": 1172.95,
|
| 286026 |
+
"learning_rate": 7.646882067851373e-06,
|
| 286027 |
+
"loss": 0.3002,
|
| 286028 |
+
"step": 146715
|
| 286029 |
+
},
|
| 286030 |
+
{
|
| 286031 |
+
"epoch": 1172.99,
|
| 286032 |
+
"learning_rate": 7.646801292407109e-06,
|
| 286033 |
+
"loss": 0.9223,
|
| 286034 |
+
"step": 146720
|
| 286035 |
+
},
|
| 286036 |
+
{
|
| 286037 |
+
"epoch": 1173.0,
|
| 286038 |
+
"eval_loss": 0.3900996744632721,
|
| 286039 |
+
"eval_runtime": 41.2507,
|
| 286040 |
+
"eval_samples_per_second": 20.291,
|
| 286041 |
+
"eval_steps_per_second": 0.655,
|
| 286042 |
+
"eval_wer": 0.1805263539476546,
|
| 286043 |
+
"step": 146721
|
| 286044 |
+
},
|
| 286045 |
+
{
|
| 286046 |
+
"epoch": 1183.03,
|
| 286047 |
+
"learning_rate": 7.646720516962845e-06,
|
| 286048 |
+
"loss": 0.3527,
|
| 286049 |
+
"step": 146725
|
| 286050 |
+
},
|
| 286051 |
+
{
|
| 286052 |
+
"epoch": 1183.07,
|
| 286053 |
+
"learning_rate": 7.646639741518579e-06,
|
| 286054 |
+
"loss": 0.2931,
|
| 286055 |
+
"step": 146730
|
| 286056 |
+
},
|
| 286057 |
+
{
|
| 286058 |
+
"epoch": 1183.11,
|
| 286059 |
+
"learning_rate": 7.646558966074315e-06,
|
| 286060 |
+
"loss": 0.2955,
|
| 286061 |
+
"step": 146735
|
| 286062 |
+
},
|
| 286063 |
+
{
|
| 286064 |
+
"epoch": 1183.15,
|
| 286065 |
+
"learning_rate": 7.646478190630049e-06,
|
| 286066 |
+
"loss": 0.3639,
|
| 286067 |
+
"step": 146740
|
| 286068 |
+
},
|
| 286069 |
+
{
|
| 286070 |
+
"epoch": 1183.19,
|
| 286071 |
+
"learning_rate": 7.646397415185785e-06,
|
| 286072 |
+
"loss": 0.8325,
|
| 286073 |
+
"step": 146745
|
| 286074 |
+
},
|
| 286075 |
+
{
|
| 286076 |
+
"epoch": 1183.23,
|
| 286077 |
+
"learning_rate": 7.646316639741519e-06,
|
| 286078 |
+
"loss": 0.5682,
|
| 286079 |
+
"step": 146750
|
| 286080 |
+
},
|
| 286081 |
+
{
|
| 286082 |
+
"epoch": 1183.27,
|
| 286083 |
+
"learning_rate": 7.646235864297255e-06,
|
| 286084 |
+
"loss": 0.3174,
|
| 286085 |
+
"step": 146755
|
| 286086 |
+
},
|
| 286087 |
+
{
|
| 286088 |
+
"epoch": 1183.31,
|
| 286089 |
+
"learning_rate": 7.646155088852989e-06,
|
| 286090 |
+
"loss": 0.3424,
|
| 286091 |
+
"step": 146760
|
| 286092 |
+
},
|
| 286093 |
+
{
|
| 286094 |
+
"epoch": 1183.35,
|
| 286095 |
+
"learning_rate": 7.646074313408725e-06,
|
| 286096 |
+
"loss": 0.3159,
|
| 286097 |
+
"step": 146765
|
| 286098 |
+
},
|
| 286099 |
+
{
|
| 286100 |
+
"epoch": 1183.39,
|
| 286101 |
+
"learning_rate": 7.645993537964459e-06,
|
| 286102 |
+
"loss": 0.7583,
|
| 286103 |
+
"step": 146770
|
| 286104 |
+
},
|
| 286105 |
+
{
|
| 286106 |
+
"epoch": 1183.43,
|
| 286107 |
+
"learning_rate": 7.645912762520195e-06,
|
| 286108 |
+
"loss": 0.5913,
|
| 286109 |
+
"step": 146775
|
| 286110 |
+
},
|
| 286111 |
+
{
|
| 286112 |
+
"epoch": 1183.47,
|
| 286113 |
+
"learning_rate": 7.645831987075929e-06,
|
| 286114 |
+
"loss": 0.2356,
|
| 286115 |
+
"step": 146780
|
| 286116 |
+
},
|
| 286117 |
+
{
|
| 286118 |
+
"epoch": 1183.51,
|
| 286119 |
+
"learning_rate": 7.645751211631665e-06,
|
| 286120 |
+
"loss": 0.3036,
|
| 286121 |
+
"step": 146785
|
| 286122 |
+
},
|
| 286123 |
+
{
|
| 286124 |
+
"epoch": 1183.55,
|
| 286125 |
+
"learning_rate": 7.6456704361874e-06,
|
| 286126 |
+
"loss": 0.4368,
|
| 286127 |
+
"step": 146790
|
| 286128 |
+
},
|
| 286129 |
+
{
|
| 286130 |
+
"epoch": 1183.59,
|
| 286131 |
+
"learning_rate": 7.645589660743134e-06,
|
| 286132 |
+
"loss": 0.8289,
|
| 286133 |
+
"step": 146795
|
| 286134 |
+
},
|
| 286135 |
+
{
|
| 286136 |
+
"epoch": 1183.63,
|
| 286137 |
+
"learning_rate": 7.64550888529887e-06,
|
| 286138 |
+
"loss": 0.6256,
|
| 286139 |
+
"step": 146800
|
| 286140 |
+
},
|
| 286141 |
+
{
|
| 286142 |
+
"epoch": 1183.67,
|
| 286143 |
+
"learning_rate": 7.645428109854604e-06,
|
| 286144 |
+
"loss": 0.2853,
|
| 286145 |
+
"step": 146805
|
| 286146 |
+
},
|
| 286147 |
+
{
|
| 286148 |
+
"epoch": 1183.71,
|
| 286149 |
+
"learning_rate": 7.64534733441034e-06,
|
| 286150 |
+
"loss": 0.3083,
|
| 286151 |
+
"step": 146810
|
| 286152 |
+
},
|
| 286153 |
+
{
|
| 286154 |
+
"epoch": 1183.76,
|
| 286155 |
+
"learning_rate": 7.645266558966074e-06,
|
| 286156 |
+
"loss": 0.4051,
|
| 286157 |
+
"step": 146815
|
| 286158 |
+
},
|
| 286159 |
+
{
|
| 286160 |
+
"epoch": 1183.8,
|
| 286161 |
+
"learning_rate": 7.64518578352181e-06,
|
| 286162 |
+
"loss": 1.006,
|
| 286163 |
+
"step": 146820
|
| 286164 |
+
},
|
| 286165 |
+
{
|
| 286166 |
+
"epoch": 1183.84,
|
| 286167 |
+
"learning_rate": 7.645105008077544e-06,
|
| 286168 |
+
"loss": 0.7374,
|
| 286169 |
+
"step": 146825
|
| 286170 |
+
},
|
| 286171 |
+
{
|
| 286172 |
+
"epoch": 1183.88,
|
| 286173 |
+
"learning_rate": 7.64502423263328e-06,
|
| 286174 |
+
"loss": 0.3358,
|
| 286175 |
+
"step": 146830
|
| 286176 |
+
},
|
| 286177 |
+
{
|
| 286178 |
+
"epoch": 1183.92,
|
| 286179 |
+
"learning_rate": 7.644943457189014e-06,
|
| 286180 |
+
"loss": 0.2254,
|
| 286181 |
+
"step": 146835
|
| 286182 |
+
},
|
| 286183 |
+
{
|
| 286184 |
+
"epoch": 1183.96,
|
| 286185 |
+
"learning_rate": 7.64486268174475e-06,
|
| 286186 |
+
"loss": 0.4519,
|
| 286187 |
+
"step": 146840
|
| 286188 |
+
},
|
| 286189 |
+
{
|
| 286190 |
+
"epoch": 1184.0,
|
| 286191 |
+
"learning_rate": 7.644781906300486e-06,
|
| 286192 |
+
"loss": 1.0456,
|
| 286193 |
+
"step": 146845
|
| 286194 |
+
},
|
| 286195 |
+
{
|
| 286196 |
+
"epoch": 1184.0,
|
| 286197 |
+
"eval_loss": 0.3610660433769226,
|
| 286198 |
+
"eval_runtime": 41.7394,
|
| 286199 |
+
"eval_samples_per_second": 20.053,
|
| 286200 |
+
"eval_steps_per_second": 0.647,
|
| 286201 |
+
"eval_wer": 0.16990750855727915,
|
| 286202 |
+
"step": 146845
|
| 286203 |
+
},
|
| 286204 |
+
{
|
| 286205 |
+
"epoch": 1174.04,
|
| 286206 |
+
"learning_rate": 7.64470113085622e-06,
|
| 286207 |
+
"loss": 0.2998,
|
| 286208 |
+
"step": 146850
|
| 286209 |
+
},
|
| 286210 |
+
{
|
| 286211 |
+
"epoch": 1174.08,
|
| 286212 |
+
"learning_rate": 7.644620355411956e-06,
|
| 286213 |
+
"loss": 0.369,
|
| 286214 |
+
"step": 146855
|
| 286215 |
+
},
|
| 286216 |
+
{
|
| 286217 |
+
"epoch": 1174.12,
|
| 286218 |
+
"learning_rate": 7.64453957996769e-06,
|
| 286219 |
+
"loss": 0.3144,
|
| 286220 |
+
"step": 146860
|
| 286221 |
+
},
|
| 286222 |
+
{
|
| 286223 |
+
"epoch": 1174.16,
|
| 286224 |
+
"learning_rate": 7.644458804523426e-06,
|
| 286225 |
+
"loss": 0.4513,
|
| 286226 |
+
"step": 146865
|
| 286227 |
+
},
|
| 286228 |
+
{
|
| 286229 |
+
"epoch": 1174.2,
|
| 286230 |
+
"learning_rate": 7.64437802907916e-06,
|
| 286231 |
+
"loss": 1.1218,
|
| 286232 |
+
"step": 146870
|
| 286233 |
+
},
|
| 286234 |
+
{
|
| 286235 |
+
"epoch": 1174.24,
|
| 286236 |
+
"learning_rate": 7.644297253634896e-06,
|
| 286237 |
+
"loss": 0.2873,
|
| 286238 |
+
"step": 146875
|
| 286239 |
+
},
|
| 286240 |
+
{
|
| 286241 |
+
"epoch": 1174.28,
|
| 286242 |
+
"learning_rate": 7.64421647819063e-06,
|
| 286243 |
+
"loss": 0.2713,
|
| 286244 |
+
"step": 146880
|
| 286245 |
+
},
|
| 286246 |
+
{
|
| 286247 |
+
"epoch": 1174.32,
|
| 286248 |
+
"learning_rate": 7.644135702746366e-06,
|
| 286249 |
+
"loss": 0.2787,
|
| 286250 |
+
"step": 146885
|
| 286251 |
+
},
|
| 286252 |
+
{
|
| 286253 |
+
"epoch": 1174.36,
|
| 286254 |
+
"learning_rate": 7.6440549273021e-06,
|
| 286255 |
+
"loss": 0.4724,
|
| 286256 |
+
"step": 146890
|
| 286257 |
+
},
|
| 286258 |
+
{
|
| 286259 |
+
"epoch": 1174.4,
|
| 286260 |
+
"learning_rate": 7.643974151857836e-06,
|
| 286261 |
+
"loss": 1.149,
|
| 286262 |
+
"step": 146895
|
| 286263 |
+
},
|
| 286264 |
+
{
|
| 286265 |
+
"epoch": 1174.44,
|
| 286266 |
+
"learning_rate": 7.643893376413572e-06,
|
| 286267 |
+
"loss": 0.3304,
|
| 286268 |
+
"step": 146900
|
| 286269 |
+
},
|
| 286270 |
+
{
|
| 286271 |
+
"epoch": 1174.48,
|
| 286272 |
+
"learning_rate": 7.643812600969306e-06,
|
| 286273 |
+
"loss": 0.2587,
|
| 286274 |
+
"step": 146905
|
| 286275 |
+
},
|
| 286276 |
+
{
|
| 286277 |
+
"epoch": 1174.52,
|
| 286278 |
+
"learning_rate": 7.643731825525042e-06,
|
| 286279 |
+
"loss": 0.2845,
|
| 286280 |
+
"step": 146910
|
| 286281 |
+
},
|
| 286282 |
+
{
|
| 286283 |
+
"epoch": 1174.56,
|
| 286284 |
+
"learning_rate": 7.643651050080776e-06,
|
| 286285 |
+
"loss": 0.4071,
|
| 286286 |
+
"step": 146915
|
| 286287 |
+
},
|
| 286288 |
+
{
|
| 286289 |
+
"epoch": 1174.6,
|
| 286290 |
+
"learning_rate": 7.643570274636512e-06,
|
| 286291 |
+
"loss": 1.1262,
|
| 286292 |
+
"step": 146920
|
| 286293 |
+
},
|
| 286294 |
+
{
|
| 286295 |
+
"epoch": 1174.64,
|
| 286296 |
+
"learning_rate": 7.643489499192246e-06,
|
| 286297 |
+
"loss": 0.3429,
|
| 286298 |
+
"step": 146925
|
| 286299 |
+
},
|
| 286300 |
+
{
|
| 286301 |
+
"epoch": 1174.68,
|
| 286302 |
+
"learning_rate": 7.643408723747982e-06,
|
| 286303 |
+
"loss": 0.3127,
|
| 286304 |
+
"step": 146930
|
| 286305 |
+
},
|
| 286306 |
+
{
|
| 286307 |
+
"epoch": 1174.72,
|
| 286308 |
+
"learning_rate": 7.643327948303716e-06,
|
| 286309 |
+
"loss": 0.2756,
|
| 286310 |
+
"step": 146935
|
| 286311 |
+
},
|
| 286312 |
+
{
|
| 286313 |
+
"epoch": 1174.76,
|
| 286314 |
+
"learning_rate": 7.643247172859452e-06,
|
| 286315 |
+
"loss": 0.5054,
|
| 286316 |
+
"step": 146940
|
| 286317 |
+
},
|
| 286318 |
+
{
|
| 286319 |
+
"epoch": 1174.8,
|
| 286320 |
+
"learning_rate": 7.643166397415186e-06,
|
| 286321 |
+
"loss": 1.0996,
|
| 286322 |
+
"step": 146945
|
| 286323 |
+
},
|
| 286324 |
+
{
|
| 286325 |
+
"epoch": 1174.84,
|
| 286326 |
+
"learning_rate": 7.643085621970922e-06,
|
| 286327 |
+
"loss": 0.3266,
|
| 286328 |
+
"step": 146950
|
| 286329 |
+
},
|
| 286330 |
+
{
|
| 286331 |
+
"epoch": 1174.88,
|
| 286332 |
+
"learning_rate": 7.643004846526656e-06,
|
| 286333 |
+
"loss": 0.2945,
|
| 286334 |
+
"step": 146955
|
| 286335 |
+
},
|
| 286336 |
+
{
|
| 286337 |
+
"epoch": 1174.92,
|
| 286338 |
+
"learning_rate": 7.642924071082392e-06,
|
| 286339 |
+
"loss": 0.3572,
|
| 286340 |
+
"step": 146960
|
| 286341 |
+
},
|
| 286342 |
+
{
|
| 286343 |
+
"epoch": 1174.96,
|
| 286344 |
+
"learning_rate": 7.642843295638127e-06,
|
| 286345 |
+
"loss": 0.4849,
|
| 286346 |
+
"step": 146965
|
| 286347 |
+
},
|
| 286348 |
+
{
|
| 286349 |
+
"epoch": 1175.0,
|
| 286350 |
+
"learning_rate": 7.642762520193862e-06,
|
| 286351 |
+
"loss": 1.3137,
|
| 286352 |
+
"step": 146970
|
| 286353 |
+
},
|
| 286354 |
+
{
|
| 286355 |
+
"epoch": 1175.0,
|
| 286356 |
+
"eval_loss": 0.3988247811794281,
|
| 286357 |
+
"eval_runtime": 41.5097,
|
| 286358 |
+
"eval_samples_per_second": 20.164,
|
| 286359 |
+
"eval_steps_per_second": 0.65,
|
| 286360 |
+
"eval_wer": 0.1729923798358734,
|
| 286361 |
+
"step": 146970
|
| 286362 |
+
},
|
| 286363 |
+
{
|
| 286364 |
+
"epoch": 1185.04,
|
| 286365 |
+
"learning_rate": 7.642681744749597e-06,
|
| 286366 |
+
"loss": 0.3428,
|
| 286367 |
+
"step": 146975
|
| 286368 |
+
},
|
| 286369 |
+
{
|
| 286370 |
+
"epoch": 1185.08,
|
| 286371 |
+
"learning_rate": 7.642600969305332e-06,
|
| 286372 |
+
"loss": 0.2385,
|
| 286373 |
+
"step": 146980
|
| 286374 |
+
},
|
| 286375 |
+
{
|
| 286376 |
+
"epoch": 1185.12,
|
| 286377 |
+
"learning_rate": 7.642520193861067e-06,
|
| 286378 |
+
"loss": 0.3427,
|
| 286379 |
+
"step": 146985
|
| 286380 |
+
},
|
| 286381 |
+
{
|
| 286382 |
+
"epoch": 1185.16,
|
| 286383 |
+
"learning_rate": 7.642439418416802e-06,
|
| 286384 |
+
"loss": 0.3936,
|
| 286385 |
+
"step": 146990
|
| 286386 |
+
},
|
| 286387 |
+
{
|
| 286388 |
+
"epoch": 1185.2,
|
| 286389 |
+
"learning_rate": 7.642358642972537e-06,
|
| 286390 |
+
"loss": 1.0638,
|
| 286391 |
+
"step": 146995
|
| 286392 |
+
},
|
| 286393 |
+
{
|
| 286394 |
+
"epoch": 1185.24,
|
| 286395 |
+
"learning_rate": 7.642277867528272e-06,
|
| 286396 |
+
"loss": 0.304,
|
| 286397 |
+
"step": 147000
|
| 286398 |
+
},
|
| 286399 |
+
{
|
| 286400 |
+
"epoch": 1185.28,
|
| 286401 |
+
"learning_rate": 7.642197092084007e-06,
|
| 286402 |
+
"loss": 0.2536,
|
| 286403 |
+
"step": 147005
|
| 286404 |
+
},
|
| 286405 |
+
{
|
| 286406 |
+
"epoch": 1185.32,
|
| 286407 |
+
"learning_rate": 7.642116316639741e-06,
|
| 286408 |
+
"loss": 0.3188,
|
| 286409 |
+
"step": 147010
|
| 286410 |
+
},
|
| 286411 |
+
{
|
| 286412 |
+
"epoch": 1185.36,
|
| 286413 |
+
"learning_rate": 7.642035541195477e-06,
|
| 286414 |
+
"loss": 0.5025,
|
| 286415 |
+
"step": 147015
|
| 286416 |
+
},
|
| 286417 |
+
{
|
| 286418 |
+
"epoch": 1185.4,
|
| 286419 |
+
"learning_rate": 7.641954765751213e-06,
|
| 286420 |
+
"loss": 1.2492,
|
| 286421 |
+
"step": 147020
|
| 286422 |
+
},
|
| 286423 |
+
{
|
| 286424 |
+
"epoch": 1185.44,
|
| 286425 |
+
"learning_rate": 7.641873990306947e-06,
|
| 286426 |
+
"loss": 0.2905,
|
| 286427 |
+
"step": 147025
|
| 286428 |
+
},
|
| 286429 |
+
{
|
| 286430 |
+
"epoch": 1185.48,
|
| 286431 |
+
"learning_rate": 7.641793214862683e-06,
|
| 286432 |
+
"loss": 0.3111,
|
| 286433 |
+
"step": 147030
|
| 286434 |
+
},
|
| 286435 |
+
{
|
| 286436 |
+
"epoch": 1185.52,
|
| 286437 |
+
"learning_rate": 7.641712439418417e-06,
|
| 286438 |
+
"loss": 0.3018,
|
| 286439 |
+
"step": 147035
|
| 286440 |
+
},
|
| 286441 |
+
{
|
| 286442 |
+
"epoch": 1185.56,
|
| 286443 |
+
"learning_rate": 7.641631663974153e-06,
|
| 286444 |
+
"loss": 0.4451,
|
| 286445 |
+
"step": 147040
|
| 286446 |
+
},
|
| 286447 |
+
{
|
| 286448 |
+
"epoch": 1185.6,
|
| 286449 |
+
"learning_rate": 7.641550888529887e-06,
|
| 286450 |
+
"loss": 1.0976,
|
| 286451 |
+
"step": 147045
|
| 286452 |
+
},
|
| 286453 |
+
{
|
| 286454 |
+
"epoch": 1185.64,
|
| 286455 |
+
"learning_rate": 7.641470113085623e-06,
|
| 286456 |
+
"loss": 0.3226,
|
| 286457 |
+
"step": 147050
|
| 286458 |
+
},
|
| 286459 |
+
{
|
| 286460 |
+
"epoch": 1185.68,
|
| 286461 |
+
"learning_rate": 7.641389337641357e-06,
|
| 286462 |
+
"loss": 0.2418,
|
| 286463 |
+
"step": 147055
|
| 286464 |
+
},
|
| 286465 |
+
{
|
| 286466 |
+
"epoch": 1185.72,
|
| 286467 |
+
"learning_rate": 7.641308562197093e-06,
|
| 286468 |
+
"loss": 0.3772,
|
| 286469 |
+
"step": 147060
|
| 286470 |
+
},
|
| 286471 |
+
{
|
| 286472 |
+
"epoch": 1185.76,
|
| 286473 |
+
"learning_rate": 7.641227786752827e-06,
|
| 286474 |
+
"loss": 0.4226,
|
| 286475 |
+
"step": 147065
|
| 286476 |
+
},
|
| 286477 |
+
{
|
| 286478 |
+
"epoch": 1185.8,
|
| 286479 |
+
"learning_rate": 7.641147011308563e-06,
|
| 286480 |
+
"loss": 1.1949,
|
| 286481 |
+
"step": 147070
|
| 286482 |
+
},
|
| 286483 |
+
{
|
| 286484 |
+
"epoch": 1185.84,
|
| 286485 |
+
"learning_rate": 7.641066235864299e-06,
|
| 286486 |
+
"loss": 0.3455,
|
| 286487 |
+
"step": 147075
|
| 286488 |
+
},
|
| 286489 |
+
{
|
| 286490 |
+
"epoch": 1185.88,
|
| 286491 |
+
"learning_rate": 7.640985460420033e-06,
|
| 286492 |
+
"loss": 0.2664,
|
| 286493 |
+
"step": 147080
|
| 286494 |
+
},
|
| 286495 |
+
{
|
| 286496 |
+
"epoch": 1185.92,
|
| 286497 |
+
"learning_rate": 7.640904684975769e-06,
|
| 286498 |
+
"loss": 0.3259,
|
| 286499 |
+
"step": 147085
|
| 286500 |
+
},
|
| 286501 |
+
{
|
| 286502 |
+
"epoch": 1185.96,
|
| 286503 |
+
"learning_rate": 7.640823909531503e-06,
|
| 286504 |
+
"loss": 0.4639,
|
| 286505 |
+
"step": 147090
|
| 286506 |
+
},
|
| 286507 |
+
{
|
| 286508 |
+
"epoch": 1186.0,
|
| 286509 |
+
"eval_loss": 0.41105982661247253,
|
| 286510 |
+
"eval_runtime": 41.11,
|
| 286511 |
+
"eval_samples_per_second": 20.36,
|
| 286512 |
+
"eval_steps_per_second": 0.657,
|
| 286513 |
+
"eval_wer": 0.1767975743574935,
|
| 286514 |
+
"step": 147094
|
| 286515 |
}
|
| 286516 |
],
|
| 286517 |
+
"max_steps": 620000,
|
| 286518 |
"num_train_epochs": 5000,
|
| 286519 |
+
"total_flos": 4.1393121623086616e+20,
|
| 286520 |
"trial_name": null,
|
| 286521 |
"trial_params": null
|
| 286522 |
}
|
model-bin/finetune/base/{checkpoint-146472 β checkpoint-147094}/training_args.bin
RENAMED
|
File without changes
|
model-bin/finetune/base/log/1630230484.4148061/events.out.tfevents.1630230484.cc93b136ebf5.1086.119
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:373e03a0065b2e921eefd867e202f84f548054ff66160ac42f034369030f3220
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630230917.0829825/events.out.tfevents.1630230917.cc93b136ebf5.1086.121
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d4941891e313e1a07773f5f32179f173fb50469156bae41861b1c5fdb30df5d6
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630231362.9249244/events.out.tfevents.1630231362.cc93b136ebf5.1086.123
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4a30f575e5d92c7236fc30cd24fb7cea6c129bed326a9687dcf47af1671d3698
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630231799.7950127/events.out.tfevents.1630231799.cc93b136ebf5.1086.125
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d581e2ac1cb2ca93023856f14c8f46437e5fbfd5ed33ec467f2dd73f4f8a4907
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/1630232234.2438745/events.out.tfevents.1630232234.cc93b136ebf5.1086.127
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b656bb8581cc3b0e8e81c57ed8bf041161d884c948fa9c079028a2323c3785ce
|
| 3 |
+
size 4194
|
model-bin/finetune/base/log/events.out.tfevents.1630230484.cc93b136ebf5.1086.118
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:21341d712d1a1d11a64108a5815e19580ec3c3b91f3adb0fcfa9f363754af853
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630230917.cc93b136ebf5.1086.120
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9207bff65ec5f514cf53a6f839848b0d88cffaba3a0673819f5b98e62a4475cd
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630231362.cc93b136ebf5.1086.122
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e143394f3a4ff3d7ad1b06893d174d817d4c9b3c2fd9b14acd48ed85157171bd
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630231799.cc93b136ebf5.1086.124
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:13d1c476968ed905c4cb56413c983b39e1d176290a9c27dfd7a277e45fbe8511
|
| 3 |
+
size 8622
|
model-bin/finetune/base/log/events.out.tfevents.1630232234.cc93b136ebf5.1086.126
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:17df0783581cbdfcc60e0903bf27fdb281f33af1809e9e2e65cabef2ab06909c
|
| 3 |
+
size 8462
|