Training in progress, step 10
Browse files
pytorch_model.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 151098921
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:67db960796c75f5af90d6ae38074eea1cad281bdbec28d10187feba4432720bf
|
| 3 |
size 151098921
|
runs/Dec13_12-46-08_d7f040c448a8/1670935580.616109/events.out.tfevents.1670935580.d7f040c448a8.26499.1
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d0a2d84b4221c89c7f654e3b9f57b33bc79012fd62ef4c4724f3902c2ce77804
|
| 3 |
+
size 5883
|
runs/Dec13_12-46-08_d7f040c448a8/events.out.tfevents.1670935580.d7f040c448a8.26499.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c2454e09e13c3ad647aeee178d68cb6420103d17f18c2d2dd06a68fb0e289d55
|
| 3 |
+
size 4736
|
src/run_debug.sh
CHANGED
|
@@ -7,14 +7,14 @@ python src/run_speech_recognition_seq2seq_streaming.py \
|
|
| 7 |
--eval_split_name="validation" \
|
| 8 |
--model_index_name="Whisper Tiny Belarusian" \
|
| 9 |
\
|
| 10 |
-
--max_steps="
|
| 11 |
--max_eval_samples="64" \
|
| 12 |
--output_dir="./" \
|
| 13 |
--per_device_train_batch_size="32" \
|
| 14 |
--per_device_eval_batch_size="32" \
|
| 15 |
--logging_steps="10" \
|
| 16 |
-
--learning_rate="1e-
|
| 17 |
-
--warmup_steps="
|
| 18 |
--evaluation_strategy="steps" \
|
| 19 |
--eval_steps="10" \
|
| 20 |
--save_strategy="steps" \
|
|
@@ -34,6 +34,7 @@ python src/run_speech_recognition_seq2seq_streaming.py \
|
|
| 34 |
\
|
| 35 |
--do_train \
|
| 36 |
--do_eval \
|
|
|
|
| 37 |
--ignore_data_skip \
|
| 38 |
--predict_with_generate \
|
| 39 |
--do_normalize_eval \
|
|
|
|
| 7 |
--eval_split_name="validation" \
|
| 8 |
--model_index_name="Whisper Tiny Belarusian" \
|
| 9 |
\
|
| 10 |
+
--max_steps="100" \
|
| 11 |
--max_eval_samples="64" \
|
| 12 |
--output_dir="./" \
|
| 13 |
--per_device_train_batch_size="32" \
|
| 14 |
--per_device_eval_batch_size="32" \
|
| 15 |
--logging_steps="10" \
|
| 16 |
+
--learning_rate="1e-4" \
|
| 17 |
+
--warmup_steps="10" \
|
| 18 |
--evaluation_strategy="steps" \
|
| 19 |
--eval_steps="10" \
|
| 20 |
--save_strategy="steps" \
|
|
|
|
| 34 |
\
|
| 35 |
--do_train \
|
| 36 |
--do_eval \
|
| 37 |
+
--overwrite_output_dir \
|
| 38 |
--ignore_data_skip \
|
| 39 |
--predict_with_generate \
|
| 40 |
--do_normalize_eval \
|
train.log
CHANGED
|
@@ -132,3 +132,4 @@
|
|
| 132 |
eval_samples_per_second = 3.804
|
| 133 |
eval_steps_per_second = 0.119
|
| 134 |
eval_wer = 52.1978
|
|
|
|
|
|
| 132 |
eval_samples_per_second = 3.804
|
| 133 |
eval_steps_per_second = 0.119
|
| 134 |
eval_wer = 52.1978
|
| 135 |
+
{'loss': 2.5622, 'learning_rate': 7e-05, 'epoch': 0.1}
|
training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 3643
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c8a86b0841271f87ec4240b29b5ca1620f243ad5ad67d99f9919cf45237c2add
|
| 3 |
size 3643
|