Model save
Browse files- config.json +1 -1
- run.sh +1 -1
- run_speech_recognition_seq2seq_streaming.py +1 -1
- runs/Dec20_19-44-59_129-146-32-172/1671565514.8929353/events.out.tfevents.1671565514.129-146-32-172.144860.1 +3 -0
- runs/Dec20_19-44-59_129-146-32-172/events.out.tfevents.1671565514.129-146-32-172.144860.0 +3 -0
- training_args.bin +1 -1
config.json
CHANGED
|
@@ -17,7 +17,7 @@
|
|
| 17 |
"decoder_layerdrop": 0.0,
|
| 18 |
"decoder_layers": 6,
|
| 19 |
"decoder_start_token_id": 50258,
|
| 20 |
-
"dropout": 0.
|
| 21 |
"encoder_attention_heads": 8,
|
| 22 |
"encoder_ffn_dim": 2048,
|
| 23 |
"encoder_layerdrop": 0.0,
|
|
|
|
| 17 |
"decoder_layerdrop": 0.0,
|
| 18 |
"decoder_layers": 6,
|
| 19 |
"decoder_start_token_id": 50258,
|
| 20 |
+
"dropout": 0.4,
|
| 21 |
"encoder_attention_heads": 8,
|
| 22 |
"encoder_ffn_dim": 2048,
|
| 23 |
"encoder_layerdrop": 0.0,
|
run.sh
CHANGED
|
@@ -12,7 +12,7 @@ python run_speech_recognition_seq2seq_streaming.py \
|
|
| 12 |
--per_device_eval_batch_size="32" \
|
| 13 |
--gradient_accumulation_steps="1" \
|
| 14 |
--logging_steps="10" \
|
| 15 |
-
--learning_rate="1e-
|
| 16 |
--warmup_steps="30" \
|
| 17 |
--evaluation_strategy="steps" \
|
| 18 |
--eval_steps="100" \
|
|
|
|
| 12 |
--per_device_eval_batch_size="32" \
|
| 13 |
--gradient_accumulation_steps="1" \
|
| 14 |
--logging_steps="10" \
|
| 15 |
+
--learning_rate="1e-6" \
|
| 16 |
--warmup_steps="30" \
|
| 17 |
--evaluation_strategy="steps" \
|
| 18 |
--eval_steps="100" \
|
run_speech_recognition_seq2seq_streaming.py
CHANGED
|
@@ -556,7 +556,7 @@ def main():
|
|
| 556 |
elif isinstance(train_dataloader.dataset, IterableDataset):
|
| 557 |
train_dataloader.dataset.set_epoch(train_dataloader.dataset._epoch + 1)
|
| 558 |
|
| 559 |
-
model.config.dropout = 0.
|
| 560 |
# Initialize Trainer
|
| 561 |
trainer = Seq2SeqTrainer(
|
| 562 |
model=model,
|
|
|
|
| 556 |
elif isinstance(train_dataloader.dataset, IterableDataset):
|
| 557 |
train_dataloader.dataset.set_epoch(train_dataloader.dataset._epoch + 1)
|
| 558 |
|
| 559 |
+
model.config.dropout = 0.4
|
| 560 |
# Initialize Trainer
|
| 561 |
trainer = Seq2SeqTrainer(
|
| 562 |
model=model,
|
runs/Dec20_19-44-59_129-146-32-172/1671565514.8929353/events.out.tfevents.1671565514.129-146-32-172.144860.1
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0f22e60461db94cd7e811ccbfb51c5ae03a2332f25e1b78df1e1fe69f8cebc15
|
| 3 |
+
size 5871
|
runs/Dec20_19-44-59_129-146-32-172/events.out.tfevents.1671565514.129-146-32-172.144860.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:57c0376246845e980e7bac514b13415fb28a3cee9165b2e20d4533d7e0033113
|
| 3 |
+
size 4614
|
training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 3579
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:82dc9b5c0b7437cb9e4f9be835037b5a9a619b3446fb5be7b09cc8bfeaff4ff4
|
| 3 |
size 3579
|