- infer_2.py +5 -3
- model_saved/save/CKPT+2025-02-24+11-42-50+00/CKPT.yaml +3 -0
- model_saved/save/CKPT+2025-02-24+11-42-50+00/brain.ckpt +3 -0
- model_saved/save/CKPT+2025-02-24+11-42-50+00/counter.ckpt +3 -0
- model_saved/save/CKPT+2025-02-24+11-42-50+00/dataloader-TRAIN.ckpt +3 -0
- model_saved/save/CKPT+2025-02-24+11-42-50+00/dataloader-VALID.ckpt +3 -0
- model_saved/save/CKPT+2025-02-24+11-42-50+00/model.ckpt +3 -0
- model_saved/save/CKPT+2025-02-24+11-42-50+00/noam_scheduler.ckpt +3 -0
- model_saved/save/CKPT+2025-02-24+11-42-50+00/normalizer.ckpt +3 -0
- model_saved/save/CKPT+2025-02-24+11-42-50+00/optimizer.ckpt +3 -0
- model_saved/save/CKPT+2025-02-24+11-42-50+00/scaler.ckpt +3 -0
- model_saved/train_log.txt +2 -2
infer_2.py
CHANGED
|
@@ -44,7 +44,7 @@ logger = structlog.get_logger(__name__)
|
|
| 44 |
|
| 45 |
# Initialize file logger
|
| 46 |
def log_step_to_file(message):
|
| 47 |
-
with open("
|
| 48 |
f.write(message + "\n")
|
| 49 |
|
| 50 |
|
|
@@ -334,6 +334,8 @@ class ASR(sb.core.Brain):
|
|
| 334 |
log_step_to_file(
|
| 335 |
f"Epoch {epoch}, WER: {stage_stats['WER']}, CER: {stage_stats['CER']}, ACC: {stage_stats['ACC']}"
|
| 336 |
)
|
|
|
|
|
|
|
| 337 |
# log stats and save checkpoint at end-of-epoch
|
| 338 |
if stage == sb.Stage.VALID:
|
| 339 |
# report different epoch stages according current stage
|
|
@@ -534,7 +536,7 @@ class TransformerPretrainedASR(ASR):
|
|
| 534 |
f"Going to epoch counter ----- epochs limit: {epoch_counter.limit} ---- current: {epoch_counter.current}"
|
| 535 |
)
|
| 536 |
|
| 537 |
-
for epoch in
|
| 538 |
self._fit_train(
|
| 539 |
train_set=train_set,
|
| 540 |
unlabeled_train_set=unlabeled_train_set,
|
|
@@ -552,7 +554,7 @@ class TransformerPretrainedASR(ASR):
|
|
| 552 |
# break
|
| 553 |
|
| 554 |
def log_step_to_file(self, message):
|
| 555 |
-
with open("
|
| 556 |
f.write(message)
|
| 557 |
|
| 558 |
def _fit_train(self, train_set, unlabeled_train_set, epoch, enable):
|
|
|
|
| 44 |
|
| 45 |
# Initialize file logger
|
| 46 |
def log_step_to_file(message):
|
| 47 |
+
with open("training_logs_file.log", "a") as f:
|
| 48 |
f.write(message + "\n")
|
| 49 |
|
| 50 |
|
|
|
|
| 334 |
log_step_to_file(
|
| 335 |
f"Epoch {epoch}, WER: {stage_stats['WER']}, CER: {stage_stats['CER']}, ACC: {stage_stats['ACC']}"
|
| 336 |
)
|
| 337 |
+
print( f"Epoch {epoch}, WER: {stage_stats['WER']}, CER: {stage_stats['CER']}, ACC: {stage_stats['ACC']}"
|
| 338 |
+
)
|
| 339 |
# log stats and save checkpoint at end-of-epoch
|
| 340 |
if stage == sb.Stage.VALID:
|
| 341 |
# report different epoch stages according current stage
|
|
|
|
| 536 |
f"Going to epoch counter ----- epochs limit: {epoch_counter.limit} ---- current: {epoch_counter.current}"
|
| 537 |
)
|
| 538 |
|
| 539 |
+
for epoch in epoch_counter:
|
| 540 |
self._fit_train(
|
| 541 |
train_set=train_set,
|
| 542 |
unlabeled_train_set=unlabeled_train_set,
|
|
|
|
| 554 |
# break
|
| 555 |
|
| 556 |
def log_step_to_file(self, message):
|
| 557 |
+
with open("training_logs_file.log", "a") as f:
|
| 558 |
f.write(message)
|
| 559 |
|
| 560 |
def _fit_train(self, train_set, unlabeled_train_set, epoch, enable):
|
model_saved/save/CKPT+2025-02-24+11-42-50+00/CKPT.yaml
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6ca0bbb2d4d204ebc62e24b56b32f7d4b2703a2632cdb4db79aba6de1210c3e7
|
| 3 |
+
size 100
|
model_saved/save/CKPT+2025-02-24+11-42-50+00/brain.ckpt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b03e2adb0d13ef61409530b8b6f71e473c648a989c71592a604810c0acf4875a
|
| 3 |
+
size 50
|
model_saved/save/CKPT+2025-02-24+11-42-50+00/counter.ckpt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d59eced1ded07f84c145592f65bdf854358e009c5cd705f5215bf18697fed103
|
| 3 |
+
size 2
|
model_saved/save/CKPT+2025-02-24+11-42-50+00/dataloader-TRAIN.ckpt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b6bc077d6675a7c8cc9e2fa5a08c86ba59b675d69af118052bb390c3cf11e5e0
|
| 3 |
+
size 3
|
model_saved/save/CKPT+2025-02-24+11-42-50+00/dataloader-VALID.ckpt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c6bd343ae0007cdb979de7540f2668fe849d68ff47fa1a650a28f89104f41f1e
|
| 3 |
+
size 3
|
model_saved/save/CKPT+2025-02-24+11-42-50+00/model.ckpt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f441db9df1145dc55e5c5d201caf6219ee836a9c32e6ad90dbe71eeeb148c1f2
|
| 3 |
+
size 175365750
|
model_saved/save/CKPT+2025-02-24+11-42-50+00/noam_scheduler.ckpt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:293b85eeb2f81f34bfc28df784f768c987f8463c99eaeb82ac26204f2fba06dd
|
| 3 |
+
size 892
|
model_saved/save/CKPT+2025-02-24+11-42-50+00/normalizer.ckpt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5774dae5a9cf61988653664f1bc19bb58ff2ede2206edb91afa8e19dd4808e1f
|
| 3 |
+
size 2218
|
model_saved/save/CKPT+2025-02-24+11-42-50+00/optimizer.ckpt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d3f2e994b42e372f985c6d499a43a1eccca1c10b622763ada3352596652c9edd
|
| 3 |
+
size 343101283
|
model_saved/save/CKPT+2025-02-24+11-42-50+00/scaler.ckpt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9234868ab6b9fde57f75fdc7db224bd8fa6e345f6414fe29041ffb0ed531aaa4
|
| 3 |
+
size 988
|
model_saved/train_log.txt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b31282bed27c5772dbb181a6099c41850e07d080bcb84dc64c728dd9371c2959
|
| 3 |
+
size 6898
|