| | import torch |
| | from utils.dataset import Speech2Text, speech_collate_fn |
| | from models.model import TransformerTransducer |
| | from tqdm import tqdm |
| | from models.loss import RNNTLoss |
| | import argparse |
| | import yaml |
| | import os |
| |
|
| | def train_one_epoch(model, dataloader, optimizer, criterion, device): |
| | model.train() |
| | total_loss = 0.0 |
| |
|
| | progress_bar = tqdm(dataloader, desc="🔁 Training", leave=False) |
| |
|
| | for batch_idx, batch in enumerate(progress_bar): |
| | speech = batch["fbank"].to(device) |
| | text = batch["text"].to(device) |
| | speech_mask = batch["fbank_mask"].to(device) |
| | text_mask = batch["text_mask"].to(device) |
| | fbank_len = batch["fbank_len"].to(device) |
| | text_len = batch["text_len"].to(device) |
| |
|
| | optimizer.zero_grad() |
| |
|
| | output, _, _ = model( |
| | speech=speech, |
| | speech_mask=speech_mask, |
| | text=text, |
| | text_mask=text_mask, |
| | ) |
| |
|
| | |
| | loss = criterion(output, text, fbank_len, text_len) |
| | loss.backward() |
| | optimizer.step() |
| |
|
| | total_loss += loss.item() |
| |
|
| | |
| | progress_bar.set_postfix(batch_loss=loss.item()) |
| |
|
| | avg_loss = total_loss / len(dataloader) |
| | print(f"✅ Average training loss: {avg_loss:.4f}") |
| | return avg_loss |
| |
|
| |
|
| | from torchaudio.functional import rnnt_loss |
| |
|
| | def evaluate(model, dataloader, criterion, device): |
| | model.eval() |
| | total_loss = 0.0 |
| |
|
| | progress_bar = tqdm(dataloader, desc="🧪 Evaluating", leave=False) |
| |
|
| | with torch.no_grad(): |
| | for batch in progress_bar: |
| | speech = batch["fbank"].to(device) |
| | text = batch["text"].to(device) |
| | speech_mask = batch["fbank_mask"].to(device) |
| | text_mask = batch["text_mask"].to(device) |
| | fbank_len = batch["fbank_len"].to(device) |
| | text_len = batch["text_len"].to(device) |
| |
|
| | output, _, _ = model( |
| | speech=speech, |
| | speech_mask=speech_mask, |
| | text=text, |
| | text_mask=text_mask, |
| | ) |
| |
|
| | loss = criterion(output, text, fbank_len, text_len) |
| | total_loss += loss.item() |
| | progress_bar.set_postfix(batch_loss=loss.item()) |
| |
|
| | avg_loss = total_loss / len(dataloader) |
| | print(f"✅ Average validation loss: {avg_loss:.4f}") |
| | return avg_loss |
| |
|
| | def load_config(config_path): |
| | with open(config_path, 'r') as f: |
| | return yaml.safe_load(f) |
| | |
| | def main(): |
| | from torch.optim import Adam |
| | |
| | parser = argparse.ArgumentParser() |
| | parser.add_argument("--config", type=str, required=True, help="Path to YAML config file") |
| | args = parser.parse_args() |
| |
|
| | config = load_config(args.config) |
| | training_cfg = config['training'] |
| | optimizer_cfg = config['optimizer'] |
| |
|
| |
|
| | |
| | train_dataset = Speech2Text( |
| | json_path=training_cfg['train_path'], |
| | vocab_path=training_cfg['vocab_path'], |
| | ) |
| |
|
| | train_loader = torch.utils.data.DataLoader( |
| | train_dataset, |
| | batch_size= training_cfg['batch_size'], |
| | shuffle=True, |
| | collate_fn = speech_collate_fn |
| | ) |
| |
|
| | dev_dataset = Speech2Text( |
| | json_path=training_cfg['dev_path'], |
| | vocab_path=training_cfg['vocab_path'] |
| | ) |
| |
|
| | dev_loader = torch.utils.data.DataLoader( |
| | dev_dataset, |
| | batch_size= training_cfg['batch_size'], |
| | shuffle=True, |
| | collate_fn = speech_collate_fn |
| | ) |
| |
|
| | device = torch.device("cuda" if torch.cuda.is_available() else "cpu") |
| | model = TransformerTransducer( |
| | in_features=config['model']['in_features'], |
| | n_classes=len(train_dataset.vocab), |
| | n_layers=config['model']['n_layers'], |
| | n_dec_layers=config['model']['n_dec_layers'], |
| | d_model=config['model']['d_model'], |
| | ff_size=config['model']['ff_size'], |
| | h=config['model']['h'], |
| | joint_size=config['model']['joint_size'], |
| | enc_left_size=config['model']['enc_left_size'], |
| | enc_right_size=config['model']['enc_right_size'], |
| | dec_left_size=config['model']['dec_left_size'], |
| | dec_right_size=config['model']['dec_right_size'], |
| | p_dropout=config['model']['p_dropout'] |
| | ).to(device) |
| |
|
| | device = torch.device("cuda" if torch.cuda.is_available() else "cpu") |
| | model.to(device) |
| |
|
| | |
| | |
| | criterion = RNNTLoss(config["rnnt_loss"]["blank"] , config["rnnt_loss"]["reduction"]) |
| |
|
| | |
| | optimizer = Adam(model.parameters(), lr=optimizer_cfg['lr']) |
| |
|
| | |
| | num_epochs = config["training"]["epochs"] |
| |
|
| | for epoch in range(1, num_epochs + 1): |
| | train_loss = train_one_epoch(model, train_loader, optimizer, criterion, device) |
| | val_loss = evaluate(model, dev_loader, criterion, device) |
| |
|
| | print(f"📘 Epoch {epoch}: Train Loss = {train_loss:.4f}, Val Loss = {val_loss:.4f}") |
| | |
| |
|
| | model_filename = os.path.join( |
| | config['training']['save_path'], |
| | f"transformer_transducer_epoch_{epoch}" |
| | ) |
| |
|
| | torch.save({ |
| | 'epoch': epoch, |
| | 'model_state_dict': model.state_dict(), |
| | 'optimizer_state_dict': optimizer.state_dict(), |
| | }, model_filename) |
| |
|
| |
|
| | if __name__ == "__main__": |
| | main() |
| |
|