| |
| |
|
|
| import os |
| import sys |
| import time |
| import traceback |
|
|
| import torch |
| from torch.utils.data import DataLoader |
| from trainer.torch import NoamLR |
| from trainer.trainer_utils import get_optimizer |
|
|
| from TTS.encoder.dataset import EncoderDataset |
| from TTS.encoder.utils.generic_utils import save_best_model, save_checkpoint, setup_encoder_model |
| from TTS.encoder.utils.training import init_training |
| from TTS.encoder.utils.visual import plot_embeddings |
| from TTS.tts.datasets import load_tts_samples |
| from TTS.utils.audio import AudioProcessor |
| from TTS.utils.generic_utils import count_parameters, remove_experiment_folder |
| from TTS.utils.io import copy_model_files |
| from TTS.utils.samplers import PerfectBatchSampler |
| from TTS.utils.training import check_update |
|
|
| torch.backends.cudnn.enabled = True |
| torch.backends.cudnn.benchmark = True |
| torch.manual_seed(54321) |
| use_cuda = torch.cuda.is_available() |
| num_gpus = torch.cuda.device_count() |
| print(" > Using CUDA: ", use_cuda) |
| print(" > Number of GPUs: ", num_gpus) |
|
|
|
|
| def setup_loader(ap: AudioProcessor, is_val: bool = False, verbose: bool = False): |
| num_utter_per_class = c.num_utter_per_class if not is_val else c.eval_num_utter_per_class |
| num_classes_in_batch = c.num_classes_in_batch if not is_val else c.eval_num_classes_in_batch |
|
|
| dataset = EncoderDataset( |
| c, |
| ap, |
| meta_data_eval if is_val else meta_data_train, |
| voice_len=c.voice_len, |
| num_utter_per_class=num_utter_per_class, |
| num_classes_in_batch=num_classes_in_batch, |
| verbose=verbose, |
| augmentation_config=c.audio_augmentation if not is_val else None, |
| use_torch_spec=c.model_params.get("use_torch_spec", False), |
| ) |
| |
| classes = dataset.get_class_list() |
|
|
| sampler = PerfectBatchSampler( |
| dataset.items, |
| classes, |
| batch_size=num_classes_in_batch * num_utter_per_class, |
| num_classes_in_batch=num_classes_in_batch, |
| num_gpus=1, |
| shuffle=not is_val, |
| drop_last=True, |
| ) |
|
|
| if len(classes) < num_classes_in_batch: |
| if is_val: |
| raise RuntimeError( |
| f"config.eval_num_classes_in_batch ({num_classes_in_batch}) need to be <= {len(classes)} (Number total of Classes in the Eval dataset) !" |
| ) |
| raise RuntimeError( |
| f"config.num_classes_in_batch ({num_classes_in_batch}) need to be <= {len(classes)} (Number total of Classes in the Train dataset) !" |
| ) |
|
|
| |
| if is_val: |
| dataset.set_classes(train_classes) |
|
|
| loader = DataLoader( |
| dataset, |
| num_workers=c.num_loader_workers, |
| batch_sampler=sampler, |
| collate_fn=dataset.collate_fn, |
| ) |
|
|
| return loader, classes, dataset.get_map_classid_to_classname() |
|
|
|
|
| def evaluation(model, criterion, data_loader, global_step): |
| eval_loss = 0 |
| for _, data in enumerate(data_loader): |
| with torch.no_grad(): |
| |
| inputs, labels = data |
|
|
| |
| labels = torch.transpose( |
| labels.view(c.eval_num_utter_per_class, c.eval_num_classes_in_batch), 0, 1 |
| ).reshape(labels.shape) |
| inputs = torch.transpose( |
| inputs.view(c.eval_num_utter_per_class, c.eval_num_classes_in_batch, -1), 0, 1 |
| ).reshape(inputs.shape) |
|
|
| |
| if use_cuda: |
| inputs = inputs.cuda(non_blocking=True) |
| labels = labels.cuda(non_blocking=True) |
|
|
| |
| outputs = model(inputs) |
|
|
| |
| loss = criterion( |
| outputs.view(c.eval_num_classes_in_batch, outputs.shape[0] // c.eval_num_classes_in_batch, -1), labels |
| ) |
|
|
| eval_loss += loss.item() |
|
|
| eval_avg_loss = eval_loss / len(data_loader) |
| |
| dashboard_logger.eval_stats(global_step, {"loss": eval_avg_loss}) |
| |
| figures = { |
| "UMAP Plot": plot_embeddings(outputs.detach().cpu().numpy(), c.num_classes_in_batch), |
| } |
| dashboard_logger.eval_figures(global_step, figures) |
| return eval_avg_loss |
|
|
|
|
| def train(model, optimizer, scheduler, criterion, data_loader, eval_data_loader, global_step): |
| model.train() |
| best_loss = float("inf") |
| avg_loader_time = 0 |
| end_time = time.time() |
| for epoch in range(c.epochs): |
| tot_loss = 0 |
| epoch_time = 0 |
| for _, data in enumerate(data_loader): |
| start_time = time.time() |
|
|
| |
| inputs, labels = data |
| |
| labels = torch.transpose(labels.view(c.num_utter_per_class, c.num_classes_in_batch), 0, 1).reshape( |
| labels.shape |
| ) |
| inputs = torch.transpose(inputs.view(c.num_utter_per_class, c.num_classes_in_batch, -1), 0, 1).reshape( |
| inputs.shape |
| ) |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| loader_time = time.time() - end_time |
| global_step += 1 |
|
|
| |
| if c.lr_decay: |
| scheduler.step() |
| optimizer.zero_grad() |
|
|
| |
| if use_cuda: |
| inputs = inputs.cuda(non_blocking=True) |
| labels = labels.cuda(non_blocking=True) |
|
|
| |
| outputs = model(inputs) |
|
|
| |
| loss = criterion( |
| outputs.view(c.num_classes_in_batch, outputs.shape[0] // c.num_classes_in_batch, -1), labels |
| ) |
| loss.backward() |
| grad_norm, _ = check_update(model, c.grad_clip) |
| optimizer.step() |
|
|
| step_time = time.time() - start_time |
| epoch_time += step_time |
|
|
| |
| tot_loss += loss.item() |
|
|
| |
| num_loader_workers = c.num_loader_workers if c.num_loader_workers > 0 else 1 |
| avg_loader_time = ( |
| 1 / num_loader_workers * loader_time + (num_loader_workers - 1) / num_loader_workers * avg_loader_time |
| if avg_loader_time != 0 |
| else loader_time |
| ) |
| current_lr = optimizer.param_groups[0]["lr"] |
|
|
| if global_step % c.steps_plot_stats == 0: |
| |
| train_stats = { |
| "loss": loss.item(), |
| "lr": current_lr, |
| "grad_norm": grad_norm, |
| "step_time": step_time, |
| "avg_loader_time": avg_loader_time, |
| } |
| dashboard_logger.train_epoch_stats(global_step, train_stats) |
| figures = { |
| "UMAP Plot": plot_embeddings(outputs.detach().cpu().numpy(), c.num_classes_in_batch), |
| } |
| dashboard_logger.train_figures(global_step, figures) |
|
|
| if global_step % c.print_step == 0: |
| print( |
| " | > Step:{} Loss:{:.5f} GradNorm:{:.5f} " |
| "StepTime:{:.2f} LoaderTime:{:.2f} AvGLoaderTime:{:.2f} LR:{:.6f}".format( |
| global_step, loss.item(), grad_norm, step_time, loader_time, avg_loader_time, current_lr |
| ), |
| flush=True, |
| ) |
|
|
| if global_step % c.save_step == 0: |
| |
| save_checkpoint(model, optimizer, criterion, loss.item(), OUT_PATH, global_step, epoch) |
|
|
| end_time = time.time() |
|
|
| print("") |
| print( |
| ">>> Epoch:{} AvgLoss: {:.5f} GradNorm:{:.5f} " |
| "EpochTime:{:.2f} AvGLoaderTime:{:.2f} ".format( |
| epoch, tot_loss / len(data_loader), grad_norm, epoch_time, avg_loader_time |
| ), |
| flush=True, |
| ) |
| |
| if c.run_eval: |
| model.eval() |
| eval_loss = evaluation(model, criterion, eval_data_loader, global_step) |
| print("\n\n") |
| print("--> EVAL PERFORMANCE") |
| print( |
| " | > Epoch:{} AvgLoss: {:.5f} ".format(epoch, eval_loss), |
| flush=True, |
| ) |
| |
| best_loss = save_best_model(model, optimizer, criterion, eval_loss, best_loss, OUT_PATH, global_step, epoch) |
| model.train() |
|
|
| return best_loss, global_step |
|
|
|
|
| def main(args): |
| |
| global meta_data_train |
| global meta_data_eval |
| global train_classes |
|
|
| ap = AudioProcessor(**c.audio) |
| model = setup_encoder_model(c) |
|
|
| optimizer = get_optimizer(c.optimizer, c.optimizer_params, c.lr, model) |
|
|
| |
| meta_data_train, meta_data_eval = load_tts_samples(c.datasets, eval_split=True) |
|
|
| train_data_loader, train_classes, map_classid_to_classname = setup_loader(ap, is_val=False, verbose=True) |
| if c.run_eval: |
| eval_data_loader, _, _ = setup_loader(ap, is_val=True, verbose=True) |
| else: |
| eval_data_loader = None |
|
|
| num_classes = len(train_classes) |
| criterion = model.get_criterion(c, num_classes) |
|
|
| if c.loss == "softmaxproto" and c.model != "speaker_encoder": |
| c.map_classid_to_classname = map_classid_to_classname |
| copy_model_files(c, OUT_PATH) |
|
|
| if args.restore_path: |
| criterion, args.restore_step = model.load_checkpoint( |
| c, args.restore_path, eval=False, use_cuda=use_cuda, criterion=criterion |
| ) |
| print(" > Model restored from step %d" % args.restore_step, flush=True) |
| else: |
| args.restore_step = 0 |
|
|
| if c.lr_decay: |
| scheduler = NoamLR(optimizer, warmup_steps=c.warmup_steps, last_epoch=args.restore_step - 1) |
| else: |
| scheduler = None |
|
|
| num_params = count_parameters(model) |
| print("\n > Model has {} parameters".format(num_params), flush=True) |
|
|
| if use_cuda: |
| model = model.cuda() |
| criterion.cuda() |
|
|
| global_step = args.restore_step |
| _, global_step = train(model, optimizer, scheduler, criterion, train_data_loader, eval_data_loader, global_step) |
|
|
|
|
| if __name__ == "__main__": |
| args, c, OUT_PATH, AUDIO_PATH, c_logger, dashboard_logger = init_training() |
|
|
| try: |
| main(args) |
| except KeyboardInterrupt: |
| remove_experiment_folder(OUT_PATH) |
| try: |
| sys.exit(0) |
| except SystemExit: |
| os._exit(0) |
| except Exception: |
| remove_experiment_folder(OUT_PATH) |
| traceback.print_exc() |
| sys.exit(1) |
|
|