| import os |
| import glob |
| import torch |
| import random |
| import numpy as np |
| from torch.utils.data import Dataset |
| from multiprocessing import Manager |
|
|
|
|
| class GANDataset(Dataset): |
| """ |
| GAN Dataset searchs for all the wav files under root path |
| and converts them to acoustic features on the fly and returns |
| random segments of (audio, feature) couples. |
| """ |
| def __init__(self, |
| ap, |
| items, |
| seq_len, |
| hop_len, |
| pad_short, |
| conv_pad=2, |
| is_training=True, |
| return_segments=True, |
| use_noise_augment=False, |
| use_cache=False, |
| verbose=False): |
|
|
| self.ap = ap |
| self.item_list = items |
| self.compute_feat = not isinstance(items[0], (tuple, list)) |
| self.seq_len = seq_len |
| self.hop_len = hop_len |
| self.pad_short = pad_short |
| self.conv_pad = conv_pad |
| self.is_training = is_training |
| self.return_segments = return_segments |
| self.use_cache = use_cache |
| self.use_noise_augment = use_noise_augment |
| self.verbose = verbose |
|
|
| assert seq_len % hop_len == 0, " [!] seq_len has to be a multiple of hop_len." |
| self.feat_frame_len = seq_len // hop_len + (2 * conv_pad) |
|
|
| |
| self.G_to_D_mappings = list(range(len(self.item_list))) |
| self.shuffle_mapping() |
|
|
| |
| if use_cache: |
| self.create_feature_cache() |
|
|
| def create_feature_cache(self): |
| self.manager = Manager() |
| self.cache = self.manager.list() |
| self.cache += [None for _ in range(len(self.item_list))] |
|
|
| @staticmethod |
| def find_wav_files(path): |
| return glob.glob(os.path.join(path, '**', '*.wav'), recursive=True) |
|
|
| def __len__(self): |
| return len(self.item_list) |
|
|
| def __getitem__(self, idx): |
| """ Return different items for Generator and Discriminator and |
| cache acoustic features """ |
| if self.return_segments: |
| idx2 = self.G_to_D_mappings[idx] |
| item1 = self.load_item(idx) |
| item2 = self.load_item(idx2) |
| return item1, item2 |
| item1 = self.load_item(idx) |
| return item1 |
|
|
| def shuffle_mapping(self): |
| random.shuffle(self.G_to_D_mappings) |
|
|
| def load_item(self, idx): |
| """ load (audio, feat) couple """ |
| if self.compute_feat: |
| |
| wavpath = self.item_list[idx] |
| |
|
|
| if self.use_cache and self.cache[idx] is not None: |
| audio, mel = self.cache[idx] |
| else: |
| audio = self.ap.load_wav(wavpath) |
|
|
| if len(audio) < self.seq_len + self.pad_short: |
| audio = np.pad(audio, (0, self.seq_len + self.pad_short - len(audio)), \ |
| mode='constant', constant_values=0.0) |
|
|
| mel = self.ap.melspectrogram(audio) |
| else: |
|
|
| |
| wavpath, feat_path = self.item_list[idx] |
|
|
| if self.use_cache and self.cache[idx] is not None: |
| audio, mel = self.cache[idx] |
| else: |
| audio = self.ap.load_wav(wavpath) |
| mel = np.load(feat_path) |
|
|
| |
| audio = np.pad(audio, (0, self.hop_len), mode="edge") |
| audio = audio[:mel.shape[-1] * self.hop_len] |
| assert mel.shape[-1] * self.hop_len == audio.shape[-1], f' [!] {mel.shape[-1] * self.hop_len} vs {audio.shape[-1]}' |
|
|
| audio = torch.from_numpy(audio).float().unsqueeze(0) |
| mel = torch.from_numpy(mel).float().squeeze(0) |
|
|
| if self.return_segments: |
| max_mel_start = mel.shape[1] - self.feat_frame_len |
| mel_start = random.randint(0, max_mel_start) |
| mel_end = mel_start + self.feat_frame_len |
| mel = mel[:, mel_start:mel_end] |
|
|
| audio_start = mel_start * self.hop_len |
| audio = audio[:, audio_start:audio_start + |
| self.seq_len] |
|
|
| if self.use_noise_augment and self.is_training and self.return_segments: |
| audio = audio + (1 / 32768) * torch.randn_like(audio) |
| return (mel, audio) |
|
|