|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
"""Utility functions.""" |
|
|
|
|
|
import fnmatch |
|
|
import logging |
|
|
import os |
|
|
import sys |
|
|
import tarfile |
|
|
|
|
|
from distutils.version import LooseVersion |
|
|
from filelock import FileLock |
|
|
|
|
|
import h5py |
|
|
import numpy as np |
|
|
import torch |
|
|
import yaml |
|
|
import soundfile as sf |
|
|
import torchaudio.transforms as transforms |
|
|
|
|
|
def read_wav_16k(audio_path): |
|
|
"""Process audio file to 16kHz sample rate""" |
|
|
if isinstance(audio_path, tuple): |
|
|
wav = audio_path[1] |
|
|
sr = audio_path[0] |
|
|
else: |
|
|
assert os.path.exists(audio_path), f"File not found: {audio_path}" |
|
|
wav, sr = sf.read(audio_path) |
|
|
|
|
|
if sr != 16000: |
|
|
audio_tensor = torch.tensor(wav, dtype=torch.float32) |
|
|
resampler = transforms.Resample(orig_freq=sr, new_freq=16000) |
|
|
wav = resampler(audio_tensor) |
|
|
wav = wav.numpy() |
|
|
return wav |
|
|
|
|
|
|
|
|
|
|
|
def find_files(root_dir, query="*.wav", include_root_dir=True): |
|
|
"""Find files recursively. |
|
|
|
|
|
Args: |
|
|
root_dir (str): Root root_dir to find. |
|
|
query (str): Query to find. |
|
|
include_root_dir (bool): If False, root_dir name is not included. |
|
|
|
|
|
Returns: |
|
|
list: List of found filenames. |
|
|
|
|
|
""" |
|
|
files = [] |
|
|
for root, dirnames, filenames in os.walk(root_dir, followlinks=True): |
|
|
for filename in fnmatch.filter(filenames, query): |
|
|
files.append(os.path.join(root, filename)) |
|
|
if not include_root_dir: |
|
|
files = [file_.replace(root_dir + "/", "") for file_ in files] |
|
|
|
|
|
return files |
|
|
|
|
|
|
|
|
def read_hdf5(hdf5_name, hdf5_path): |
|
|
"""Read hdf5 dataset. |
|
|
|
|
|
Args: |
|
|
hdf5_name (str): Filename of hdf5 file. |
|
|
hdf5_path (str): Dataset name in hdf5 file. |
|
|
|
|
|
Return: |
|
|
any: Dataset values. |
|
|
|
|
|
""" |
|
|
if not os.path.exists(hdf5_name): |
|
|
logging.error(f"There is no such a hdf5 file ({hdf5_name}).") |
|
|
sys.exit(1) |
|
|
|
|
|
hdf5_file = h5py.File(hdf5_name, "r") |
|
|
|
|
|
if hdf5_path not in hdf5_file: |
|
|
logging.error(f"There is no such a data in hdf5 file. ({hdf5_path})") |
|
|
sys.exit(1) |
|
|
|
|
|
hdf5_data = hdf5_file[hdf5_path][()] |
|
|
hdf5_file.close() |
|
|
|
|
|
return hdf5_data |
|
|
|
|
|
|
|
|
def write_hdf5(hdf5_name, hdf5_path, write_data, is_overwrite=True): |
|
|
"""Write dataset to hdf5. |
|
|
|
|
|
Args: |
|
|
hdf5_name (str): Hdf5 dataset filename. |
|
|
hdf5_path (str): Dataset path in hdf5. |
|
|
write_data (ndarray): Data to write. |
|
|
is_overwrite (bool): Whether to overwrite dataset. |
|
|
|
|
|
""" |
|
|
|
|
|
write_data = np.array(write_data) |
|
|
|
|
|
|
|
|
folder_name, _ = os.path.split(hdf5_name) |
|
|
if not os.path.exists(folder_name) and len(folder_name) != 0: |
|
|
os.makedirs(folder_name) |
|
|
|
|
|
|
|
|
if os.path.exists(hdf5_name): |
|
|
|
|
|
hdf5_file = h5py.File(hdf5_name, "r+") |
|
|
|
|
|
if hdf5_path in hdf5_file: |
|
|
if is_overwrite: |
|
|
logging.warning( |
|
|
"Dataset in hdf5 file already exists. " "recreate dataset in hdf5." |
|
|
) |
|
|
hdf5_file.__delitem__(hdf5_path) |
|
|
else: |
|
|
logging.error( |
|
|
"Dataset in hdf5 file already exists. " |
|
|
"if you want to overwrite, please set is_overwrite = True." |
|
|
) |
|
|
hdf5_file.close() |
|
|
sys.exit(1) |
|
|
else: |
|
|
|
|
|
hdf5_file = h5py.File(hdf5_name, "w") |
|
|
|
|
|
|
|
|
hdf5_file.create_dataset(hdf5_path, data=write_data) |
|
|
hdf5_file.flush() |
|
|
hdf5_file.close() |
|
|
|
|
|
|
|
|
class HDF5ScpLoader(object): |
|
|
"""Loader class for a fests.scp file of hdf5 file. |
|
|
|
|
|
Examples: |
|
|
key1 /some/path/a.h5:feats |
|
|
key2 /some/path/b.h5:feats |
|
|
key3 /some/path/c.h5:feats |
|
|
key4 /some/path/d.h5:feats |
|
|
... |
|
|
>>> loader = HDF5ScpLoader("hdf5.scp") |
|
|
>>> array = loader["key1"] |
|
|
|
|
|
key1 /some/path/a.h5 |
|
|
key2 /some/path/b.h5 |
|
|
key3 /some/path/c.h5 |
|
|
key4 /some/path/d.h5 |
|
|
... |
|
|
>>> loader = HDF5ScpLoader("hdf5.scp", "feats") |
|
|
>>> array = loader["key1"] |
|
|
|
|
|
key1 /some/path/a.h5:feats_1,feats_2 |
|
|
key2 /some/path/b.h5:feats_1,feats_2 |
|
|
key3 /some/path/c.h5:feats_1,feats_2 |
|
|
key4 /some/path/d.h5:feats_1,feats_2 |
|
|
... |
|
|
>>> loader = HDF5ScpLoader("hdf5.scp") |
|
|
# feats_1 and feats_2 will be concatenated |
|
|
>>> array = loader["key1"] |
|
|
|
|
|
""" |
|
|
|
|
|
def __init__(self, feats_scp, default_hdf5_path="feats"): |
|
|
"""Initialize HDF5 scp loader. |
|
|
|
|
|
Args: |
|
|
feats_scp (str): Kaldi-style feats.scp file with hdf5 format. |
|
|
default_hdf5_path (str): Path in hdf5 file. If the scp contain the info, not used. |
|
|
|
|
|
""" |
|
|
self.default_hdf5_path = default_hdf5_path |
|
|
with open(feats_scp) as f: |
|
|
lines = [line.replace("\n", "") for line in f.readlines()] |
|
|
self.data = {} |
|
|
for line in lines: |
|
|
key, value = line.split() |
|
|
self.data[key] = value |
|
|
|
|
|
def get_path(self, key): |
|
|
"""Get hdf5 file path for a given key.""" |
|
|
return self.data[key] |
|
|
|
|
|
def __getitem__(self, key): |
|
|
"""Get ndarray for a given key.""" |
|
|
p = self.data[key] |
|
|
if ":" in p: |
|
|
if len(p.split(",")) == 1: |
|
|
return read_hdf5(*p.split(":")) |
|
|
else: |
|
|
p1, p2 = p.split(":") |
|
|
feats = [read_hdf5(p1, p) for p in p2.split(",")] |
|
|
return np.concatenate( |
|
|
[f if len(f.shape) != 1 else f.reshape(-1, 1) for f in feats], 1 |
|
|
) |
|
|
else: |
|
|
return read_hdf5(p, self.default_hdf5_path) |
|
|
|
|
|
def __len__(self): |
|
|
"""Return the length of the scp file.""" |
|
|
return len(self.data) |
|
|
|
|
|
def __iter__(self): |
|
|
"""Return the iterator of the scp file.""" |
|
|
return iter(self.data) |
|
|
|
|
|
def keys(self): |
|
|
"""Return the keys of the scp file.""" |
|
|
return self.data.keys() |
|
|
|
|
|
def values(self): |
|
|
"""Return the values of the scp file.""" |
|
|
for key in self.keys(): |
|
|
yield self[key] |
|
|
|
|
|
|
|
|
class NpyScpLoader(object): |
|
|
"""Loader class for a fests.scp file of npy file. |
|
|
|
|
|
Examples: |
|
|
key1 /some/path/a.npy |
|
|
key2 /some/path/b.npy |
|
|
key3 /some/path/c.npy |
|
|
key4 /some/path/d.npy |
|
|
... |
|
|
>>> loader = NpyScpLoader("feats.scp") |
|
|
>>> array = loader["key1"] |
|
|
|
|
|
""" |
|
|
|
|
|
def __init__(self, feats_scp): |
|
|
"""Initialize npy scp loader. |
|
|
|
|
|
Args: |
|
|
feats_scp (str): Kaldi-style feats.scp file with npy format. |
|
|
|
|
|
""" |
|
|
with open(feats_scp) as f: |
|
|
lines = [line.replace("\n", "") for line in f.readlines()] |
|
|
self.data = {} |
|
|
for line in lines: |
|
|
key, value = line.split() |
|
|
self.data[key] = value |
|
|
|
|
|
def get_path(self, key): |
|
|
"""Get npy file path for a given key.""" |
|
|
return self.data[key] |
|
|
|
|
|
def __getitem__(self, key): |
|
|
"""Get ndarray for a given key.""" |
|
|
return np.load(self.data[key]) |
|
|
|
|
|
def __len__(self): |
|
|
"""Return the length of the scp file.""" |
|
|
return len(self.data) |
|
|
|
|
|
def __iter__(self): |
|
|
"""Return the iterator of the scp file.""" |
|
|
return iter(self.data) |
|
|
|
|
|
def keys(self): |
|
|
"""Return the keys of the scp file.""" |
|
|
return self.data.keys() |
|
|
|
|
|
def values(self): |
|
|
"""Return the values of the scp file.""" |
|
|
for key in self.keys(): |
|
|
yield self[key] |
|
|
|
|
|
|
|
|
def load_model(checkpoint, config=None): |
|
|
"""Load trained model. |
|
|
|
|
|
Args: |
|
|
checkpoint (str): Checkpoint path. |
|
|
config (dict): Configuration dict. |
|
|
|
|
|
Return: |
|
|
torch.nn.Module: Model instance. |
|
|
|
|
|
""" |
|
|
|
|
|
if config is None: |
|
|
dirname = os.path.dirname(checkpoint) |
|
|
config = os.path.join(dirname, "config.yml") |
|
|
with open(config) as f: |
|
|
config = yaml.load(f, Loader=yaml.Loader) |
|
|
|
|
|
|
|
|
import vec2wav2.models |
|
|
|
|
|
|
|
|
model_class = getattr( |
|
|
vec2wav2.models, |
|
|
config.get("generator_type", "BigVGAN"), |
|
|
) |
|
|
model = vec2wav2.models.VEC2WAV2Generator( |
|
|
vec2wav2.models.CTXVEC2WAVFrontend(config["prompt_net_type"], config["num_mels"], **config["frontend_params"]), |
|
|
model_class(**config["generator_params"]) |
|
|
) |
|
|
model.load_state_dict( |
|
|
torch.load(checkpoint, map_location="cpu")["model"]["generator"] |
|
|
) |
|
|
|
|
|
return model |
|
|
|
|
|
def load_feat_codebook(codebook: np.ndarray, device: str="cuda"): |
|
|
"""Given a codebook of shape [G, V, D], convert into a torch Module that can be called. |
|
|
""" |
|
|
feat_codebook = torch.tensor(codebook).to(device) |
|
|
feat_codebook_numgroups = feat_codebook.shape[0] |
|
|
feat_codebook = torch.nn.ModuleList([torch.nn.Embedding.from_pretrained(feat_codebook[i], freeze=True) for i in range(feat_codebook_numgroups)]).to(device) |
|
|
return feat_codebook, feat_codebook_numgroups |
|
|
|
|
|
def idx2vec(codebook: torch.nn.Module, idx: torch.Tensor, num_groups: int): |
|
|
"""Given a codebook (converted, so can be called), and a idx tensor with shape [L, groups] or [B, L, groups] |
|
|
Return the corresponding vectors |
|
|
""" |
|
|
return torch.cat([codebook[i](idx[..., i]) for i in range(num_groups)], dim=-1) |
|
|
|
|
|
def crop_seq(x, offsets, length): |
|
|
"""Crop padded tensor with specified length. |
|
|
|
|
|
:param x: (torch.Tensor) The shape is (B, C, D) |
|
|
:param offsets: (list) |
|
|
:param min_len: (int) |
|
|
:return: |
|
|
""" |
|
|
B, C, T = x.shape |
|
|
x_ = x.new_zeros(B, C, length) |
|
|
for i in range(B): |
|
|
x_[i, :] = x[i, :, offsets[i]: offsets[i] + length] |
|
|
return x_ |
|
|
|
|
|
|