id
int64
0
190k
prompt
stringlengths
21
13.4M
docstring
stringlengths
1
12k
13,989
import os import glob import re import sys import argparse import logging import json import subprocess import warnings import functools import numpy as np from scipy.io.wavfile import read import torch def summarize(writer, global_step, scalars={}, histograms={}, images={}, audios={}, audio_sampling_rate=22050): for k, v in scalars.items(): writer.add_scalar(k, v, global_step) for k, v in histograms.items(): writer.add_histogram(k, v, global_step) for k, v in images.items(): writer.add_image(k, v, global_step, dataformats="HWC") for k, v in audios.items(): writer.add_audio(k, v, global_step, audio_sampling_rate)
null
13,990
import os import glob import re import sys import argparse import logging import json import subprocess import warnings import functools import numpy as np from scipy.io.wavfile import read import torch def latest_checkpoint_path(dir_path, regex="G_*.pth"): f_list = glob.glob(os.path.join(dir_path, regex)) f_list.sort(key=lambda f: int("".join(filter(str.isdigit, f)))) x = f_list[-1] print(x) return x
null
13,991
import os import glob import re import sys import argparse import logging import json import subprocess import warnings import functools import numpy as np from scipy.io.wavfile import read import torch MATPLOTLIB_FLAG = False def plot_spectrogram_to_numpy(spectrogram): global MATPLOTLIB_FLAG if not MATPLOTLIB_FLAG: import matplotlib matplotlib.use("Agg") MATPLOTLIB_FLAG = True mpl_logger = logging.getLogger("matplotlib") mpl_logger.setLevel(logging.WARNING) import matplotlib.pylab as plt import numpy as np fig, ax = plt.subplots(figsize=(10, 2)) im = ax.imshow(spectrogram, aspect="auto", origin="lower", interpolation="none") plt.colorbar(im, ax=ax) plt.xlabel("Frames") plt.ylabel("Channels") plt.tight_layout() fig.canvas.draw() data = np.fromstring(fig.canvas.tostring_rgb(), dtype=np.uint8, sep="") data = data.reshape(fig.canvas.get_width_height()[::-1] + (3,)) plt.close() return data
null
13,992
import os import glob import re import sys import argparse import logging import json import subprocess import warnings import functools import numpy as np from scipy.io.wavfile import read import torch MATPLOTLIB_FLAG = False def plot_alignment_to_numpy(alignment, info=None): global MATPLOTLIB_FLAG if not MATPLOTLIB_FLAG: import matplotlib matplotlib.use("Agg") MATPLOTLIB_FLAG = True mpl_logger = logging.getLogger("matplotlib") mpl_logger.setLevel(logging.WARNING) import matplotlib.pylab as plt import numpy as np fig, ax = plt.subplots(figsize=(6, 4)) im = ax.imshow(alignment.transpose(), aspect="auto", origin="lower", interpolation="none") fig.colorbar(im, ax=ax) xlabel = "Decoder timestep" if info is not None: xlabel += "\n\n" + info plt.xlabel(xlabel) plt.ylabel("Encoder timestep") plt.tight_layout() fig.canvas.draw() data = np.fromstring(fig.canvas.tostring_rgb(), dtype=np.uint8, sep="") data = data.reshape(fig.canvas.get_width_height()[::-1] + (3,)) plt.close() return data
null
13,993
import os import glob import re import sys import argparse import logging import json import subprocess import warnings import functools import numpy as np from scipy.io.wavfile import read import torch def load_wav_to_torch(full_path): sampling_rate, data = read(full_path) return torch.FloatTensor(data.astype(np.float32)), sampling_rate
null
13,994
import os import glob import re import sys import argparse import logging import json import subprocess import warnings import functools import numpy as np from scipy.io.wavfile import read import torch def load_filepaths_and_text(filename, split="|"): with open(filename, encoding="utf-8") as f: filepaths_and_text = [line.strip().split(split) for line in f] return filepaths_and_text
null
13,995
import os import glob import re import sys import argparse import logging import json import subprocess import warnings import functools import numpy as np from scipy.io.wavfile import read import torch class HParams: def __init__(self, **kwargs): def keys(self): def items(self): def values(self): def __len__(self): def __getitem__(self, key): def __setitem__(self, key, value): def __contains__(self, key): def __repr__(self): def get_hparams(init=True): parser = argparse.ArgumentParser() parser.add_argument("-c", "--config", type=str, default="./configs/base.json", help="JSON file for configuration") parser.add_argument("-m", "--model", type=str, required=True, help="Model name") args = parser.parse_args() model_dir = os.path.join("./logs", args.model) if not os.path.exists(model_dir): os.makedirs(model_dir) config_path = args.config config_save_path = os.path.join(model_dir, "config.json") if init: with open(config_path, "r") as f: data = f.read() with open(config_save_path, "w") as f: f.write(data) else: with open(config_save_path, "r") as f: data = f.read() config = json.loads(data) hparams = HParams(**config) hparams.model_dir = model_dir return hparams
null
13,996
import os import glob import re import sys import argparse import logging import json import subprocess import warnings import functools import numpy as np from scipy.io.wavfile import read import torch class HParams: def __init__(self, **kwargs): for k, v in kwargs.items(): if type(v) == dict: v = HParams(**v) self[k] = v def keys(self): return self.__dict__.keys() def items(self): return self.__dict__.items() def values(self): return self.__dict__.values() def __len__(self): return len(self.__dict__) def __getitem__(self, key): return getattr(self, key) def __setitem__(self, key, value): return setattr(self, key, value) def __contains__(self, key): return key in self.__dict__ def __repr__(self): return self.__dict__.__repr__() def get_hparams_from_dir(model_dir): config_save_path = os.path.join(model_dir, "config.json") with open(config_save_path, "r") as f: data = f.read() config = json.loads(data) hparams = HParams(**config) hparams.model_dir = model_dir return hparams
null
13,997
import os import glob import re import sys import argparse import logging import json import subprocess import warnings import functools import numpy as np from scipy.io.wavfile import read import torch class HParams: def __init__(self, **kwargs): for k, v in kwargs.items(): if type(v) == dict: v = HParams(**v) self[k] = v def keys(self): return self.__dict__.keys() def items(self): return self.__dict__.items() def values(self): return self.__dict__.values() def __len__(self): return len(self.__dict__) def __getitem__(self, key): return getattr(self, key) def __setitem__(self, key, value): return setattr(self, key, value) def __contains__(self, key): return key in self.__dict__ def __repr__(self): return self.__dict__.__repr__() def get_hparams_from_file(config_path): with open(config_path, "r") as f: data = f.read() config = json.loads(data) hparams = HParams(**config) return hparams
null
13,998
import os import glob import re import sys import argparse import logging import json import subprocess import warnings import functools import numpy as np from scipy.io.wavfile import read import torch logger = logging def check_git_hash(model_dir): source_dir = os.path.dirname(os.path.realpath(__file__)) if not os.path.exists(os.path.join(source_dir, ".git")): logger.warn("{} is not a git repository, therefore hash value comparison will be ignored.".format(source_dir)) return cur_hash = subprocess.getoutput("git rev-parse HEAD") path = os.path.join(model_dir, "githash") if os.path.exists(path): saved_hash = open(path).read() if saved_hash != cur_hash: logger.warn("git hash values are different. {}(saved) != {}(current)".format(saved_hash[:8], cur_hash[:8])) else: open(path, "w").write(cur_hash)
null
13,999
import os import glob import re import sys import argparse import logging import json import subprocess import warnings import functools import numpy as np from scipy.io.wavfile import read import torch logger = logging def get_logger(model_dir, filename="train.log"): global logger logger = logging.getLogger(os.path.basename(model_dir)) logger.setLevel(logging.DEBUG) formatter = logging.Formatter("%(asctime)s\t%(name)s\t%(levelname)s\t%(message)s") if not os.path.exists(model_dir): os.makedirs(model_dir) h = logging.FileHandler(os.path.join(model_dir, filename)) h.setLevel(logging.DEBUG) h.setFormatter(formatter) logger.addHandler(h) return logger
null
14,000
import os import glob import re import sys import argparse import logging import json import subprocess import warnings import functools import numpy as np from scipy.io.wavfile import read import torch def repeat_expand_2d(content, target_len): # content : [h, t] src_len = content.shape[-1] target = torch.zeros([content.shape[0], target_len], dtype=torch.float).to(content.device) temp = torch.arange(src_len + 1) * target_len / src_len current_pos = 0 for i in range(target_len): if i < temp[current_pos + 1]: target[:, i] = content[:, current_pos] else: current_pos += 1 target[:, i] = content[:, current_pos] return target
null
14,001
import os from pathlib import Path import torch import logging import argparse import numpy as np from sklearn.cluster import KMeans, MiniBatchKMeans import tqdm logger = logging.getLogger(__name__) import time def train_cluster(in_dir, n_clusters, use_minibatch=True, verbose=False): logger.info(f"Loading features from {in_dir}") features = [] nums = 0 for path in tqdm.tqdm(in_dir.glob("*.soft.pt")): features.append(torch.load(path).squeeze(0).numpy().T) # print(features[-1].shape) features = np.concatenate(features, axis=0) print(nums, features.nbytes / 1024**2, "MB , shape:", features.shape, features.dtype) features = features.astype(np.float32) logger.info(f"Clustering features of shape: {features.shape}") t = time.time() if use_minibatch: kmeans = MiniBatchKMeans(n_clusters=n_clusters, verbose=verbose, batch_size=4096, max_iter=80).fit(features) else: kmeans = KMeans(n_clusters=n_clusters, verbose=verbose).fit(features) print(time.time() - t, "s") x = { "n_features_in_": kmeans.n_features_in_, "_n_threads": kmeans._n_threads, "cluster_centers_": kmeans.cluster_centers_, } print("end") return x
null
14,002
import glob import os import torch from torch.nn.utils import weight_norm import matplotlib.pylab as plt def plot_spectrogram(spectrogram): fig, ax = plt.subplots(figsize=(10, 2)) im = ax.imshow(spectrogram, aspect="auto", origin="lower", interpolation="none") plt.colorbar(im, ax=ax) fig.canvas.draw() plt.close() return fig
null
14,003
import glob import os import torch from torch.nn.utils import weight_norm import matplotlib.pylab as plt def init_weights(m, mean=0.0, std=0.01): classname = m.__class__.__name__ if classname.find("Conv") != -1: m.weight.data.normal_(mean, std)
null
14,004
import glob import os import torch from torch.nn.utils import weight_norm import matplotlib.pylab as plt def apply_weight_norm(m): classname = m.__class__.__name__ if classname.find("Conv") != -1: weight_norm(m)
null
14,005
import glob import os import torch from torch.nn.utils import weight_norm import matplotlib.pylab as plt def get_padding(kernel_size, dilation=1): return int((kernel_size * dilation - dilation) / 2)
null
14,006
import glob import os import torch from torch.nn.utils import weight_norm import matplotlib.pylab as plt def load_checkpoint(filepath, device): assert os.path.isfile(filepath) print("Loading '{}'".format(filepath)) checkpoint_dict = torch.load(filepath, map_location=device) print("Complete.") return checkpoint_dict
null
14,007
import glob import os import torch from torch.nn.utils import weight_norm import matplotlib.pylab as plt def save_checkpoint(filepath, obj): print("Saving checkpoint to {}".format(filepath)) torch.save(obj, filepath) print("Complete.")
null
14,008
import glob import os import torch from torch.nn.utils import weight_norm import matplotlib.pylab as plt def del_old_checkpoints(cp_dir, prefix, n_models=2): pattern = os.path.join(cp_dir, prefix + "????????") cp_list = glob.glob(pattern) # get checkpoint paths cp_list = sorted(cp_list) # sort by iter if len(cp_list) > n_models: # if more than n_models models are found for cp in cp_list[:-n_models]: # delete the oldest models other than lastest n_models open(cp, "w").close() # empty file contents os.unlink(cp) # delete file (move to trash when using Colab)
null
14,009
import glob import os import torch from torch.nn.utils import weight_norm import matplotlib.pylab as plt def scan_checkpoint(cp_dir, prefix): pattern = os.path.join(cp_dir, prefix + "????????") cp_list = glob.glob(pattern) if len(cp_list) == 0: return None return sorted(cp_list)[-1]
null
14,010
import os import torch import torch.utils.data import numpy as np import librosa from librosa.filters import mel as librosa_mel_fn import soundfile as sf def load_wav_to_torch(full_path, target_sr=None, return_empty_on_exception=False): sampling_rate = None try: data, sampling_rate = sf.read(full_path, always_2d=True) # than soundfile. except Exception as ex: print(f"'{full_path}' failed to load.\nException:") print(ex) if return_empty_on_exception: return [], sampling_rate or target_sr or 32000 else: raise Exception(ex) if len(data.shape) > 1: data = data[:, 0] assert len(data) > 2 # check duration of audio file is > 2 samples (because otherwise the slice operation was on the wrong dimension) if np.issubdtype(data.dtype, np.integer): # if audio data is type int max_mag = -np.iinfo(data.dtype).min # maximum magnitude = min possible value of intXX else: # if audio data is type fp32 max_mag = max(np.amax(data), -np.amin(data)) max_mag = (2**31) + 1 if max_mag > (2**15) else ((2**15) + 1 if max_mag > 1.01 else 1.0) # data should be either 16-bit INT, 32-bit INT or [-1 to 1] float32 data = torch.FloatTensor(data.astype(np.float32)) / max_mag if (torch.isinf(data) | torch.isnan(data)).any() and return_empty_on_exception: # resample will crash with inf/NaN inputs. return_empty_on_exception will return empty arr instead of except return [], sampling_rate or target_sr or 32000 if target_sr is not None and sampling_rate != target_sr: data = torch.from_numpy(librosa.core.resample(data.numpy(), orig_sr=sampling_rate, target_sr=target_sr)) sampling_rate = target_sr return data, sampling_rate
null
14,011
import os import torch import torch.utils.data import numpy as np import librosa from librosa.filters import mel as librosa_mel_fn import soundfile as sf def dynamic_range_compression(x, C=1, clip_val=1e-5): return np.log(np.clip(x, a_min=clip_val, a_max=None) * C)
null
14,012
import os import torch import torch.utils.data import numpy as np import librosa from librosa.filters import mel as librosa_mel_fn import soundfile as sf def dynamic_range_decompression(x, C=1): return np.exp(x) / C
null
14,013
import os import torch import torch.utils.data import numpy as np import librosa from librosa.filters import mel as librosa_mel_fn import soundfile as sf def dynamic_range_compression_torch(x, C=1, clip_val=1e-5): return torch.log(torch.clamp(x, min=clip_val) * C)
null
14,014
import os import torch import torch.utils.data import numpy as np import librosa from librosa.filters import mel as librosa_mel_fn import soundfile as sf def dynamic_range_decompression_torch(x, C=1): return torch.exp(x) / C
null
14,015
import os import json from .env import AttrDict import numpy as np import torch import torch.nn.functional as F import torch.nn as nn from torch.nn import Conv1d, ConvTranspose1d, AvgPool1d, Conv2d from torch.nn.utils import weight_norm, remove_weight_norm, spectral_norm from .utils import init_weights, get_padding class Generator(torch.nn.Module): def __init__(self, h): super(Generator, self).__init__() self.h = h self.num_kernels = len(h["resblock_kernel_sizes"]) self.num_upsamples = len(h["upsample_rates"]) self.f0_upsamp = torch.nn.Upsample(scale_factor=np.prod(h["upsample_rates"])) self.m_source = SourceModuleHnNSF(sampling_rate=h["sampling_rate"], harmonic_num=8) self.noise_convs = nn.ModuleList() self.conv_pre = weight_norm(Conv1d(h["inter_channels"], h["upsample_initial_channel"], 7, 1, padding=3)) resblock = ResBlock1 if h["resblock"] == "1" else ResBlock2 self.ups = nn.ModuleList() for i, (u, k) in enumerate(zip(h["upsample_rates"], h["upsample_kernel_sizes"])): c_cur = h["upsample_initial_channel"] // (2 ** (i + 1)) self.ups.append(weight_norm(ConvTranspose1d(h["upsample_initial_channel"] // (2**i), h["upsample_initial_channel"] // (2 ** (i + 1)), k, u, padding=(k - u) // 2))) if i + 1 < len(h["upsample_rates"]): # stride_f0 = np.prod(h["upsample_rates"][i + 1 :]) self.noise_convs.append(Conv1d(1, c_cur, kernel_size=stride_f0 * 2, stride=stride_f0, padding=stride_f0 // 2)) else: self.noise_convs.append(Conv1d(1, c_cur, kernel_size=1)) self.resblocks = nn.ModuleList() for i in range(len(self.ups)): ch = h["upsample_initial_channel"] // (2 ** (i + 1)) for j, (k, d) in enumerate(zip(h["resblock_kernel_sizes"], h["resblock_dilation_sizes"])): self.resblocks.append(resblock(h, ch, k, d)) self.conv_post = weight_norm(Conv1d(ch, 1, 7, 1, padding=3)) self.ups.apply(init_weights) self.conv_post.apply(init_weights) self.cond = nn.Conv1d(h["gin_channels"], h["upsample_initial_channel"], 1) def forward(self, x, f0, g=None): # print(1,x.shape,f0.shape,f0[:, None].shape) f0 = self.f0_upsamp(f0[:, None]).transpose(1, 2) # bs,n,t # print(2,f0.shape) har_source, noi_source, uv = self.m_source(f0) har_source = har_source.transpose(1, 2) x = self.conv_pre(x) x = x + self.cond(g) # print(124,x.shape,har_source.shape) for i in range(self.num_upsamples): x = F.leaky_relu(x, LRELU_SLOPE) # print(3,x.shape) x = self.ups[i](x) x_source = self.noise_convs[i](har_source) # print(4,x_source.shape,har_source.shape,x.shape) x = x + x_source xs = None for j in range(self.num_kernels): if xs is None: xs = self.resblocks[i * self.num_kernels + j](x) else: xs += self.resblocks[i * self.num_kernels + j](x) x = xs / self.num_kernels x = F.leaky_relu(x) x = self.conv_post(x) x = torch.tanh(x) return x def remove_weight_norm(self): print("Removing weight norm...") for l in self.ups: remove_weight_norm(l) for l in self.resblocks: l.remove_weight_norm() remove_weight_norm(self.conv_pre) remove_weight_norm(self.conv_post) class AttrDict(dict): def __init__(self, *args, **kwargs): super(AttrDict, self).__init__(*args, **kwargs) self.__dict__ = self def load_model(model_path, device="cuda"): config_file = os.path.join(os.path.split(model_path)[0], "config.json") with open(config_file) as f: data = f.read() global h json_config = json.loads(data) h = AttrDict(json_config) generator = Generator(h).to(device) cp_dict = torch.load(model_path) generator.load_state_dict(cp_dict["generator"]) generator.eval() generator.remove_weight_norm() del cp_dict return generator, h
null
14,016
import os import json from .env import AttrDict import numpy as np import torch import torch.nn.functional as F import torch.nn as nn from torch.nn import Conv1d, ConvTranspose1d, AvgPool1d, Conv2d from torch.nn.utils import weight_norm, remove_weight_norm, spectral_norm from .utils import init_weights, get_padding def padDiff(x): return F.pad(F.pad(x, (0, 0, -1, 1), "constant", 0) - x, (0, 0, 0, -1), "constant", 0)
null
14,017
import os import json from .env import AttrDict import numpy as np import torch import torch.nn.functional as F import torch.nn as nn from torch.nn import Conv1d, ConvTranspose1d, AvgPool1d, Conv2d from torch.nn.utils import weight_norm, remove_weight_norm, spectral_norm from .utils import init_weights, get_padding def feature_loss(fmap_r, fmap_g): loss = 0 for dr, dg in zip(fmap_r, fmap_g): for rl, gl in zip(dr, dg): loss += torch.mean(torch.abs(rl - gl)) return loss * 2
null
14,018
import os import json from .env import AttrDict import numpy as np import torch import torch.nn.functional as F import torch.nn as nn from torch.nn import Conv1d, ConvTranspose1d, AvgPool1d, Conv2d from torch.nn.utils import weight_norm, remove_weight_norm, spectral_norm from .utils import init_weights, get_padding def discriminator_loss(disc_real_outputs, disc_generated_outputs): loss = 0 r_losses = [] g_losses = [] for dr, dg in zip(disc_real_outputs, disc_generated_outputs): r_loss = torch.mean((1 - dr) ** 2) g_loss = torch.mean(dg**2) loss += r_loss + g_loss r_losses.append(r_loss.item()) g_losses.append(g_loss.item()) return loss, r_losses, g_losses
null
14,019
import os import json from .env import AttrDict import numpy as np import torch import torch.nn.functional as F import torch.nn as nn from torch.nn import Conv1d, ConvTranspose1d, AvgPool1d, Conv2d from torch.nn.utils import weight_norm, remove_weight_norm, spectral_norm from .utils import init_weights, get_padding def generator_loss(disc_outputs): loss = 0 gen_losses = [] for dg in disc_outputs: l = torch.mean((1 - dg) ** 2) gen_losses.append(l) loss += l return loss, gen_losses
null
14,020
import os import shutil def build_env(config, config_name, path): t_path = os.path.join(path, config_name) if config != t_path: os.makedirs(path, exist_ok=True) shutil.copyfile(config, os.path.join(path, config_name))
null
14,021
import glob import os import matplotlib import torch from torch.nn.utils import weight_norm import matplotlib.pylab as plt def plot_spectrogram(spectrogram): fig, ax = plt.subplots(figsize=(10, 2)) im = ax.imshow(spectrogram, aspect="auto", origin="lower", interpolation="none") plt.colorbar(im, ax=ax) fig.canvas.draw() plt.close() return fig
null
14,022
import glob import os import matplotlib import torch from torch.nn.utils import weight_norm import matplotlib.pylab as plt def init_weights(m, mean=0.0, std=0.01): classname = m.__class__.__name__ if classname.find("Conv") != -1: m.weight.data.normal_(mean, std)
null
14,023
import glob import os import matplotlib import torch from torch.nn.utils import weight_norm import matplotlib.pylab as plt def apply_weight_norm(m): classname = m.__class__.__name__ if classname.find("Conv") != -1: weight_norm(m)
null
14,024
import glob import os import matplotlib import torch from torch.nn.utils import weight_norm import matplotlib.pylab as plt def get_padding(kernel_size, dilation=1): return int((kernel_size * dilation - dilation) / 2)
null
14,025
import glob import os import matplotlib import torch from torch.nn.utils import weight_norm import matplotlib.pylab as plt def load_checkpoint(filepath, device): assert os.path.isfile(filepath) print("Loading '{}'".format(filepath)) checkpoint_dict = torch.load(filepath, map_location=device) print("Complete.") return checkpoint_dict
null
14,026
import glob import os import matplotlib import torch from torch.nn.utils import weight_norm import matplotlib.pylab as plt def save_checkpoint(filepath, obj): print("Saving checkpoint to {}".format(filepath)) torch.save(obj, filepath) print("Complete.")
null
14,027
import glob import os import matplotlib import torch from torch.nn.utils import weight_norm import matplotlib.pylab as plt def del_old_checkpoints(cp_dir, prefix, n_models=2): pattern = os.path.join(cp_dir, prefix + "????????") cp_list = glob.glob(pattern) # get checkpoint paths cp_list = sorted(cp_list) # sort by iter if len(cp_list) > n_models: # if more than n_models models are found for cp in cp_list[:-n_models]: # delete the oldest models other than lastest n_models open(cp, "w").close() # empty file contents os.unlink(cp) # delete file (move to trash when using Colab)
null
14,028
import glob import os import matplotlib import torch from torch.nn.utils import weight_norm import matplotlib.pylab as plt def scan_checkpoint(cp_dir, prefix): pattern = os.path.join(cp_dir, prefix + "????????") cp_list = glob.glob(pattern) if len(cp_list) == 0: return None return sorted(cp_list)[-1]
null
14,029
import os import torch import torch.utils.data import numpy as np import librosa from librosa.filters import mel as librosa_mel_fn import soundfile as sf import torch.nn.functional as F def load_wav_to_torch(full_path, target_sr=None, return_empty_on_exception=False): sampling_rate = None try: data, sampling_rate = sf.read(full_path, always_2d=True) # than soundfile. except Exception as ex: print(f"'{full_path}' failed to load.\nException:") print(ex) if return_empty_on_exception: return [], sampling_rate or target_sr or 48000 else: raise Exception(ex) if len(data.shape) > 1: data = data[:, 0] assert len(data) > 2 # check duration of audio file is > 2 samples (because otherwise the slice operation was on the wrong dimension) if np.issubdtype(data.dtype, np.integer): # if audio data is type int max_mag = -np.iinfo(data.dtype).min # maximum magnitude = min possible value of intXX else: # if audio data is type fp32 max_mag = max(np.amax(data), -np.amin(data)) max_mag = (2**31) + 1 if max_mag > (2**15) else ((2**15) + 1 if max_mag > 1.01 else 1.0) # data should be either 16-bit INT, 32-bit INT or [-1 to 1] float32 data = torch.FloatTensor(data.astype(np.float32)) / max_mag if (torch.isinf(data) | torch.isnan(data)).any() and return_empty_on_exception: # resample will crash with inf/NaN inputs. return_empty_on_exception will return empty arr instead of except return [], sampling_rate or target_sr or 48000 if target_sr is not None and sampling_rate != target_sr: data = torch.from_numpy(librosa.core.resample(data.numpy(), orig_sr=sampling_rate, target_sr=target_sr)) sampling_rate = target_sr return data, sampling_rate
null
14,030
import os import torch import torch.utils.data import numpy as np import librosa from librosa.filters import mel as librosa_mel_fn import soundfile as sf import torch.nn.functional as F def dynamic_range_compression(x, C=1, clip_val=1e-5): return np.log(np.clip(x, a_min=clip_val, a_max=None) * C)
null
14,031
import os import torch import torch.utils.data import numpy as np import librosa from librosa.filters import mel as librosa_mel_fn import soundfile as sf import torch.nn.functional as F def dynamic_range_decompression(x, C=1): return np.exp(x) / C
null
14,032
import os import torch import torch.utils.data import numpy as np import librosa from librosa.filters import mel as librosa_mel_fn import soundfile as sf import torch.nn.functional as F def dynamic_range_compression_torch(x, C=1, clip_val=1e-5): return torch.log(torch.clamp(x, min=clip_val) * C)
null
14,033
import os import torch import torch.utils.data import numpy as np import librosa from librosa.filters import mel as librosa_mel_fn import soundfile as sf import torch.nn.functional as F def dynamic_range_decompression_torch(x, C=1): return torch.exp(x) / C
null
14,034
import os import json from .env import AttrDict import numpy as np import torch import torch.nn.functional as F import torch.nn as nn from torch.nn import Conv1d, ConvTranspose1d, AvgPool1d, Conv2d from torch.nn.utils import weight_norm, remove_weight_norm, spectral_norm from .utils import init_weights, get_padding class Generator(torch.nn.Module): def __init__(self, h): super(Generator, self).__init__() self.h = h self.num_kernels = len(h.resblock_kernel_sizes) self.num_upsamples = len(h.upsample_rates) self.m_source = SourceModuleHnNSF(sampling_rate=h.sampling_rate, harmonic_num=8) self.noise_convs = nn.ModuleList() self.conv_pre = weight_norm(Conv1d(h.num_mels, h.upsample_initial_channel, 7, 1, padding=3)) resblock = ResBlock1 if h.resblock == "1" else ResBlock2 self.ups = nn.ModuleList() for i, (u, k) in enumerate(zip(h.upsample_rates, h.upsample_kernel_sizes)): c_cur = h.upsample_initial_channel // (2 ** (i + 1)) self.ups.append(weight_norm(ConvTranspose1d(h.upsample_initial_channel // (2**i), h.upsample_initial_channel // (2 ** (i + 1)), k, u, padding=(k - u) // 2))) if i + 1 < len(h.upsample_rates): # stride_f0 = int(np.prod(h.upsample_rates[i + 1 :])) self.noise_convs.append(Conv1d(1, c_cur, kernel_size=stride_f0 * 2, stride=stride_f0, padding=stride_f0 // 2)) else: self.noise_convs.append(Conv1d(1, c_cur, kernel_size=1)) self.resblocks = nn.ModuleList() ch = h.upsample_initial_channel for i in range(len(self.ups)): ch //= 2 for j, (k, d) in enumerate(zip(h.resblock_kernel_sizes, h.resblock_dilation_sizes)): self.resblocks.append(resblock(h, ch, k, d)) self.conv_post = weight_norm(Conv1d(ch, 1, 7, 1, padding=3)) self.ups.apply(init_weights) self.conv_post.apply(init_weights) self.upp = int(np.prod(h.upsample_rates)) def forward(self, x, f0): har_source = self.m_source(f0, self.upp).transpose(1, 2) x = self.conv_pre(x) for i in range(self.num_upsamples): x = F.leaky_relu(x, LRELU_SLOPE) x = self.ups[i](x) x_source = self.noise_convs[i](har_source) x = x + x_source xs = None for j in range(self.num_kernels): if xs is None: xs = self.resblocks[i * self.num_kernels + j](x) else: xs += self.resblocks[i * self.num_kernels + j](x) x = xs / self.num_kernels x = F.leaky_relu(x) x = self.conv_post(x) x = torch.tanh(x) return x def remove_weight_norm(self): print("Removing weight norm...") for l in self.ups: remove_weight_norm(l) for l in self.resblocks: l.remove_weight_norm() remove_weight_norm(self.conv_pre) remove_weight_norm(self.conv_post) class AttrDict(dict): def __init__(self, *args, **kwargs): super(AttrDict, self).__init__(*args, **kwargs) self.__dict__ = self def load_model(model_path, device="cuda"): config_file = os.path.join(os.path.split(model_path)[0], "config.json") with open(config_file) as f: data = f.read() json_config = json.loads(data) h = AttrDict(json_config) generator = Generator(h).to(device) cp_dict = torch.load(model_path, map_location=device) generator.load_state_dict(cp_dict["generator"]) generator.eval() generator.remove_weight_norm() del cp_dict return generator, h
null
14,039
import sys import os from dataclasses import asdict import numpy as np import torch from data.ModelSlot import DDSPSVCModelSlot from voice_changer.DDSP_SVC.deviceManager.DeviceManager import DeviceManager from voice_changer.VoiceChangerParamsManager import VoiceChangerParamsManager from .models.diffusion.infer_gt_mel import DiffGtMel from voice_changer.utils.VoiceChangerModel import AudioInOut, VoiceChangerModel from voice_changer.utils.VoiceChangerParams import VoiceChangerParams from voice_changer.DDSP_SVC.DDSP_SVCSetting import DDSP_SVCSettings from voice_changer.RVC.embedder.EmbedderManager import EmbedderManager from voice_changer.DDSP_SVC.SvcDDSP import SvcDDSP def phase_vocoder(a, b, fade_out, fade_in): fa = torch.fft.rfft(a) fb = torch.fft.rfft(b) absab = torch.abs(fa) + torch.abs(fb) n = a.shape[0] if n % 2 == 0: absab[1:-1] *= 2 else: absab[1:] *= 2 phia = torch.angle(fa) phib = torch.angle(fb) deltaphase = phib - phia deltaphase = deltaphase - 2 * np.pi * torch.floor(deltaphase / 2 / np.pi + 0.5) w = 2 * np.pi * torch.arange(n // 2 + 1).to(a) + deltaphase t = torch.arange(n).unsqueeze(-1).to(a) / n result = ( a * (fade_out**2) + b * (fade_in**2) + torch.sum(absab * torch.cos(w * t + phia), -1) * fade_out * fade_in / n ) return result
null
14,040
import copy from typing import Optional, Tuple import random from sklearn.cluster import KMeans import torch import torch.nn as nn import torch.nn.functional as F from torch.nn.modules.utils import consume_prefix_in_state_dict_if_present def _compute_mask( shape: Tuple[int, int], mask_prob: float, mask_length: int, device: torch.device, min_masks: int = 0, ) -> torch.Tensor: batch_size, sequence_length = shape if mask_length < 1: raise ValueError("`mask_length` has to be bigger than 0.") if mask_length > sequence_length: raise ValueError( f"`mask_length` has to be smaller than `sequence_length`, but got `mask_length`: {mask_length} and `sequence_length`: {sequence_length}`" ) # compute number of masked spans in batch num_masked_spans = int(mask_prob * sequence_length / mask_length + random.random()) num_masked_spans = max(num_masked_spans, min_masks) # make sure num masked indices <= sequence_length if num_masked_spans * mask_length > sequence_length: num_masked_spans = sequence_length // mask_length # SpecAugment mask to fill mask = torch.zeros((batch_size, sequence_length), device=device, dtype=torch.bool) # uniform distribution to sample from, make sure that offset samples are < sequence_length uniform_dist = torch.ones( (batch_size, sequence_length - (mask_length - 1)), device=device ) # get random indices to mask mask_indices = torch.multinomial(uniform_dist, num_masked_spans) # expand masked indices to masked spans mask_indices = ( mask_indices.unsqueeze(dim=-1) .expand((batch_size, num_masked_spans, mask_length)) .reshape(batch_size, num_masked_spans * mask_length) ) offsets = ( torch.arange(mask_length, device=device)[None, None, :] .expand((batch_size, num_masked_spans, mask_length)) .reshape(batch_size, num_masked_spans * mask_length) ) mask_idxs = mask_indices + offsets # scatter indices to mask mask = mask.scatter(1, mask_idxs, True) return mask
null
14,041
import copy from typing import Optional, Tuple import random from sklearn.cluster import KMeans import torch import torch.nn as nn import torch.nn.functional as F from torch.nn.modules.utils import consume_prefix_in_state_dict_if_present URLS = { "hubert-discrete": "https://github.com/bshall/hubert/releases/download/v0.1/hubert-discrete-e9416457.pt", "hubert-soft": "https://github.com/bshall/hubert/releases/download/v0.1/hubert-soft-0d54a1f4.pt", "kmeans100": "https://github.com/bshall/hubert/releases/download/v0.1/kmeans100-50f36a95.pt", } class HubertDiscrete(Hubert): def __init__(self, kmeans): super().__init__(504) self.kmeans = kmeans def units(self, wav: torch.Tensor) -> torch.LongTensor: wav = F.pad(wav, ((400 - 320) // 2, (400 - 320) // 2)) x, _ = self.encode(wav, layer=7) x = self.kmeans.predict(x.squeeze().cpu().numpy()) return torch.tensor(x, dtype=torch.long, device=wav.device) def kmeans100(pretrained: bool = True, progress: bool = True) -> KMeans: r""" k-means checkpoint for HuBERT-Discrete with 100 clusters. Args: pretrained (bool): load pretrained weights into the model progress (bool): show progress bar when downloading model """ return _kmeans(100, pretrained, progress) The provided code snippet includes necessary dependencies for implementing the `hubert_discrete` function. Write a Python function `def hubert_discrete( pretrained: bool = True, progress: bool = True, ) -> HubertDiscrete` to solve the following problem: r"""HuBERT-Discrete from `"A Comparison of Discrete and Soft Speech Units for Improved Voice Conversion"`. Args: pretrained (bool): load pretrained weights into the model progress (bool): show progress bar when downloading model Here is the function: def hubert_discrete( pretrained: bool = True, progress: bool = True, ) -> HubertDiscrete: r"""HuBERT-Discrete from `"A Comparison of Discrete and Soft Speech Units for Improved Voice Conversion"`. Args: pretrained (bool): load pretrained weights into the model progress (bool): show progress bar when downloading model """ kmeans = kmeans100(pretrained=pretrained, progress=progress) hubert = HubertDiscrete(kmeans) if pretrained: checkpoint = torch.hub.load_state_dict_from_url( URLS["hubert-discrete"], progress=progress ) consume_prefix_in_state_dict_if_present(checkpoint, "module.") hubert.load_state_dict(checkpoint) hubert.eval() return hubert
r"""HuBERT-Discrete from `"A Comparison of Discrete and Soft Speech Units for Improved Voice Conversion"`. Args: pretrained (bool): load pretrained weights into the model progress (bool): show progress bar when downloading model
14,042
import copy from typing import Optional, Tuple import random from sklearn.cluster import KMeans import torch import torch.nn as nn import torch.nn.functional as F from torch.nn.modules.utils import consume_prefix_in_state_dict_if_present URLS = { "hubert-discrete": "https://github.com/bshall/hubert/releases/download/v0.1/hubert-discrete-e9416457.pt", "hubert-soft": "https://github.com/bshall/hubert/releases/download/v0.1/hubert-soft-0d54a1f4.pt", "kmeans100": "https://github.com/bshall/hubert/releases/download/v0.1/kmeans100-50f36a95.pt", } class HubertSoft(Hubert): def __init__(self): super().__init__() def units(self, wav: torch.Tensor) -> torch.Tensor: wav = F.pad(wav, ((400 - 320) // 2, (400 - 320) // 2)) x, _ = self.encode(wav) return self.proj(x) The provided code snippet includes necessary dependencies for implementing the `hubert_soft` function. Write a Python function `def hubert_soft( pretrained: bool = True, progress: bool = True, ) -> HubertSoft` to solve the following problem: r"""HuBERT-Soft from `"A Comparison of Discrete and Soft Speech Units for Improved Voice Conversion"`. Args: pretrained (bool): load pretrained weights into the model progress (bool): show progress bar when downloading model Here is the function: def hubert_soft( pretrained: bool = True, progress: bool = True, ) -> HubertSoft: r"""HuBERT-Soft from `"A Comparison of Discrete and Soft Speech Units for Improved Voice Conversion"`. Args: pretrained (bool): load pretrained weights into the model progress (bool): show progress bar when downloading model """ hubert = HubertSoft() if pretrained: checkpoint = torch.hub.load_state_dict_from_url( URLS["hubert-soft"], progress=progress ) consume_prefix_in_state_dict_if_present(checkpoint, "module.") hubert.load_state_dict(checkpoint) hubert.eval() return hubert
r"""HuBERT-Soft from `"A Comparison of Discrete and Soft Speech Units for Improved Voice Conversion"`. Args: pretrained (bool): load pretrained weights into the model progress (bool): show progress bar when downloading model
14,043
import os import time import numpy as np import torch import librosa from logger.saver import Saver from logger import utils from torch import autocast from torch.cuda.amp import GradScaler def test(args, model, vocoder, loader_test, saver): def train(args, initial_global_step, model, optimizer, scheduler, vocoder, loader_train, loader_test): # saver saver = Saver(args, initial_global_step=initial_global_step) # model size params_count = utils.get_network_paras_amount({'model': model}) saver.log_info('--- model size ---') saver.log_info(params_count) # run num_batches = len(loader_train) model.train() saver.log_info('======= start training =======') scaler = GradScaler() if args.train.amp_dtype == 'fp32': dtype = torch.float32 elif args.train.amp_dtype == 'fp16': dtype = torch.float16 elif args.train.amp_dtype == 'bf16': dtype = torch.bfloat16 else: raise ValueError(' [x] Unknown amp_dtype: ' + args.train.amp_dtype) for epoch in range(args.train.epochs): for batch_idx, data in enumerate(loader_train): saver.global_step_increment() optimizer.zero_grad() # unpack data for k in data.keys(): if not k.startswith('name'): data[k] = data[k].to(args.device) # forward if dtype == torch.float32: loss = model(data['units'].float(), data['f0'], data['volume'], data['spk_id'], aug_shift = data['aug_shift'], gt_spec=data['mel'].float(), infer=False) else: with autocast(device_type=args.device, dtype=dtype): loss = model(data['units'], data['f0'], data['volume'], data['spk_id'], aug_shift = data['aug_shift'], gt_spec=data['mel'], infer=False) # handle nan loss if torch.isnan(loss): raise ValueError(' [x] nan loss ') else: # backpropagate if dtype == torch.float32: loss.backward() optimizer.step() else: scaler.scale(loss).backward() scaler.step(optimizer) scaler.update() scheduler.step() # log loss if saver.global_step % args.train.interval_log == 0: current_lr = optimizer.param_groups[0]['lr'] saver.log_info( 'epoch: {} | {:3d}/{:3d} | {} | batch/s: {:.2f} | lr: {:.6} | loss: {:.3f} | time: {} | step: {}'.format( epoch, batch_idx, num_batches, args.env.expdir, args.train.interval_log/saver.get_interval_time(), current_lr, loss.item(), saver.get_total_time(), saver.global_step ) ) saver.log_value({ 'train/loss': loss.item() }) saver.log_value({ 'train/lr': current_lr }) # validation if saver.global_step % args.train.interval_val == 0: optimizer_save = optimizer if args.train.save_opt else None # save latest saver.save_model(model, optimizer_save, postfix=f'{saver.global_step}') last_val_step = saver.global_step - args.train.interval_val if last_val_step % args.train.interval_force_save != 0: saver.delete_model(postfix=f'{last_val_step}') # run testing set test_loss = test(args, model, vocoder, loader_test, saver) # log loss saver.log_info( ' --- <validation> --- \nloss: {:.3f}. '.format( test_loss, ) ) saver.log_value({ 'validation/loss': test_loss }) model.train()
null
14,044
import torch import torch.nn.functional as F import math The provided code snippet includes necessary dependencies for implementing the `model_wrapper` function. Write a Python function `def model_wrapper( model, noise_schedule, model_type="noise", model_kwargs={}, guidance_type="uncond", condition=None, unconditional_condition=None, guidance_scale=1., classifier_fn=None, classifier_kwargs={}, )` to solve the following problem: Create a wrapper function for the noise prediction model. Here is the function: def model_wrapper( model, noise_schedule, model_type="noise", model_kwargs={}, guidance_type="uncond", condition=None, unconditional_condition=None, guidance_scale=1., classifier_fn=None, classifier_kwargs={}, ): """Create a wrapper function for the noise prediction model. """ def get_model_input_time(t_continuous): """ Convert the continuous-time `t_continuous` (in [epsilon, T]) to the model input time. For discrete-time DPMs, we convert `t_continuous` in [1 / N, 1] to `t_input` in [0, 1000 * (N - 1) / N]. For continuous-time DPMs, we just use `t_continuous`. """ if noise_schedule.schedule == 'discrete': return (t_continuous - 1. / noise_schedule.total_N) * noise_schedule.total_N else: return t_continuous def noise_pred_fn(x, t_continuous, cond=None): t_input = get_model_input_time(t_continuous) if cond is None: output = model(x, t_input, **model_kwargs) else: output = model(x, t_input, cond, **model_kwargs) if model_type == "noise": return output elif model_type == "x_start": alpha_t, sigma_t = noise_schedule.marginal_alpha(t_continuous), noise_schedule.marginal_std(t_continuous) return (x - alpha_t * output) / sigma_t elif model_type == "v": alpha_t, sigma_t = noise_schedule.marginal_alpha(t_continuous), noise_schedule.marginal_std(t_continuous) return alpha_t * output + sigma_t * x elif model_type == "score": sigma_t = noise_schedule.marginal_std(t_continuous) return -sigma_t * output def cond_grad_fn(x, t_input): """ Compute the gradient of the classifier, i.e. nabla_{x} log p_t(cond | x_t). """ with torch.enable_grad(): x_in = x.detach().requires_grad_(True) log_prob = classifier_fn(x_in, t_input, condition, **classifier_kwargs) return torch.autograd.grad(log_prob.sum(), x_in)[0] def model_fn(x, t_continuous): """ The noise predicition model function that is used for DPM-Solver. """ if guidance_type == "uncond": return noise_pred_fn(x, t_continuous) elif guidance_type == "classifier": assert classifier_fn is not None t_input = get_model_input_time(t_continuous) cond_grad = cond_grad_fn(x, t_input) sigma_t = noise_schedule.marginal_std(t_continuous) noise = noise_pred_fn(x, t_continuous) return noise - guidance_scale * sigma_t * cond_grad elif guidance_type == "classifier-free": if guidance_scale == 1. or unconditional_condition is None: return noise_pred_fn(x, t_continuous, cond=condition) else: x_in = torch.cat([x] * 2) t_in = torch.cat([t_continuous] * 2) c_in = torch.cat([unconditional_condition, condition]) noise_uncond, noise = noise_pred_fn(x_in, t_in, cond=c_in).chunk(2) return noise_uncond + guidance_scale * (noise - noise_uncond) assert model_type in ["noise", "x_start", "v"] assert guidance_type in ["uncond", "classifier", "classifier-free"] return model_fn
Create a wrapper function for the noise prediction model.
14,045
import torch import torch.nn.functional as F import math The provided code snippet includes necessary dependencies for implementing the `interpolate_fn` function. Write a Python function `def interpolate_fn(x, xp, yp)` to solve the following problem: A piecewise linear function y = f(x), using xp and yp as keypoints. We implement f(x) in a differentiable way (i.e. applicable for autograd). The function f(x) is well-defined for all x-axis. (For x beyond the bounds of xp, we use the outmost points of xp to define the linear function.) Args: x: PyTorch tensor with shape [N, C], where N is the batch size, C is the number of channels (we use C = 1 for DPM-Solver). xp: PyTorch tensor with shape [C, K], where K is the number of keypoints. yp: PyTorch tensor with shape [C, K]. Returns: The function values f(x), with shape [N, C]. Here is the function: def interpolate_fn(x, xp, yp): """ A piecewise linear function y = f(x), using xp and yp as keypoints. We implement f(x) in a differentiable way (i.e. applicable for autograd). The function f(x) is well-defined for all x-axis. (For x beyond the bounds of xp, we use the outmost points of xp to define the linear function.) Args: x: PyTorch tensor with shape [N, C], where N is the batch size, C is the number of channels (we use C = 1 for DPM-Solver). xp: PyTorch tensor with shape [C, K], where K is the number of keypoints. yp: PyTorch tensor with shape [C, K]. Returns: The function values f(x), with shape [N, C]. """ N, K = x.shape[0], xp.shape[1] all_x = torch.cat([x.unsqueeze(2), xp.unsqueeze(0).repeat((N, 1, 1))], dim=2) sorted_all_x, x_indices = torch.sort(all_x, dim=2) x_idx = torch.argmin(x_indices, dim=2) cand_start_idx = x_idx - 1 start_idx = torch.where( torch.eq(x_idx, 0), torch.tensor(1, device=x.device), torch.where( torch.eq(x_idx, K), torch.tensor(K - 2, device=x.device), cand_start_idx, ), ) end_idx = torch.where(torch.eq(start_idx, cand_start_idx), start_idx + 2, start_idx + 1) start_x = torch.gather(sorted_all_x, dim=2, index=start_idx.unsqueeze(2)).squeeze(2) end_x = torch.gather(sorted_all_x, dim=2, index=end_idx.unsqueeze(2)).squeeze(2) start_idx2 = torch.where( torch.eq(x_idx, 0), torch.tensor(0, device=x.device), torch.where( torch.eq(x_idx, K), torch.tensor(K - 2, device=x.device), cand_start_idx, ), ) y_positions_expanded = yp.unsqueeze(0).expand(N, -1, -1) start_y = torch.gather(y_positions_expanded, dim=2, index=start_idx2.unsqueeze(2)).squeeze(2) end_y = torch.gather(y_positions_expanded, dim=2, index=(start_idx2 + 1).unsqueeze(2)).squeeze(2) cand = start_y + (x - start_x) * (end_y - start_y) / (end_x - start_x) return cand
A piecewise linear function y = f(x), using xp and yp as keypoints. We implement f(x) in a differentiable way (i.e. applicable for autograd). The function f(x) is well-defined for all x-axis. (For x beyond the bounds of xp, we use the outmost points of xp to define the linear function.) Args: x: PyTorch tensor with shape [N, C], where N is the batch size, C is the number of channels (we use C = 1 for DPM-Solver). xp: PyTorch tensor with shape [C, K], where K is the number of keypoints. yp: PyTorch tensor with shape [C, K]. Returns: The function values f(x), with shape [N, C].
14,046
import torch import torch.nn.functional as F import math The provided code snippet includes necessary dependencies for implementing the `expand_dims` function. Write a Python function `def expand_dims(v, dims)` to solve the following problem: Expand the tensor `v` to the dim `dims`. Args: `v`: a PyTorch tensor with shape [N]. `dim`: a `int`. Returns: a PyTorch tensor with shape [N, 1, 1, ..., 1] and the total dimension is `dims`. Here is the function: def expand_dims(v, dims): """ Expand the tensor `v` to the dim `dims`. Args: `v`: a PyTorch tensor with shape [N]. `dim`: a `int`. Returns: a PyTorch tensor with shape [N, 1, 1, ..., 1] and the total dimension is `dims`. """ return v[(...,) + (None,)*(dims - 1)]
Expand the tensor `v` to the dim `dims`. Args: `v`: a PyTorch tensor with shape [N]. `dim`: a `int`. Returns: a PyTorch tensor with shape [N, 1, 1, ..., 1] and the total dimension is `dims`.
14,047
from collections import deque from functools import partial from inspect import isfunction import torch.nn.functional as F import numpy as np from torch.nn import Conv1d from torch.nn import Mish import torch from torch import nn from tqdm import tqdm import math def exists(x): def default(val, d): if exists(val): return val return d() if isfunction(d) else d
null
14,048
from collections import deque from functools import partial from inspect import isfunction import torch.nn.functional as F import numpy as np from torch.nn import Conv1d from torch.nn import Mish import torch from torch import nn from tqdm import tqdm import math def extract(a, t): return a[t].reshape((1, 1, 1, 1))
null
14,049
from collections import deque from functools import partial from inspect import isfunction import torch.nn.functional as F import numpy as np from torch.nn import Conv1d from torch.nn import Mish import torch from torch import nn from tqdm import tqdm import math def noise_like(shape, device, repeat=False): repeat_noise = lambda: torch.randn((1, *shape[1:]), device=device).repeat(shape[0], *((1,) * (len(shape) - 1))) # NOQA noise = lambda: torch.randn(shape, device=device) # NOQA return repeat_noise() if repeat else noise()
null
14,050
from collections import deque from functools import partial from inspect import isfunction import torch.nn.functional as F import numpy as np from torch.nn import Conv1d from torch.nn import Mish import torch from torch import nn from tqdm import tqdm import math The provided code snippet includes necessary dependencies for implementing the `linear_beta_schedule` function. Write a Python function `def linear_beta_schedule(timesteps, max_beta=0.02)` to solve the following problem: linear schedule Here is the function: def linear_beta_schedule(timesteps, max_beta=0.02): """ linear schedule """ betas = np.linspace(1e-4, max_beta, timesteps) return betas
linear schedule
14,051
from collections import deque from functools import partial from inspect import isfunction import torch.nn.functional as F import numpy as np from torch.nn import Conv1d from torch.nn import Mish import torch from torch import nn from tqdm import tqdm import math The provided code snippet includes necessary dependencies for implementing the `cosine_beta_schedule` function. Write a Python function `def cosine_beta_schedule(timesteps, s=0.008)` to solve the following problem: cosine schedule as proposed in https://openreview.net/forum?id=-NEXDKk8gZ Here is the function: def cosine_beta_schedule(timesteps, s=0.008): """ cosine schedule as proposed in https://openreview.net/forum?id=-NEXDKk8gZ """ steps = timesteps + 1 x = np.linspace(0, steps, steps) alphas_cumprod = np.cos(((x / steps) + s) / (1 + s) * np.pi * 0.5) ** 2 alphas_cumprod = alphas_cumprod / alphas_cumprod[0] betas = 1 - (alphas_cumprod[1:] / alphas_cumprod[:-1]) return np.clip(betas, a_min=0, a_max=0.999)
cosine schedule as proposed in https://openreview.net/forum?id=-NEXDKk8gZ
14,052
from collections import deque from functools import partial from inspect import isfunction import torch.nn.functional as F import numpy as np from torch.nn import Conv1d from torch.nn import Mish import torch from torch import nn from tqdm import tqdm import math def extract_1(a, t): return a[t].reshape((1, 1, 1, 1))
null
14,053
from collections import deque from functools import partial from inspect import isfunction import torch.nn.functional as F import numpy as np from torch.nn import Conv1d from torch.nn import Mish import torch from torch import nn from tqdm import tqdm import math def predict_stage0(noise_pred, noise_pred_prev): return (noise_pred + noise_pred_prev) / 2
null
14,054
from collections import deque from functools import partial from inspect import isfunction import torch.nn.functional as F import numpy as np from torch.nn import Conv1d from torch.nn import Mish import torch from torch import nn from tqdm import tqdm import math def predict_stage1(noise_pred, noise_list): return (noise_pred * 3 - noise_list[-1]) / 2
null
14,055
from collections import deque from functools import partial from inspect import isfunction import torch.nn.functional as F import numpy as np from torch.nn import Conv1d from torch.nn import Mish import torch from torch import nn from tqdm import tqdm import math def predict_stage2(noise_pred, noise_list): return (noise_pred * 23 - noise_list[-1] * 16 + noise_list[-2] * 5) / 12
null
14,056
from collections import deque from functools import partial from inspect import isfunction import torch.nn.functional as F import numpy as np from torch.nn import Conv1d from torch.nn import Mish import torch from torch import nn from tqdm import tqdm import math def predict_stage3(noise_pred, noise_list): return (noise_pred * 55 - noise_list[-1] * 59 + noise_list[-2] * 37 - noise_list[-3] * 9) / 24
null
14,057
import os import random import re import numpy as np import librosa import torch from tqdm import tqdm from torch.utils.data import Dataset def traverse_dir(root_dir, extensions, amount=None, str_include=None, str_exclude=None, is_pure=False, is_sort=False, is_ext=True): file_list = [] cnt = 0 for root, _, files in os.walk(root_dir): for file in files: if any([file.endswith(f".{ext}") for ext in extensions]): # path mix_path = os.path.join(root, file) pure_path = mix_path[len(root_dir) + 1 :] if is_pure else mix_path # amount if (amount is not None) and (cnt == amount): if is_sort: file_list.sort() return file_list # check string if (str_include is not None) and (str_include not in pure_path): continue if (str_exclude is not None) and (str_exclude in pure_path): continue if not is_ext: ext = pure_path.split(".")[-1] pure_path = pure_path[: -(len(ext) + 1)] file_list.append(pure_path) cnt += 1 if is_sort: file_list.sort() return file_list
null
14,058
import os import random import re import numpy as np import librosa import torch from tqdm import tqdm from torch.utils.data import Dataset class AudioDataset(Dataset): def __init__( self, path_root, waveform_sec, hop_size, sample_rate, load_all_data=True, whole_audio=False, extensions=["wav"], n_spk=1, device="cpu", fp16=False, use_aug=False, ): def __getitem__(self, file_idx): def get_data(self, name_ext, data_buffer): def __len__(self): def get_data_loaders(args, whole_audio=False): data_train = AudioDataset(args.data.train_path, waveform_sec=args.data.duration, hop_size=args.data.block_size, sample_rate=args.data.sampling_rate, load_all_data=args.train.cache_all_data, whole_audio=whole_audio, extensions=args.data.extensions, n_spk=args.model.n_spk, device=args.train.cache_device, fp16=args.train.cache_fp16, use_aug=True) loader_train = torch.utils.data.DataLoader(data_train, batch_size=args.train.batch_size if not whole_audio else 1, shuffle=True, num_workers=args.train.num_workers if args.train.cache_device == "cpu" else 0, persistent_workers=(args.train.num_workers > 0) if args.train.cache_device == "cpu" else False, pin_memory=True if args.train.cache_device == "cpu" else False) data_valid = AudioDataset(args.data.valid_path, waveform_sec=args.data.duration, hop_size=args.data.block_size, sample_rate=args.data.sampling_rate, load_all_data=args.train.cache_all_data, whole_audio=True, extensions=args.data.extensions, n_spk=args.model.n_spk) loader_valid = torch.utils.data.DataLoader(data_valid, batch_size=1, shuffle=False, num_workers=0, pin_memory=True) return loader_train, loader_valid
null
14,059
from collections import deque from functools import partial from inspect import isfunction import torch.nn.functional as F import numpy as np import torch from torch import nn from tqdm import tqdm def exists(x): def default(val, d): if exists(val): return val return d() if isfunction(d) else d
null
14,060
from collections import deque from functools import partial from inspect import isfunction import torch.nn.functional as F import numpy as np import torch from torch import nn from tqdm import tqdm def extract(a, t, x_shape): b, *_ = t.shape out = a.gather(-1, t) return out.reshape(b, *((1,) * (len(x_shape) - 1)))
null
14,061
from collections import deque from functools import partial from inspect import isfunction import torch.nn.functional as F import numpy as np import torch from torch import nn from tqdm import tqdm def noise_like(shape, device, repeat=False): repeat_noise = lambda: torch.randn((1, *shape[1:]), device=device).repeat(shape[0], *((1,) * (len(shape) - 1))) # NOQA noise = lambda: torch.randn(shape, device=device) # NOQA return repeat_noise() if repeat else noise()
null
14,062
from collections import deque from functools import partial from inspect import isfunction import torch.nn.functional as F import numpy as np import torch from torch import nn from tqdm import tqdm The provided code snippet includes necessary dependencies for implementing the `linear_beta_schedule` function. Write a Python function `def linear_beta_schedule(timesteps, max_beta=0.02)` to solve the following problem: linear schedule Here is the function: def linear_beta_schedule(timesteps, max_beta=0.02): """ linear schedule """ betas = np.linspace(1e-4, max_beta, timesteps) return betas
linear schedule
14,063
from collections import deque from functools import partial from inspect import isfunction import torch.nn.functional as F import numpy as np import torch from torch import nn from tqdm import tqdm The provided code snippet includes necessary dependencies for implementing the `cosine_beta_schedule` function. Write a Python function `def cosine_beta_schedule(timesteps, s=0.008)` to solve the following problem: cosine schedule as proposed in https://openreview.net/forum?id=-NEXDKk8gZ Here is the function: def cosine_beta_schedule(timesteps, s=0.008): """ cosine schedule as proposed in https://openreview.net/forum?id=-NEXDKk8gZ """ steps = timesteps + 1 x = np.linspace(0, steps, steps) alphas_cumprod = np.cos(((x / steps) + s) / (1 + s) * np.pi * 0.5) ** 2 alphas_cumprod = alphas_cumprod / alphas_cumprod[0] betas = 1 - (alphas_cumprod[1:] / alphas_cumprod[:-1]) return np.clip(betas, a_min=0, a_max=0.999)
cosine schedule as proposed in https://openreview.net/forum?id=-NEXDKk8gZ
14,064
from diffusion_onnx import GaussianDiffusion import os import yaml import torch import torch.nn as nn import numpy as np from wavenet import WaveNet import torch.nn.functional as F import diffusion class DotDict(dict): def __getattr__(*args): val = dict.get(*args) return DotDict(val) if type(val) is dict else val __setattr__ = dict.__setitem__ __delattr__ = dict.__delitem__ class Unit2Mel(nn.Module): def __init__( self, input_channel, n_spk, use_pitch_aug=False, out_dims=128, n_layers=20, n_chans=384, n_hidden=256): super().__init__() self.unit_embed = nn.Linear(input_channel, n_hidden) self.f0_embed = nn.Linear(1, n_hidden) self.volume_embed = nn.Linear(1, n_hidden) if use_pitch_aug: self.aug_shift_embed = nn.Linear(1, n_hidden, bias=False) else: self.aug_shift_embed = None self.n_spk = n_spk if n_spk is not None and n_spk > 1: self.spk_embed = nn.Embedding(n_spk, n_hidden) # diffusion self.decoder = GaussianDiffusion(out_dims, n_layers, n_chans, n_hidden) self.hidden_size = n_hidden self.speaker_map = torch.zeros((self.n_spk,1,1,n_hidden)) def forward(self, units, mel2ph, f0, volume, g = None): ''' input: B x n_frames x n_unit return: dict of B x n_frames x feat ''' decoder_inp = F.pad(units, [0, 0, 1, 0]) mel2ph_ = mel2ph.unsqueeze(2).repeat([1, 1, units.shape[-1]]) units = torch.gather(decoder_inp, 1, mel2ph_) # [B, T, H] x = self.unit_embed(units) + self.f0_embed((1 + f0.unsqueeze(-1) / 700).log()) + self.volume_embed(volume.unsqueeze(-1)) if self.n_spk is not None and self.n_spk > 1: # [N, S] * [S, B, 1, H] g = g.reshape((g.shape[0], g.shape[1], 1, 1, 1)) # [N, S, B, 1, 1] g = g * self.speaker_map # [N, S, B, 1, H] g = torch.sum(g, dim=1) # [N, 1, B, 1, H] g = g.transpose(0, -1).transpose(0, -2).squeeze(0) # [B, H, N] x = x.transpose(1, 2) + g return x else: return x.transpose(1, 2) def init_spkembed(self, units, f0, volume, spk_id = None, spk_mix_dict = None, aug_shift = None, gt_spec=None, infer=True, infer_speedup=10, method='dpm-solver', k_step=300, use_tqdm=True): ''' input: B x n_frames x n_unit return: dict of B x n_frames x feat ''' x = self.unit_embed(units) + self.f0_embed((1+ f0 / 700).log()) + self.volume_embed(volume) if self.n_spk is not None and self.n_spk > 1: if spk_mix_dict is not None: spk_embed_mix = torch.zeros((1,1,self.hidden_size)) for k, v in spk_mix_dict.items(): spk_id_torch = torch.LongTensor(np.array([[k]])).to(units.device) spk_embeddd = self.spk_embed(spk_id_torch) self.speaker_map[k] = spk_embeddd spk_embed_mix = spk_embed_mix + v * spk_embeddd x = x + spk_embed_mix else: x = x + self.spk_embed(spk_id - 1) self.speaker_map = self.speaker_map.unsqueeze(0) self.speaker_map = self.speaker_map.detach() return x.transpose(1, 2) def OnnxExport(self, project_name=None, init_noise=None, export_encoder=True, export_denoise=True, export_pred=True, export_after=True): hubert_hidden_size = 768 n_frames = 100 hubert = torch.randn((1, n_frames, hubert_hidden_size)) mel2ph = torch.arange(end=n_frames).unsqueeze(0).long() f0 = torch.randn((1, n_frames)) volume = torch.randn((1, n_frames)) spk_mix = [] spks = {} if self.n_spk is not None and self.n_spk > 1: for i in range(self.n_spk): spk_mix.append(1.0/float(self.n_spk)) spks.update({i:1.0/float(self.n_spk)}) spk_mix = torch.tensor(spk_mix) spk_mix = spk_mix.repeat(n_frames, 1) orgouttt = self.init_spkembed(hubert, f0.unsqueeze(-1), volume.unsqueeze(-1), spk_mix_dict=spks) outtt = self.forward(hubert, mel2ph, f0, volume, spk_mix) if export_encoder: torch.onnx.export( self, (hubert, mel2ph, f0, volume, spk_mix), f"{project_name}_encoder.onnx", input_names=["hubert", "mel2ph", "f0", "volume", "spk_mix"], output_names=["mel_pred"], dynamic_axes={ "hubert": [1], "f0": [1], "volume": [1], "mel2ph": [1], "spk_mix": [0], }, opset_version=16 ) self.decoder.OnnxExport(project_name, init_noise=init_noise, export_denoise=export_denoise, export_pred=export_pred, export_after=export_after) def ExportOnnx(self, project_name=None): hubert_hidden_size = 768 n_frames = 100 hubert = torch.randn((1, n_frames, hubert_hidden_size)) mel2ph = torch.arange(end=n_frames).unsqueeze(0).long() f0 = torch.randn((1, n_frames)) volume = torch.randn((1, n_frames)) spk_mix = [] spks = {} if self.n_spk is not None and self.n_spk > 1: for i in range(self.n_spk): spk_mix.append(1.0/float(self.n_spk)) spks.update({i:1.0/float(self.n_spk)}) spk_mix = torch.tensor(spk_mix) orgouttt = self.orgforward(hubert, f0.unsqueeze(-1), volume.unsqueeze(-1), spk_mix_dict=spks) outtt = self.forward(hubert, mel2ph, f0, volume, spk_mix) torch.onnx.export( self, (hubert, mel2ph, f0, volume, spk_mix), f"{project_name}_encoder.onnx", input_names=["hubert", "mel2ph", "f0", "volume", "spk_mix"], output_names=["mel_pred"], dynamic_axes={ "hubert": [1], "f0": [1], "volume": [1], "mel2ph": [1] }, opset_version=16 ) condition = torch.randn(1,self.decoder.n_hidden,n_frames) noise = torch.randn((1, 1, self.decoder.mel_bins, condition.shape[2]), dtype=torch.float32) pndm_speedup = torch.LongTensor([100]) K_steps = torch.LongTensor([1000]) self.decoder = torch.jit.script(self.decoder) self.decoder(condition, noise, pndm_speedup, K_steps) torch.onnx.export( self.decoder, (condition, noise, pndm_speedup, K_steps), f"{project_name}_diffusion.onnx", input_names=["condition", "noise", "pndm_speedup", "K_steps"], output_names=["mel"], dynamic_axes={ "condition": [2], "noise": [3], }, opset_version=16 ) def load_model_vocoder( model_path, device='cpu'): config_file = os.path.join(os.path.split(model_path)[0], 'config.yaml') with open(config_file, "r") as config: args = yaml.safe_load(config) args = DotDict(args) # load model model = Unit2Mel( args.data.encoder_out_channels, args.model.n_spk, args.model.use_pitch_aug, 128, args.model.n_layers, args.model.n_chans, args.model.n_hidden) print(' [Loading] ' + model_path) ckpt = torch.load(model_path, map_location=torch.device(device)) model.to(device) model.load_state_dict(ckpt['model']) model.eval() return model, args
null
14,065
import os import yaml import torch import torch.nn as nn import numpy as np from .diffusion import GaussianDiffusion from .wavenet import WaveNet from .vocoder import Vocoder class DotDict(dict): def __getattr__(*args): # type: ignore val = dict.get(*args) return DotDict(val) if type(val) is dict else val __setattr__ = dict.__setitem__ # type: ignore __delattr__ = dict.__delitem__ # type: ignore class Unit2Mel(nn.Module): def __init__(self, input_channel, n_spk, use_pitch_aug=False, out_dims=128, n_layers=20, n_chans=384, n_hidden=256): super().__init__() self.unit_embed = nn.Linear(input_channel, n_hidden) self.f0_embed = nn.Linear(1, n_hidden) self.volume_embed = nn.Linear(1, n_hidden) if use_pitch_aug: self.aug_shift_embed = nn.Linear(1, n_hidden, bias=False) else: self.aug_shift_embed = None self.n_spk = n_spk if n_spk is not None and n_spk > 1: self.spk_embed = nn.Embedding(n_spk, n_hidden) # diffusion self.decoder = GaussianDiffusion(WaveNet(out_dims, n_layers, n_chans, n_hidden), out_dims=out_dims) def forward(self, units, f0, volume, spk_id=None, spk_mix_dict=None, aug_shift=None, gt_spec=None, infer=True, infer_speedup=10, method="dpm-solver", k_step=300, use_tqdm=True): """ input: B x n_frames x n_unit return: dict of B x n_frames x feat """ x = self.unit_embed(units) + self.f0_embed((1 + f0 / 700).log()) + self.volume_embed(volume) if self.n_spk is not None and self.n_spk > 1: if spk_mix_dict is not None: for k, v in spk_mix_dict.items(): spk_id_torch = torch.LongTensor(np.array([[k]])).to(units.device) x = x + v * self.spk_embed(spk_id_torch - 1) else: x = x + self.spk_embed(spk_id - 1) if self.aug_shift_embed is not None and aug_shift is not None: x = x + self.aug_shift_embed(aug_shift / 5) x = self.decoder(x, gt_spec=gt_spec, infer=infer, infer_speedup=infer_speedup, method=method, k_step=k_step, use_tqdm=use_tqdm) return x class Vocoder: def __init__(self, vocoder_type, vocoder_ckpt, device=None): if device is None: device = "cuda" if torch.cuda.is_available() else "cpu" self.device = device if vocoder_type == "nsf-hifigan": self.vocoder = NsfHifiGAN(vocoder_ckpt, device=device) elif vocoder_type == "nsf-hifigan-log10": self.vocoder = NsfHifiGANLog10(vocoder_ckpt, device=device) else: raise ValueError(f" [x] Unknown vocoder: {vocoder_type}") self.resample_kernel = {} self.vocoder_sample_rate = self.vocoder.sample_rate() self.vocoder_hop_size = self.vocoder.hop_size() self.dimension = self.vocoder.dimension() def extract(self, audio, sample_rate, keyshift=0): # resample if sample_rate == self.vocoder_sample_rate: audio_res = audio else: key_str = str(sample_rate) if key_str not in self.resample_kernel: self.resample_kernel[key_str] = Resample(sample_rate, self.vocoder_sample_rate, lowpass_filter_width=128).to(self.device) audio_res = self.resample_kernel[key_str](audio) # extract mel = self.vocoder.extract(audio_res, keyshift=keyshift) # B, n_frames, bins return mel def infer(self, mel, f0): f0 = f0[:, : mel.size(1), 0] # B, n_frames audio = self.vocoder(mel, f0) return audio def load_model_vocoder(model_path, device="cpu"): config_file = os.path.join(os.path.split(model_path)[0], "config.yaml") with open(config_file, "r") as config: args = yaml.safe_load(config) args = DotDict(args) # load vocoder vocoder = Vocoder(args.vocoder.type, args.vocoder.ckpt, device=device) # load model model = Unit2Mel(args.data.encoder_out_channels, args.model.n_spk, args.model.use_pitch_aug, vocoder.dimension, args.model.n_layers, args.model.n_chans, args.model.n_hidden) print(" [Loading] " + model_path) ckpt = torch.load(model_path, map_location=torch.device(device)) model.to(device) model.load_state_dict(ckpt["model"]) model.eval() return model, vocoder, args
null
14,066
import torch def expand_dims(v, dims): """ Expand the tensor `v` to the dim `dims`. Args: `v`: a PyTorch tensor with shape [N]. `dim`: a `int`. Returns: a PyTorch tensor with shape [N, 1, 1, ..., 1] and the total dimension is `dims`. """ return v[(...,) + (None,) * (dims - 1)] The provided code snippet includes necessary dependencies for implementing the `model_wrapper` function. Write a Python function `def model_wrapper( model, noise_schedule, model_type="noise", model_kwargs={}, guidance_type="uncond", condition=None, unconditional_condition=None, guidance_scale=1.0, classifier_fn=None, classifier_kwargs={}, )` to solve the following problem: Create a wrapper function for the noise prediction model. DPM-Solver needs to solve the continuous-time diffusion ODEs. For DPMs trained on discrete-time labels, we need to firstly wrap the model function to a noise prediction model that accepts the continuous time as the input. We support four types of the diffusion model by setting `model_type`: 1. "noise": noise prediction model. (Trained by predicting noise). 2. "x_start": data prediction model. (Trained by predicting the data x_0 at time 0). 3. "v": velocity prediction model. (Trained by predicting the velocity). The "v" prediction is derivation detailed in Appendix D of [1], and is used in Imagen-Video [2]. [1] Salimans, Tim, and Jonathan Ho. "Progressive distillation for fast sampling of diffusion models." arXiv preprint arXiv:2202.00512 (2022). [2] Ho, Jonathan, et al. "Imagen Video: High Definition Video Generation with Diffusion Models." arXiv preprint arXiv:2210.02303 (2022). 4. "score": marginal score function. (Trained by denoising score matching). Note that the score function and the noise prediction model follows a simple relationship: ``` noise(x_t, t) = -sigma_t * score(x_t, t) ``` We support three types of guided sampling by DPMs by setting `guidance_type`: 1. "uncond": unconditional sampling by DPMs. The input `model` has the following format: `` model(x, t_input, **model_kwargs) -> noise | x_start | v | score `` 2. "classifier": classifier guidance sampling [3] by DPMs and another classifier. The input `model` has the following format: `` model(x, t_input, **model_kwargs) -> noise | x_start | v | score `` The input `classifier_fn` has the following format: `` classifier_fn(x, t_input, cond, **classifier_kwargs) -> logits(x, t_input, cond) `` [3] P. Dhariwal and A. Q. Nichol, "Diffusion models beat GANs on image synthesis," in Advances in Neural Information Processing Systems, vol. 34, 2021, pp. 8780-8794. 3. "classifier-free": classifier-free guidance sampling by conditional DPMs. The input `model` has the following format: `` model(x, t_input, cond, **model_kwargs) -> noise | x_start | v | score `` And if cond == `unconditional_condition`, the model output is the unconditional DPM output. [4] Ho, Jonathan, and Tim Salimans. "Classifier-free diffusion guidance." arXiv preprint arXiv:2207.12598 (2022). The `t_input` is the time label of the model, which may be discrete-time labels (i.e. 0 to 999) or continuous-time labels (i.e. epsilon to T). We wrap the model function to accept only `x` and `t_continuous` as inputs, and outputs the predicted noise: `` def model_fn(x, t_continuous) -> noise: t_input = get_model_input_time(t_continuous) return noise_pred(model, x, t_input, **model_kwargs) `` where `t_continuous` is the continuous time labels (i.e. epsilon to T). And we use `model_fn` for DPM-Solver. =============================================================== Args: model: A diffusion model with the corresponding format described above. noise_schedule: A noise schedule object, such as NoiseScheduleVP. model_type: A `str`. The parameterization type of the diffusion model. "noise" or "x_start" or "v" or "score". model_kwargs: A `dict`. A dict for the other inputs of the model function. guidance_type: A `str`. The type of the guidance for sampling. "uncond" or "classifier" or "classifier-free". condition: A pytorch tensor. The condition for the guided sampling. Only used for "classifier" or "classifier-free" guidance type. unconditional_condition: A pytorch tensor. The condition for the unconditional sampling. Only used for "classifier-free" guidance type. guidance_scale: A `float`. The scale for the guided sampling. classifier_fn: A classifier function. Only used for the classifier guidance. classifier_kwargs: A `dict`. A dict for the other inputs of the classifier function. Returns: A noise prediction model that accepts the noised data and the continuous time as the inputs. Here is the function: def model_wrapper( model, noise_schedule, model_type="noise", model_kwargs={}, guidance_type="uncond", condition=None, unconditional_condition=None, guidance_scale=1.0, classifier_fn=None, classifier_kwargs={}, ): """Create a wrapper function for the noise prediction model. DPM-Solver needs to solve the continuous-time diffusion ODEs. For DPMs trained on discrete-time labels, we need to firstly wrap the model function to a noise prediction model that accepts the continuous time as the input. We support four types of the diffusion model by setting `model_type`: 1. "noise": noise prediction model. (Trained by predicting noise). 2. "x_start": data prediction model. (Trained by predicting the data x_0 at time 0). 3. "v": velocity prediction model. (Trained by predicting the velocity). The "v" prediction is derivation detailed in Appendix D of [1], and is used in Imagen-Video [2]. [1] Salimans, Tim, and Jonathan Ho. "Progressive distillation for fast sampling of diffusion models." arXiv preprint arXiv:2202.00512 (2022). [2] Ho, Jonathan, et al. "Imagen Video: High Definition Video Generation with Diffusion Models." arXiv preprint arXiv:2210.02303 (2022). 4. "score": marginal score function. (Trained by denoising score matching). Note that the score function and the noise prediction model follows a simple relationship: ``` noise(x_t, t) = -sigma_t * score(x_t, t) ``` We support three types of guided sampling by DPMs by setting `guidance_type`: 1. "uncond": unconditional sampling by DPMs. The input `model` has the following format: `` model(x, t_input, **model_kwargs) -> noise | x_start | v | score `` 2. "classifier": classifier guidance sampling [3] by DPMs and another classifier. The input `model` has the following format: `` model(x, t_input, **model_kwargs) -> noise | x_start | v | score `` The input `classifier_fn` has the following format: `` classifier_fn(x, t_input, cond, **classifier_kwargs) -> logits(x, t_input, cond) `` [3] P. Dhariwal and A. Q. Nichol, "Diffusion models beat GANs on image synthesis," in Advances in Neural Information Processing Systems, vol. 34, 2021, pp. 8780-8794. 3. "classifier-free": classifier-free guidance sampling by conditional DPMs. The input `model` has the following format: `` model(x, t_input, cond, **model_kwargs) -> noise | x_start | v | score `` And if cond == `unconditional_condition`, the model output is the unconditional DPM output. [4] Ho, Jonathan, and Tim Salimans. "Classifier-free diffusion guidance." arXiv preprint arXiv:2207.12598 (2022). The `t_input` is the time label of the model, which may be discrete-time labels (i.e. 0 to 999) or continuous-time labels (i.e. epsilon to T). We wrap the model function to accept only `x` and `t_continuous` as inputs, and outputs the predicted noise: `` def model_fn(x, t_continuous) -> noise: t_input = get_model_input_time(t_continuous) return noise_pred(model, x, t_input, **model_kwargs) `` where `t_continuous` is the continuous time labels (i.e. epsilon to T). And we use `model_fn` for DPM-Solver. =============================================================== Args: model: A diffusion model with the corresponding format described above. noise_schedule: A noise schedule object, such as NoiseScheduleVP. model_type: A `str`. The parameterization type of the diffusion model. "noise" or "x_start" or "v" or "score". model_kwargs: A `dict`. A dict for the other inputs of the model function. guidance_type: A `str`. The type of the guidance for sampling. "uncond" or "classifier" or "classifier-free". condition: A pytorch tensor. The condition for the guided sampling. Only used for "classifier" or "classifier-free" guidance type. unconditional_condition: A pytorch tensor. The condition for the unconditional sampling. Only used for "classifier-free" guidance type. guidance_scale: A `float`. The scale for the guided sampling. classifier_fn: A classifier function. Only used for the classifier guidance. classifier_kwargs: A `dict`. A dict for the other inputs of the classifier function. Returns: A noise prediction model that accepts the noised data and the continuous time as the inputs. """ def get_model_input_time(t_continuous): """ Convert the continuous-time `t_continuous` (in [epsilon, T]) to the model input time. For discrete-time DPMs, we convert `t_continuous` in [1 / N, 1] to `t_input` in [0, 1000 * (N - 1) / N]. For continuous-time DPMs, we just use `t_continuous`. """ if noise_schedule.schedule == "discrete": return (t_continuous - 1.0 / noise_schedule.total_N) * noise_schedule.total_N else: return t_continuous def noise_pred_fn(x, t_continuous, cond=None): t_input = get_model_input_time(t_continuous) if cond is None: output = model(x, t_input, **model_kwargs) else: output = model(x, t_input, cond, **model_kwargs) if model_type == "noise": return output elif model_type == "x_start": alpha_t, sigma_t = noise_schedule.marginal_alpha(t_continuous), noise_schedule.marginal_std(t_continuous) return (x - expand_dims(alpha_t, x.dim()) * output) / expand_dims(sigma_t, x.dim()) elif model_type == "v": alpha_t, sigma_t = noise_schedule.marginal_alpha(t_continuous), noise_schedule.marginal_std(t_continuous) return expand_dims(alpha_t, x.dim()) * output + expand_dims(sigma_t, x.dim()) * x elif model_type == "score": sigma_t = noise_schedule.marginal_std(t_continuous) return -expand_dims(sigma_t, x.dim()) * output def cond_grad_fn(x, t_input): """ Compute the gradient of the classifier, i.e. nabla_{x} log p_t(cond | x_t). """ with torch.enable_grad(): x_in = x.detach().requires_grad_(True) log_prob = classifier_fn(x_in, t_input, condition, **classifier_kwargs) return torch.autograd.grad(log_prob.sum(), x_in)[0] def model_fn(x, t_continuous): """ The noise predicition model function that is used for DPM-Solver. """ if guidance_type == "uncond": return noise_pred_fn(x, t_continuous) elif guidance_type == "classifier": assert classifier_fn is not None t_input = get_model_input_time(t_continuous) cond_grad = cond_grad_fn(x, t_input) sigma_t = noise_schedule.marginal_std(t_continuous) noise = noise_pred_fn(x, t_continuous) return noise - guidance_scale * expand_dims(sigma_t, x.dim()) * cond_grad elif guidance_type == "classifier-free": if guidance_scale == 1.0 or unconditional_condition is None: return noise_pred_fn(x, t_continuous, cond=condition) else: x_in = torch.cat([x] * 2) t_in = torch.cat([t_continuous] * 2) c_in = torch.cat([unconditional_condition, condition]) noise_uncond, noise = noise_pred_fn(x_in, t_in, cond=c_in).chunk(2) return noise_uncond + guidance_scale * (noise - noise_uncond) assert model_type in ["noise", "x_start", "v", "score"] assert guidance_type in ["uncond", "classifier", "classifier-free"] return model_fn
Create a wrapper function for the noise prediction model. DPM-Solver needs to solve the continuous-time diffusion ODEs. For DPMs trained on discrete-time labels, we need to firstly wrap the model function to a noise prediction model that accepts the continuous time as the input. We support four types of the diffusion model by setting `model_type`: 1. "noise": noise prediction model. (Trained by predicting noise). 2. "x_start": data prediction model. (Trained by predicting the data x_0 at time 0). 3. "v": velocity prediction model. (Trained by predicting the velocity). The "v" prediction is derivation detailed in Appendix D of [1], and is used in Imagen-Video [2]. [1] Salimans, Tim, and Jonathan Ho. "Progressive distillation for fast sampling of diffusion models." arXiv preprint arXiv:2202.00512 (2022). [2] Ho, Jonathan, et al. "Imagen Video: High Definition Video Generation with Diffusion Models." arXiv preprint arXiv:2210.02303 (2022). 4. "score": marginal score function. (Trained by denoising score matching). Note that the score function and the noise prediction model follows a simple relationship: ``` noise(x_t, t) = -sigma_t * score(x_t, t) ``` We support three types of guided sampling by DPMs by setting `guidance_type`: 1. "uncond": unconditional sampling by DPMs. The input `model` has the following format: `` model(x, t_input, **model_kwargs) -> noise | x_start | v | score `` 2. "classifier": classifier guidance sampling [3] by DPMs and another classifier. The input `model` has the following format: `` model(x, t_input, **model_kwargs) -> noise | x_start | v | score `` The input `classifier_fn` has the following format: `` classifier_fn(x, t_input, cond, **classifier_kwargs) -> logits(x, t_input, cond) `` [3] P. Dhariwal and A. Q. Nichol, "Diffusion models beat GANs on image synthesis," in Advances in Neural Information Processing Systems, vol. 34, 2021, pp. 8780-8794. 3. "classifier-free": classifier-free guidance sampling by conditional DPMs. The input `model` has the following format: `` model(x, t_input, cond, **model_kwargs) -> noise | x_start | v | score `` And if cond == `unconditional_condition`, the model output is the unconditional DPM output. [4] Ho, Jonathan, and Tim Salimans. "Classifier-free diffusion guidance." arXiv preprint arXiv:2207.12598 (2022). The `t_input` is the time label of the model, which may be discrete-time labels (i.e. 0 to 999) or continuous-time labels (i.e. epsilon to T). We wrap the model function to accept only `x` and `t_continuous` as inputs, and outputs the predicted noise: `` def model_fn(x, t_continuous) -> noise: t_input = get_model_input_time(t_continuous) return noise_pred(model, x, t_input, **model_kwargs) `` where `t_continuous` is the continuous time labels (i.e. epsilon to T). And we use `model_fn` for DPM-Solver. =============================================================== Args: model: A diffusion model with the corresponding format described above. noise_schedule: A noise schedule object, such as NoiseScheduleVP. model_type: A `str`. The parameterization type of the diffusion model. "noise" or "x_start" or "v" or "score". model_kwargs: A `dict`. A dict for the other inputs of the model function. guidance_type: A `str`. The type of the guidance for sampling. "uncond" or "classifier" or "classifier-free". condition: A pytorch tensor. The condition for the guided sampling. Only used for "classifier" or "classifier-free" guidance type. unconditional_condition: A pytorch tensor. The condition for the unconditional sampling. Only used for "classifier-free" guidance type. guidance_scale: A `float`. The scale for the guided sampling. classifier_fn: A classifier function. Only used for the classifier guidance. classifier_kwargs: A `dict`. A dict for the other inputs of the classifier function. Returns: A noise prediction model that accepts the noised data and the continuous time as the inputs.
14,067
import torch The provided code snippet includes necessary dependencies for implementing the `interpolate_fn` function. Write a Python function `def interpolate_fn(x, xp, yp)` to solve the following problem: A piecewise linear function y = f(x), using xp and yp as keypoints. We implement f(x) in a differentiable way (i.e. applicable for autograd). The function f(x) is well-defined for all x-axis. (For x beyond the bounds of xp, we use the outmost points of xp to define the linear function.) Args: x: PyTorch tensor with shape [N, C], where N is the batch size, C is the number of channels (we use C = 1 for DPM-Solver). xp: PyTorch tensor with shape [C, K], where K is the number of keypoints. yp: PyTorch tensor with shape [C, K]. Returns: The function values f(x), with shape [N, C]. Here is the function: def interpolate_fn(x, xp, yp): """ A piecewise linear function y = f(x), using xp and yp as keypoints. We implement f(x) in a differentiable way (i.e. applicable for autograd). The function f(x) is well-defined for all x-axis. (For x beyond the bounds of xp, we use the outmost points of xp to define the linear function.) Args: x: PyTorch tensor with shape [N, C], where N is the batch size, C is the number of channels (we use C = 1 for DPM-Solver). xp: PyTorch tensor with shape [C, K], where K is the number of keypoints. yp: PyTorch tensor with shape [C, K]. Returns: The function values f(x), with shape [N, C]. """ N, K = x.shape[0], xp.shape[1] all_x = torch.cat([x.unsqueeze(2), xp.unsqueeze(0).repeat((N, 1, 1))], dim=2) sorted_all_x, x_indices = torch.sort(all_x, dim=2) x_idx = torch.argmin(x_indices, dim=2) cand_start_idx = x_idx - 1 start_idx = torch.where( torch.eq(x_idx, 0), torch.tensor(1, device=x.device), torch.where( torch.eq(x_idx, K), torch.tensor(K - 2, device=x.device), cand_start_idx, ), ) end_idx = torch.where(torch.eq(start_idx, cand_start_idx), start_idx + 2, start_idx + 1) start_x = torch.gather(sorted_all_x, dim=2, index=start_idx.unsqueeze(2)).squeeze(2) end_x = torch.gather(sorted_all_x, dim=2, index=end_idx.unsqueeze(2)).squeeze(2) start_idx2 = torch.where( torch.eq(x_idx, 0), torch.tensor(0, device=x.device), torch.where( torch.eq(x_idx, K), torch.tensor(K - 2, device=x.device), cand_start_idx, ), ) y_positions_expanded = yp.unsqueeze(0).expand(N, -1, -1) start_y = torch.gather(y_positions_expanded, dim=2, index=start_idx2.unsqueeze(2)).squeeze(2) end_y = torch.gather(y_positions_expanded, dim=2, index=(start_idx2 + 1).unsqueeze(2)).squeeze(2) cand = start_y + (x - start_x) * (end_y - start_y) / (end_x - start_x) return cand
A piecewise linear function y = f(x), using xp and yp as keypoints. We implement f(x) in a differentiable way (i.e. applicable for autograd). The function f(x) is well-defined for all x-axis. (For x beyond the bounds of xp, we use the outmost points of xp to define the linear function.) Args: x: PyTorch tensor with shape [N, C], where N is the batch size, C is the number of channels (we use C = 1 for DPM-Solver). xp: PyTorch tensor with shape [C, K], where K is the number of keypoints. yp: PyTorch tensor with shape [C, K]. Returns: The function values f(x), with shape [N, C].
14,068
import torch import torch.nn as nn from torch.nn import functional as F import math import numpy as np def MaskedAvgPool1d(x, kernel_size): x = x.unsqueeze(1) x = F.pad(x, ((kernel_size - 1) // 2, kernel_size // 2), mode="reflect") mask = ~torch.isnan(x) masked_x = torch.where(mask, x, torch.zeros_like(x)) ones_kernel = torch.ones(x.size(1), 1, kernel_size, device=x.device) # Perform sum pooling sum_pooled = F.conv1d( masked_x, ones_kernel, stride=1, padding=0, groups=x.size(1), ) # Count the non-masked (valid) elements in each pooling window valid_count = F.conv1d( mask.float(), ones_kernel, stride=1, padding=0, groups=x.size(1), ) valid_count = valid_count.clamp(min=1) # Avoid division by zero # Perform masked average pooling avg_pooled = sum_pooled / valid_count return avg_pooled.squeeze(1)
null
14,069
import torch import torch.nn as nn from torch.nn import functional as F import math import numpy as np def MedianPool1d(x, kernel_size): x = x.unsqueeze(1) x = F.pad(x, ((kernel_size - 1) // 2, kernel_size // 2), mode="reflect") x = x.squeeze(1) x = x.unfold(1, kernel_size, 1) x, _ = torch.sort(x, dim=-1) return x[:, :, (kernel_size - 1) // 2]
null
14,070
import torch import torch.nn as nn from torch.nn import functional as F import math import numpy as np def upsample(signal, factor): signal = signal.permute(0, 2, 1) signal = nn.functional.interpolate(torch.cat((signal,signal[:,:,-1:]),2), size=signal.shape[-1] * factor + 1, mode='linear', align_corners=True) signal = signal[:,:,:-1] return signal.permute(0, 2, 1)
null
14,071
import torch import torch.nn as nn from torch.nn import functional as F import math import numpy as np def remove_above_fmax(amplitudes, pitch, fmax, level_start=1): n_harm = amplitudes.shape[-1] pitches = pitch * torch.arange(level_start, n_harm + level_start).to(pitch) aa = (pitches < fmax).float() + 1e-7 return amplitudes * aa
null
14,072
import torch import torch.nn as nn from torch.nn import functional as F import math import numpy as np def fft_convolve(audio, impulse_response): def frequency_impulse_response(magnitudes, hann_window = True, half_width_frames = None): def frequency_filter(audio, magnitudes, hann_window=True, half_width_frames=None): impulse_response = frequency_impulse_response(magnitudes, hann_window, half_width_frames) return fft_convolve(audio, impulse_response)
null
14,073
import torch from torch import nn import math from functools import partial from einops import rearrange, repeat from local_attention import LocalAttention import torch.nn.functional as F def softmax_kernel(data, *, projection_matrix, is_query, normalize_data=True, eps=1e-4, device = None): b, h, *_ = data.shape # (batch size, head, length, model_dim) # normalize model dim data_normalizer = (data.shape[-1] ** -0.25) if normalize_data else 1. # what is ration?, projection_matrix.shape[0] --> 266 ratio = (projection_matrix.shape[0] ** -0.5) projection = repeat(projection_matrix, 'j d -> b h j d', b = b, h = h) projection = projection.type_as(data) #data_dash = w^T x data_dash = torch.einsum('...id,...jd->...ij', (data_normalizer * data), projection) # diag_data = D**2 diag_data = data ** 2 diag_data = torch.sum(diag_data, dim=-1) diag_data = (diag_data / 2.0) * (data_normalizer ** 2) diag_data = diag_data.unsqueeze(dim=-1) #print () if is_query: data_dash = ratio * ( torch.exp(data_dash - diag_data - torch.max(data_dash, dim=-1, keepdim=True).values) + eps) else: data_dash = ratio * ( torch.exp(data_dash - diag_data + eps))#- torch.max(data_dash)) + eps) return data_dash.type_as(data)
null
14,074
import torch from torch import nn import math from functools import partial from einops import rearrange, repeat from local_attention import LocalAttention import torch.nn.functional as F def empty(tensor): return tensor.numel() == 0
null
14,075
import torch from torch import nn import math from functools import partial from einops import rearrange, repeat from local_attention import LocalAttention import torch.nn.functional as F def exists(val): return val is not None def default(val, d): return val if exists(val) else d
null
14,076
import torch from torch import nn import math from functools import partial from einops import rearrange, repeat from local_attention import LocalAttention import torch.nn.functional as F def cast_tuple(val): return (val,) if not isinstance(val, tuple) else val
null
14,077
import torch from torch import nn import math from functools import partial from einops import rearrange, repeat from local_attention import LocalAttention import torch.nn.functional as F def calc_same_padding(kernel_size): pad = kernel_size // 2 return (pad, pad - (kernel_size + 1) % 2)
null
14,078
import torch from torch import nn import math from functools import partial from einops import rearrange, repeat from local_attention import LocalAttention import torch.nn.functional as F def linear_attention(q, k, v): if v is None: #print (k.size(), q.size()) out = torch.einsum('...ed,...nd->...ne', k, q) return out else: k_cumsum = k.sum(dim = -2) #k_cumsum = k.sum(dim = -2) D_inv = 1. / (torch.einsum('...nd,...d->...n', q, k_cumsum.type_as(q)) + 1e-8) context = torch.einsum('...nd,...ne->...de', k, v) #print ("TRUEEE: ", context.size(), q.size(), D_inv.size()) out = torch.einsum('...de,...nd,...n->...ne', context, q, D_inv) return out
null
14,079
import torch from torch import nn import math from functools import partial from einops import rearrange, repeat from local_attention import LocalAttention import torch.nn.functional as F def orthogonal_matrix_chunk(cols, qr_uniform_q = False, device = None): unstructured_block = torch.randn((cols, cols), device = device) q, r = torch.linalg.qr(unstructured_block.cpu(), mode='reduced') q, r = map(lambda t: t.to(device), (q, r)) # proposed by @Parskatt # to make sure Q is uniform https://arxiv.org/pdf/math-ph/0609050.pdf if qr_uniform_q: d = torch.diag(r, 0) q *= d.sign() return q.t() def gaussian_orthogonal_random_matrix(nb_rows, nb_columns, scaling = 0, qr_uniform_q = False, device = None): nb_full_blocks = int(nb_rows / nb_columns) #print (nb_full_blocks) block_list = [] for _ in range(nb_full_blocks): q = orthogonal_matrix_chunk(nb_columns, qr_uniform_q = qr_uniform_q, device = device) block_list.append(q) # block_list[n] is a orthogonal matrix ... (model_dim * model_dim) #print (block_list[0].size(), torch.einsum('...nd,...nd->...n', block_list[0], torch.roll(block_list[0],1,1))) #print (nb_rows, nb_full_blocks, nb_columns) remaining_rows = nb_rows - nb_full_blocks * nb_columns #print (remaining_rows) if remaining_rows > 0: q = orthogonal_matrix_chunk(nb_columns, qr_uniform_q = qr_uniform_q, device = device) #print (q[:remaining_rows].size()) block_list.append(q[:remaining_rows]) final_matrix = torch.cat(block_list) if scaling == 0: multiplier = torch.randn((nb_rows, nb_columns), device = device).norm(dim = 1) elif scaling == 1: multiplier = math.sqrt((float(nb_columns))) * torch.ones((nb_rows,), device = device) else: raise ValueError(f'Invalid scaling {scaling}') return torch.diag(multiplier) @ final_matrix
null
14,080
import os import numpy as np import yaml import torch import torch.nn.functional as F import pyworld as pw import parselmouth import torchcrepe from transformers import HubertModel, Wav2Vec2FeatureExtractor from fairseq import checkpoint_utils from torch.nn.modules.utils import consume_prefix_in_state_dict_if_present from torchaudio.transforms import Resample from .unit2control import Unit2Control from .core import frequency_filter, upsample, remove_above_fmax, MaskedAvgPool1d, MedianPool1d from ..encoder.hubert.model import HubertSoft class DotDict(dict): def __getattr__(*args): # type: ignore val = dict.get(*args) return DotDict(val) if type(val) is dict else val __setattr__ = dict.__setitem__ # type: ignore __delattr__ = dict.__delitem__ # type: ignore class Sins(torch.nn.Module): def __init__(self, sampling_rate, block_size, n_harmonics, n_mag_allpass, n_mag_noise, n_unit=256, n_spk=1): super().__init__() print(" [DDSP Model] Sinusoids Additive Synthesiser") # params self.register_buffer("sampling_rate", torch.tensor(sampling_rate)) self.register_buffer("block_size", torch.tensor(block_size)) # Unit2Control split_map = { "amplitudes": n_harmonics, "group_delay": n_mag_allpass, "noise_magnitude": n_mag_noise, } self.unit2ctrl = Unit2Control(n_unit, n_spk, split_map) def forward(self, units_frames, f0_frames, volume_frames, spk_id=None, spk_mix_dict=None, initial_phase=None, infer=True, max_upsample_dim=32): """ units_frames: B x n_frames x n_unit f0_frames: B x n_frames x 1 volume_frames: B x n_frames x 1 spk_id: B x 1 """ # exciter phase f0 = upsample(f0_frames, self.block_size) if infer: x = torch.cumsum(f0.double() / self.sampling_rate, axis=1) else: x = torch.cumsum(f0 / self.sampling_rate, axis=1) if initial_phase is not None: x += initial_phase.to(x) / 2 / np.pi x = x - torch.round(x) x = x.to(f0) phase = 2 * np.pi * x phase_frames = phase[:, :: self.block_size, :] # parameter prediction ctrls = self.unit2ctrl(units_frames, f0_frames, phase_frames, volume_frames, spk_id=spk_id, spk_mix_dict=spk_mix_dict) amplitudes_frames = torch.exp(ctrls["amplitudes"]) / 128 group_delay = np.pi * torch.tanh(ctrls["group_delay"]) noise_param = torch.exp(ctrls["noise_magnitude"]) / 128 # sinusoids exciter signal amplitudes_frames = remove_above_fmax(amplitudes_frames, f0_frames, self.sampling_rate / 2, level_start=1) n_harmonic = amplitudes_frames.shape[-1] level_harmonic = torch.arange(1, n_harmonic + 1).to(phase) sinusoids = 0.0 for n in range((n_harmonic - 1) // max_upsample_dim + 1): start = n * max_upsample_dim end = (n + 1) * max_upsample_dim phases = phase * level_harmonic[start:end] amplitudes = upsample(amplitudes_frames[:, :, start:end], self.block_size) sinusoids += (torch.sin(phases) * amplitudes).sum(-1) # harmonic part filter (apply group-delay) harmonic = frequency_filter(sinusoids, torch.exp(1.0j * torch.cumsum(group_delay, axis=-1)), hann_window=False) # noise part filter noise = torch.rand_like(harmonic) * 2 - 1 noise = frequency_filter(noise, torch.complex(noise_param, torch.zeros_like(noise_param)), hann_window=True) signal = harmonic + noise return signal, phase, (harmonic, noise) # , (noise_param, noise_param) class CombSubFast(torch.nn.Module): def __init__(self, sampling_rate, block_size, n_unit=256, n_spk=1): super().__init__() print(" [DDSP Model] Combtooth Subtractive Synthesiser") # params self.register_buffer("sampling_rate", torch.tensor(sampling_rate)) self.register_buffer("block_size", torch.tensor(block_size)) self.register_buffer("window", torch.sqrt(torch.hann_window(2 * block_size))) # Unit2Control split_map = {"harmonic_magnitude": block_size + 1, "harmonic_phase": block_size + 1, "noise_magnitude": block_size + 1} self.unit2ctrl = Unit2Control(n_unit, n_spk, split_map) def forward(self, units_frames, f0_frames, volume_frames, spk_id=None, spk_mix_dict=None, initial_phase=None, infer=True, **kwargs): """ units_frames: B x n_frames x n_unit f0_frames: B x n_frames x 1 volume_frames: B x n_frames x 1 spk_id: B x 1 """ # exciter phase f0 = upsample(f0_frames, self.block_size) if infer: x = torch.cumsum(f0.double() / self.sampling_rate, axis=1) else: x = torch.cumsum(f0 / self.sampling_rate, axis=1) if initial_phase is not None: x += initial_phase.to(x) / 2 / np.pi x = x - torch.round(x) x = x.to(f0) phase_frames = 2 * np.pi * x[:, :: self.block_size, :] # parameter prediction ctrls = self.unit2ctrl(units_frames, f0_frames, phase_frames, volume_frames, spk_id=spk_id, spk_mix_dict=spk_mix_dict) src_filter = torch.exp(ctrls["harmonic_magnitude"] + 1.0j * np.pi * ctrls["harmonic_phase"]) src_filter = torch.cat((src_filter, src_filter[:, -1:, :]), 1) noise_filter = torch.exp(ctrls["noise_magnitude"]) / 128 noise_filter = torch.cat((noise_filter, noise_filter[:, -1:, :]), 1) # combtooth exciter signal combtooth = torch.sinc(self.sampling_rate * x / (f0 + 1e-3)) combtooth = combtooth.squeeze(-1) combtooth_frames = F.pad(combtooth, (self.block_size, self.block_size)).unfold(1, 2 * self.block_size, self.block_size) combtooth_frames = combtooth_frames * self.window combtooth_fft = torch.fft.rfft(combtooth_frames, 2 * self.block_size) # noise exciter signal noise = torch.rand_like(combtooth) * 2 - 1 noise_frames = F.pad(noise, (self.block_size, self.block_size)).unfold(1, 2 * self.block_size, self.block_size) noise_frames = noise_frames * self.window noise_fft = torch.fft.rfft(noise_frames, 2 * self.block_size) # apply the filters signal_fft = combtooth_fft * src_filter + noise_fft * noise_filter # take the ifft to resynthesize audio. signal_frames_out = torch.fft.irfft(signal_fft, 2 * self.block_size) * self.window # overlap add fold = torch.nn.Fold(output_size=(1, (signal_frames_out.size(1) + 1) * self.block_size), kernel_size=(1, 2 * self.block_size), stride=(1, self.block_size)) signal = fold(signal_frames_out.transpose(1, 2))[:, 0, 0, self.block_size : -self.block_size] return signal, phase_frames, (signal, signal) class CombSub(torch.nn.Module): def __init__(self, sampling_rate, block_size, n_mag_allpass, n_mag_harmonic, n_mag_noise, n_unit=256, n_spk=1): super().__init__() print(" [DDSP Model] Combtooth Subtractive Synthesiser (Old Version)") # params self.register_buffer("sampling_rate", torch.tensor(sampling_rate)) self.register_buffer("block_size", torch.tensor(block_size)) # Unit2Control split_map = {"group_delay": n_mag_allpass, "harmonic_magnitude": n_mag_harmonic, "noise_magnitude": n_mag_noise} self.unit2ctrl = Unit2Control(n_unit, n_spk, split_map) def forward(self, units_frames, f0_frames, volume_frames, spk_id=None, spk_mix_dict=None, initial_phase=None, infer=True, **kwargs): """ units_frames: B x n_frames x n_unit f0_frames: B x n_frames x 1 volume_frames: B x n_frames x 1 spk_id: B x 1 """ # exciter phase f0 = upsample(f0_frames, self.block_size) if infer: x = torch.cumsum(f0.double() / self.sampling_rate, axis=1) else: x = torch.cumsum(f0 / self.sampling_rate, axis=1) if initial_phase is not None: x += initial_phase.to(x) / 2 / np.pi x = x - torch.round(x) x = x.to(f0) phase_frames = 2 * np.pi * x[:, :: self.block_size, :] # parameter prediction ctrls = self.unit2ctrl(units_frames, f0_frames, phase_frames, volume_frames, spk_id=spk_id, spk_mix_dict=spk_mix_dict) group_delay = np.pi * torch.tanh(ctrls["group_delay"]) src_param = torch.exp(ctrls["harmonic_magnitude"]) noise_param = torch.exp(ctrls["noise_magnitude"]) / 128 # combtooth exciter signal combtooth = torch.sinc(self.sampling_rate * x / (f0 + 1e-3)) combtooth = combtooth.squeeze(-1) # harmonic part filter (using dynamic-windowed LTV-FIR, with group-delay prediction) harmonic = frequency_filter(combtooth, torch.exp(1.0j * torch.cumsum(group_delay, axis=-1)), hann_window=False) harmonic = frequency_filter(harmonic, torch.complex(src_param, torch.zeros_like(src_param)), hann_window=True, half_width_frames=1.5 * self.sampling_rate / (f0_frames + 1e-3)) # noise part filter (using constant-windowed LTV-FIR, without group-delay) noise = torch.rand_like(harmonic) * 2 - 1 noise = frequency_filter(noise, torch.complex(noise_param, torch.zeros_like(noise_param)), hann_window=True) signal = harmonic + noise return signal, phase_frames, (harmonic, noise) def load_model(model_path, device="cpu"): config_file = os.path.join(os.path.split(model_path)[0], "config.yaml") with open(config_file, "r") as config: args = yaml.safe_load(config) args = DotDict(args) # load model model = None if args.model.type == "Sins": model = Sins(sampling_rate=args.data.sampling_rate, block_size=args.data.block_size, n_harmonics=args.model.n_harmonics, n_mag_allpass=args.model.n_mag_allpass, n_mag_noise=args.model.n_mag_noise, n_unit=args.data.encoder_out_channels, n_spk=args.model.n_spk) elif args.model.type == "CombSub": model = CombSub(sampling_rate=args.data.sampling_rate, block_size=args.data.block_size, n_mag_allpass=args.model.n_mag_allpass, n_mag_harmonic=args.model.n_mag_harmonic, n_mag_noise=args.model.n_mag_noise, n_unit=args.data.encoder_out_channels, n_spk=args.model.n_spk) elif args.model.type == "CombSubFast": model = CombSubFast(sampling_rate=args.data.sampling_rate, block_size=args.data.block_size, n_unit=args.data.encoder_out_channels, n_spk=args.model.n_spk) else: raise ValueError(f" [x] Unknown Model: {args.model.type}") print(" [Loading] " + model_path) ckpt = torch.load(model_path, map_location=torch.device(device)) model.to(device) model.load_state_dict(ckpt["model"]) model.eval() return model, args
null
14,081
import gin import numpy as np import torch import torch.nn as nn from torch.nn.utils import weight_norm from .pcmer import PCmer The provided code snippet includes necessary dependencies for implementing the `split_to_dict` function. Write a Python function `def split_to_dict(tensor, tensor_splits)` to solve the following problem: Split a tensor into a dictionary of multiple tensors. Here is the function: def split_to_dict(tensor, tensor_splits): """Split a tensor into a dictionary of multiple tensors.""" labels = [] sizes = [] for k, v in tensor_splits.items(): labels.append(k) sizes.append(v) tensors = torch.split(tensor, sizes, dim=-1) return dict(zip(labels, tensors))
Split a tensor into a dictionary of multiple tensors.
14,082
import glob import os import matplotlib import torch from torch.nn.utils import weight_norm import matplotlib.pylab as plt def plot_spectrogram(spectrogram): fig, ax = plt.subplots(figsize=(10, 2)) im = ax.imshow(spectrogram, aspect="auto", origin="lower", interpolation='none') plt.colorbar(im, ax=ax) fig.canvas.draw() plt.close() return fig
null
14,085
import glob import os import matplotlib import torch from torch.nn.utils import weight_norm import matplotlib.pylab as plt def get_padding(kernel_size, dilation=1): return int((kernel_size*dilation - dilation)/2)
null
14,088
import glob import os import matplotlib import torch from torch.nn.utils import weight_norm import matplotlib.pylab as plt def del_old_checkpoints(cp_dir, prefix, n_models=2): pattern = os.path.join(cp_dir, prefix + '????????') cp_list = glob.glob(pattern) # get checkpoint paths cp_list = sorted(cp_list)# sort by iter if len(cp_list) > n_models: # if more than n_models models are found for cp in cp_list[:-n_models]:# delete the oldest models other than lastest n_models open(cp, 'w').close()# empty file contents os.unlink(cp)# delete file (move to trash when using Colab)
null
14,089
import glob import os import matplotlib import torch from torch.nn.utils import weight_norm import matplotlib.pylab as plt def scan_checkpoint(cp_dir, prefix): pattern = os.path.join(cp_dir, prefix + '????????') cp_list = glob.glob(pattern) if len(cp_list) == 0: return None return sorted(cp_list)[-1]
null
14,090
import math import os import random import torch import torch.utils.data import numpy as np import librosa from librosa.util import normalize from librosa.filters import mel as librosa_mel_fn from scipy.io.wavfile import read import soundfile as sf import torch.nn.functional as F def load_wav_to_torch(full_path, target_sr=None, return_empty_on_exception=False): sampling_rate = None try: data, sampling_rate = sf.read(full_path, always_2d=True)# than soundfile. except Exception as ex: print(f"'{full_path}' failed to load.\nException:") print(ex) if return_empty_on_exception: return [], sampling_rate or target_sr or 48000 else: raise Exception(ex) if len(data.shape) > 1: data = data[:, 0] assert len(data) > 2# check duration of audio file is > 2 samples (because otherwise the slice operation was on the wrong dimension) if np.issubdtype(data.dtype, np.integer): # if audio data is type int max_mag = -np.iinfo(data.dtype).min # maximum magnitude = min possible value of intXX else: # if audio data is type fp32 max_mag = max(np.amax(data), -np.amin(data)) max_mag = (2**31)+1 if max_mag > (2**15) else ((2**15)+1 if max_mag > 1.01 else 1.0) # data should be either 16-bit INT, 32-bit INT or [-1 to 1] float32 data = torch.FloatTensor(data.astype(np.float32))/max_mag if (torch.isinf(data) | torch.isnan(data)).any() and return_empty_on_exception:# resample will crash with inf/NaN inputs. return_empty_on_exception will return empty arr instead of except return [], sampling_rate or target_sr or 48000 if target_sr is not None and sampling_rate != target_sr: data = torch.from_numpy(librosa.core.resample(data.numpy(), orig_sr=sampling_rate, target_sr=target_sr)) sampling_rate = target_sr return data, sampling_rate
null
14,091
import math import os import random import torch import torch.utils.data import numpy as np import librosa from librosa.util import normalize from librosa.filters import mel as librosa_mel_fn from scipy.io.wavfile import read import soundfile as sf import torch.nn.functional as F def dynamic_range_compression(x, C=1, clip_val=1e-5): return np.log(np.clip(x, a_min=clip_val, a_max=None) * C)
null
14,092
import math import os import random import torch import torch.utils.data import numpy as np import librosa from librosa.util import normalize from librosa.filters import mel as librosa_mel_fn from scipy.io.wavfile import read import soundfile as sf import torch.nn.functional as F def dynamic_range_decompression(x, C=1): return np.exp(x) / C
null
14,093
import math import os import random import torch import torch.utils.data import numpy as np import librosa from librosa.util import normalize from librosa.filters import mel as librosa_mel_fn from scipy.io.wavfile import read import soundfile as sf import torch.nn.functional as F def dynamic_range_compression_torch(x, C=1, clip_val=1e-5): return torch.log(torch.clamp(x, min=clip_val) * C)
null
14,094
import math import os import random import torch import torch.utils.data import numpy as np import librosa from librosa.util import normalize from librosa.filters import mel as librosa_mel_fn from scipy.io.wavfile import read import soundfile as sf import torch.nn.functional as F def dynamic_range_decompression_torch(x, C=1): return torch.exp(x) / C
null
14,095
import os import json from .env import AttrDict import numpy as np import torch import torch.nn.functional as F import torch.nn as nn from torch.nn import Conv1d, ConvTranspose1d, AvgPool1d, Conv2d from torch.nn.utils import weight_norm, remove_weight_norm, spectral_norm from .utils import init_weights, get_padding def load_config(model_path): config_file = os.path.join(os.path.split(model_path)[0], 'config.json') with open(config_file) as f: data = f.read() json_config = json.loads(data) h = AttrDict(json_config) return h class Generator(torch.nn.Module): def __init__(self, h): super(Generator, self).__init__() self.h = h self.num_kernels = len(h.resblock_kernel_sizes) self.num_upsamples = len(h.upsample_rates) self.m_source = SourceModuleHnNSF( sampling_rate=h.sampling_rate, harmonic_num=8 ) self.noise_convs = nn.ModuleList() self.conv_pre = weight_norm(Conv1d(h.num_mels, h.upsample_initial_channel, 7, 1, padding=3)) resblock = ResBlock1 if h.resblock == '1' else ResBlock2 self.ups = nn.ModuleList() for i, (u, k) in enumerate(zip(h.upsample_rates, h.upsample_kernel_sizes)): c_cur = h.upsample_initial_channel // (2 ** (i + 1)) self.ups.append(weight_norm( ConvTranspose1d(h.upsample_initial_channel // (2 ** i), h.upsample_initial_channel // (2 ** (i + 1)), k, u, padding=(k - u) // 2))) if i + 1 < len(h.upsample_rates): # stride_f0 = int(np.prod(h.upsample_rates[i + 1:])) self.noise_convs.append(Conv1d( 1, c_cur, kernel_size=stride_f0 * 2, stride=stride_f0, padding=stride_f0 // 2)) else: self.noise_convs.append(Conv1d(1, c_cur, kernel_size=1)) self.resblocks = nn.ModuleList() ch = h.upsample_initial_channel for i in range(len(self.ups)): ch //= 2 for j, (k, d) in enumerate(zip(h.resblock_kernel_sizes, h.resblock_dilation_sizes)): self.resblocks.append(resblock(h, ch, k, d)) self.conv_post = weight_norm(Conv1d(ch, 1, 7, 1, padding=3)) self.ups.apply(init_weights) self.conv_post.apply(init_weights) self.upp = int(np.prod(h.upsample_rates)) def forward(self, x, f0): har_source = self.m_source(f0, self.upp).transpose(1, 2) x = self.conv_pre(x) for i in range(self.num_upsamples): x = F.leaky_relu(x, LRELU_SLOPE) x = self.ups[i](x) x_source = self.noise_convs[i](har_source) x = x + x_source xs = None for j in range(self.num_kernels): if xs is None: xs = self.resblocks[i * self.num_kernels + j](x) else: xs += self.resblocks[i * self.num_kernels + j](x) x = xs / self.num_kernels x = F.leaky_relu(x) x = self.conv_post(x) x = torch.tanh(x) return x def remove_weight_norm(self): print('Removing weight norm...') for l in self.ups: remove_weight_norm(l) for l in self.resblocks: l.remove_weight_norm() remove_weight_norm(self.conv_pre) remove_weight_norm(self.conv_post) def load_model(model_path, device='cuda'): h = load_config(model_path) generator = Generator(h).to(device) cp_dict = torch.load(model_path, map_location=device) generator.load_state_dict(cp_dict['generator']) generator.eval() generator.remove_weight_norm() del cp_dict return generator, h
null
14,097
import os import json from .env import AttrDict import numpy as np import torch import torch.nn.functional as F import torch.nn as nn from torch.nn import Conv1d, ConvTranspose1d, AvgPool1d, Conv2d from torch.nn.utils import weight_norm, remove_weight_norm, spectral_norm from .utils import init_weights, get_padding def discriminator_loss(disc_real_outputs, disc_generated_outputs): loss = 0 r_losses = [] g_losses = [] for dr, dg in zip(disc_real_outputs, disc_generated_outputs): r_loss = torch.mean((1 - dr) ** 2) g_loss = torch.mean(dg ** 2) loss += (r_loss + g_loss) r_losses.append(r_loss.item()) g_losses.append(g_loss.item()) return loss, r_losses, g_losses
null