id
int64
0
190k
prompt
stringlengths
21
13.4M
docstring
stringlengths
1
12k
8,285
import numpy as np import pandas as pd import torch import gensim from gensim.models import Word2Vec from tqdm import tqdm import fire import sys import os from utils.build_vocab import Vocabulary def create_embedding(vocab_file: str, embed_size: int, output: str, caption_file: str = None, pretrained_weights_path: str = None, **word2vec_kwargs): vocabulary = torch.load(vocab_file, map_location="cpu") if pretrained_weights_path: model = gensim.models.KeyedVectors.load_word2vec_format( fname=pretrained_weights_path, binary=True, ) if model.vector_size != embed_size: assert embed_size < model.vector_size, f"only reduce dimension, cannot add dimesion {model.vector_size} to {embed_size}" from sklearn.decomposition import PCA pca = PCA(n_components=embed_size) model.vectors = pca.fit_transform(model.vectors) else: caption_df = pd.read_json(caption_file) caption_df["tokens"] = caption_df["tokens"].apply(lambda x: ["<start>"] + [token for token in x] + ["<end>"]) sentences = list(caption_df["tokens"].values) epochs = word2vec_kwargs.get("epochs", 10) if "epochs" in word2vec_kwargs: del word2vec_kwargs["epochs"] model = Word2Vec(size=embed_size, min_count=1, **word2vec_kwargs) model.build_vocab(sentences=sentences) model.train(sentences=sentences, total_examples=len(sentences), epochs=epochs) word_embeddings = np.random.randn(len(vocabulary), embed_size) if isinstance(model, gensim.models.word2vec.Word2Vec): model = model.wv with tqdm(total=len(vocabulary), ascii=True) as pbar: for word, idx in vocabulary.word2idx.items(): try: word_embeddings[idx] = model.get_vector(word) except KeyError: print(f"word {word} not found in word2vec model, it is random initialized!") pbar.update() np.save(output, word_embeddings) print("Finish writing word2vec embeddings to " + output)
null
8,286
import sys import os import librosa import numpy as np import torch import audio_to_text.captioning.models import audio_to_text.captioning.models.encoder import audio_to_text.captioning.models.decoder import audio_to_text.captioning.utils.train_util as train_util def load_model(config, checkpoint): ckpt = torch.load(checkpoint, "cpu") encoder_cfg = config["model"]["encoder"] encoder = train_util.init_obj( audio_to_text.captioning.models.encoder, encoder_cfg ) if "pretrained" in encoder_cfg: pretrained = encoder_cfg["pretrained"] train_util.load_pretrained_model(encoder, pretrained, sys.stdout.write) decoder_cfg = config["model"]["decoder"] if "vocab_size" not in decoder_cfg["args"]: decoder_cfg["args"]["vocab_size"] = len(ckpt["vocabulary"]) decoder = train_util.init_obj( audio_to_text.captioning.models.decoder, decoder_cfg ) if "word_embedding" in decoder_cfg: decoder.load_word_embedding(**decoder_cfg["word_embedding"]) if "pretrained" in decoder_cfg: pretrained = decoder_cfg["pretrained"] train_util.load_pretrained_model(decoder, pretrained, sys.stdout.write) model = train_util.init_obj(audio_to_text.captioning.models, config["model"], encoder=encoder, decoder=decoder) train_util.load_pretrained_model(model, ckpt) model.eval() return { "model": model, "vocabulary": ckpt["vocabulary"] }
null
8,287
import sys import os import librosa import numpy as np import torch import audio_to_text.captioning.models import audio_to_text.captioning.models.encoder import audio_to_text.captioning.models.decoder import audio_to_text.captioning.utils.train_util as train_util def decode_caption(word_ids, vocabulary): candidate = [] for word_id in word_ids: word = vocabulary[word_id] if word == "<end>": break elif word == "<start>": continue candidate.append(word) candidate = " ".join(candidate) return candidate
null
8,288
import math import random from collections import deque from functools import partial from inspect import isfunction from pathlib import Path import numpy as np import torch import torch.nn.functional as F from torch import nn from tqdm import tqdm from einops import rearrange from modules.fastspeech.fs2 import FastSpeech2 from modules.diffsinger_midi.fs2 import FastSpeech2MIDI from utils.hparams import hparams def exists(x): return x is not None def default(val, d): if exists(val): return val return d() if isfunction(d) else d
null
8,289
import math import random from collections import deque from functools import partial from inspect import isfunction from pathlib import Path import numpy as np import torch import torch.nn.functional as F from torch import nn from tqdm import tqdm from einops import rearrange from modules.fastspeech.fs2 import FastSpeech2 from modules.diffsinger_midi.fs2 import FastSpeech2MIDI from utils.hparams import hparams def extract(a, t, x_shape): b, *_ = t.shape out = a.gather(-1, t) return out.reshape(b, *((1,) * (len(x_shape) - 1)))
null
8,290
import math import random from collections import deque from functools import partial from inspect import isfunction from pathlib import Path import numpy as np import torch import torch.nn.functional as F from torch import nn from tqdm import tqdm from einops import rearrange from modules.fastspeech.fs2 import FastSpeech2 from modules.diffsinger_midi.fs2 import FastSpeech2MIDI from utils.hparams import hparams def noise_like(shape, device, repeat=False): repeat_noise = lambda: torch.randn((1, *shape[1:]), device=device).repeat(shape[0], *((1,) * (len(shape) - 1))) noise = lambda: torch.randn(shape, device=device) return repeat_noise() if repeat else noise()
null
8,291
import math import random from collections import deque from functools import partial from inspect import isfunction from pathlib import Path import numpy as np import torch import torch.nn.functional as F from torch import nn from tqdm import tqdm from einops import rearrange from modules.fastspeech.fs2 import FastSpeech2 from modules.diffsinger_midi.fs2 import FastSpeech2MIDI from utils.hparams import hparams hparams = {} The provided code snippet includes necessary dependencies for implementing the `linear_beta_schedule` function. Write a Python function `def linear_beta_schedule(timesteps, max_beta=hparams.get('max_beta', 0.01))` to solve the following problem: linear schedule Here is the function: def linear_beta_schedule(timesteps, max_beta=hparams.get('max_beta', 0.01)): """ linear schedule """ betas = np.linspace(1e-4, max_beta, timesteps) return betas
linear schedule
8,292
import math import random from collections import deque from functools import partial from inspect import isfunction from pathlib import Path import numpy as np import torch import torch.nn.functional as F from torch import nn from tqdm import tqdm from einops import rearrange from modules.fastspeech.fs2 import FastSpeech2 from modules.diffsinger_midi.fs2 import FastSpeech2MIDI from utils.hparams import hparams The provided code snippet includes necessary dependencies for implementing the `cosine_beta_schedule` function. Write a Python function `def cosine_beta_schedule(timesteps, s=0.008)` to solve the following problem: cosine schedule as proposed in https://openreview.net/forum?id=-NEXDKk8gZ Here is the function: def cosine_beta_schedule(timesteps, s=0.008): """ cosine schedule as proposed in https://openreview.net/forum?id=-NEXDKk8gZ """ steps = timesteps + 1 x = np.linspace(0, steps, steps) alphas_cumprod = np.cos(((x / steps) + s) / (1 + s) * np.pi * 0.5) ** 2 alphas_cumprod = alphas_cumprod / alphas_cumprod[0] betas = 1 - (alphas_cumprod[1:] / alphas_cumprod[:-1]) return np.clip(betas, a_min=0, a_max=0.999)
cosine schedule as proposed in https://openreview.net/forum?id=-NEXDKk8gZ
8,293
import math import random from functools import partial from inspect import isfunction from pathlib import Path import numpy as np import torch import torch.nn.functional as F from torch import nn from tqdm import tqdm from einops import rearrange from modules.fastspeech.fs2 import FastSpeech2 from modules.diffsinger_midi.fs2 import FastSpeech2MIDI from utils.hparams import hparams def exists(x): return x is not None def default(val, d): if exists(val): return val return d() if isfunction(d) else d
null
8,294
import math import random from functools import partial from inspect import isfunction from pathlib import Path import numpy as np import torch import torch.nn.functional as F from torch import nn from tqdm import tqdm from einops import rearrange from modules.fastspeech.fs2 import FastSpeech2 from modules.diffsinger_midi.fs2 import FastSpeech2MIDI from utils.hparams import hparams def cycle(dl): while True: for data in dl: yield data
null
8,295
import math import random from functools import partial from inspect import isfunction from pathlib import Path import numpy as np import torch import torch.nn.functional as F from torch import nn from tqdm import tqdm from einops import rearrange from modules.fastspeech.fs2 import FastSpeech2 from modules.diffsinger_midi.fs2 import FastSpeech2MIDI from utils.hparams import hparams def num_to_groups(num, divisor): groups = num // divisor remainder = num % divisor arr = [divisor] * groups if remainder > 0: arr.append(remainder) return arr
null
8,296
import math import random from functools import partial from inspect import isfunction from pathlib import Path import numpy as np import torch import torch.nn.functional as F from torch import nn from tqdm import tqdm from einops import rearrange from modules.fastspeech.fs2 import FastSpeech2 from modules.diffsinger_midi.fs2 import FastSpeech2MIDI from utils.hparams import hparams def extract(a, t, x_shape): b, *_ = t.shape out = a.gather(-1, t) return out.reshape(b, *((1,) * (len(x_shape) - 1)))
null
8,297
import math import random from functools import partial from inspect import isfunction from pathlib import Path import numpy as np import torch import torch.nn.functional as F from torch import nn from tqdm import tqdm from einops import rearrange from modules.fastspeech.fs2 import FastSpeech2 from modules.diffsinger_midi.fs2 import FastSpeech2MIDI from utils.hparams import hparams def noise_like(shape, device, repeat=False): repeat_noise = lambda: torch.randn((1, *shape[1:]), device=device).repeat(shape[0], *((1,) * (len(shape) - 1))) noise = lambda: torch.randn(shape, device=device) return repeat_noise() if repeat else noise()
null
8,298
import math import random from functools import partial from inspect import isfunction from pathlib import Path import numpy as np import torch import torch.nn.functional as F from torch import nn from tqdm import tqdm from einops import rearrange from modules.fastspeech.fs2 import FastSpeech2 from modules.diffsinger_midi.fs2 import FastSpeech2MIDI from utils.hparams import hparams The provided code snippet includes necessary dependencies for implementing the `cosine_beta_schedule` function. Write a Python function `def cosine_beta_schedule(timesteps, s=0.008)` to solve the following problem: cosine schedule as proposed in https://openreview.net/forum?id=-NEXDKk8gZ Here is the function: def cosine_beta_schedule(timesteps, s=0.008): """ cosine schedule as proposed in https://openreview.net/forum?id=-NEXDKk8gZ """ steps = timesteps + 1 x = np.linspace(0, steps, steps) alphas_cumprod = np.cos(((x / steps) + s) / (1 + s) * np.pi * 0.5) ** 2 alphas_cumprod = alphas_cumprod / alphas_cumprod[0] betas = 1 - (alphas_cumprod[1:] / alphas_cumprod[:-1]) return np.clip(betas, a_min=0, a_max=0.999)
cosine schedule as proposed in https://openreview.net/forum?id=-NEXDKk8gZ
8,299
import math import torch import torch.nn as nn import torch.nn.functional as F from math import sqrt from .diffusion import Mish from utils.hparams import hparams def Conv1d(*args, **kwargs): layer = nn.Conv1d(*args, **kwargs) nn.init.kaiming_normal_(layer.weight) return layer
null
8,300
import math import torch import torch.nn as nn import torch.nn.functional as F from math import sqrt from .diffusion import Mish from utils.hparams import hparams def silu(x): return x * torch.sigmoid(x)
null
8,301
from modules.fastspeech.tts_modules import FastspeechDecoder import torch from torch.nn import functional as F import torch.nn as nn import math from utils.hparams import hparams from .diffusion import Mish def Conv1d(*args, **kwargs): layer = nn.Conv1d(*args, **kwargs) nn.init.kaiming_normal_(layer.weight) return layer
null
8,302
from copy import deepcopy import torch import dgl import stanza import networkx as nx The provided code snippet includes necessary dependencies for implementing the `plot_dgl_sentence_graph` function. Write a Python function `def plot_dgl_sentence_graph(dgl_graph, labels)` to solve the following problem: labels = {idx: word for idx,word in enumerate(sentence.split(" ")) } Here is the function: def plot_dgl_sentence_graph(dgl_graph, labels): """ labels = {idx: word for idx,word in enumerate(sentence.split(" ")) } """ import matplotlib.pyplot as plt nx_graph = dgl_graph.to_networkx() pos = nx.random_layout(nx_graph) nx.draw(nx_graph, pos, with_labels=False) nx.draw_networkx_labels(nx_graph, pos, labels) plt.show()
labels = {idx: word for idx,word in enumerate(sentence.split(" ")) }
8,303
import torch import torch.nn as nn import torch.nn.functional as F import dgl from dgl.nn.pytorch import GatedGraphConv def sequence_mask(lengths, maxlen, dtype=torch.bool): if maxlen is None: maxlen = lengths.max() mask = ~(torch.ones((len(lengths), maxlen)).to(lengths.device).cumsum(dim=1).t() > lengths).t() mask.type(dtype) return mask
null
8,304
import torch import torch.nn as nn import torch.nn.functional as F import dgl from dgl.nn.pytorch import GatedGraphConv The provided code snippet includes necessary dependencies for implementing the `group_hidden_by_segs` function. Write a Python function `def group_hidden_by_segs(h, seg_ids, max_len)` to solve the following problem: :param h: [B, T, H] :param seg_ids: [B, T] :return: h_ph: [B, T_ph, H] Here is the function: def group_hidden_by_segs(h, seg_ids, max_len): """ :param h: [B, T, H] :param seg_ids: [B, T] :return: h_ph: [B, T_ph, H] """ B, T, H = h.shape h_gby_segs = h.new_zeros([B, max_len + 1, H]).scatter_add_(1, seg_ids[:, :, None].repeat([1, 1, H]), h) all_ones = h.new_ones(h.shape[:2]) cnt_gby_segs = h.new_zeros([B, max_len + 1]).scatter_add_(1, seg_ids, all_ones).contiguous() h_gby_segs = h_gby_segs[:, 1:] cnt_gby_segs = cnt_gby_segs[:, 1:] h_gby_segs = h_gby_segs / torch.clamp(cnt_gby_segs[:, :, None], min=1) # assert h_gby_segs.shape[-1] == 192 return h_gby_segs
:param h: [B, T, H] :param seg_ids: [B, T] :return: h_ph: [B, T_ph, H]
8,305
import torch import torch.nn.functional as F def build_word_mask(x2word, y2word): return (x2word[:, :, None] == y2word[:, None, :]).long()
null
8,306
import torch import torch.nn.functional as F def mel2ph_to_mel2word(mel2ph, ph2word): mel2word = (ph2word - 1).gather(1, (mel2ph - 1).clamp(min=0)) + 1 mel2word = mel2word * (mel2ph > 0).long() return mel2word
null
8,307
import torch import torch.nn.functional as F def clip_mel2token_to_multiple(mel2token, frames_multiple): max_frames = mel2token.shape[1] // frames_multiple * frames_multiple mel2token = mel2token[:, :max_frames] return mel2token
null
8,308
import torch import torch.nn.functional as F def expand_states(h, mel2token): h = F.pad(h, [0, 0, 1, 0]) mel2token_ = mel2token[..., None].repeat([1, 1, h.shape[-1]]) h = torch.gather(h, 1, mel2token_) # [B, T, H] return h
null
8,309
import torch from torch import nn def fused_add_tanh_sigmoid_multiply(input_a, input_b, n_channels): n_channels_int = n_channels[0] in_act = input_a + input_b t_act = torch.tanh(in_act[:, :n_channels_int, :]) s_act = torch.sigmoid(in_act[:, n_channels_int:, :]) acts = t_act * s_act return acts
null
8,310
import torch def squeeze(x, x_mask=None, n_sqz=2): b, c, t = x.size() t = (t // n_sqz) * n_sqz x = x[:, :, :t] x_sqz = x.view(b, c, t // n_sqz, n_sqz) x_sqz = x_sqz.permute(0, 3, 1, 2).contiguous().view(b, c * n_sqz, t // n_sqz) if x_mask is not None: x_mask = x_mask[:, :, n_sqz - 1::n_sqz] else: x_mask = torch.ones(b, 1, t // n_sqz).to(device=x.device, dtype=x.dtype) return x_sqz * x_mask, x_mask
null
8,311
import torch def unsqueeze(x, x_mask=None, n_sqz=2): b, c, t = x.size() x_unsqz = x.view(b, n_sqz, c // n_sqz, t) x_unsqz = x_unsqz.permute(0, 2, 3, 1).contiguous().view(b, c // n_sqz, t * n_sqz) if x_mask is not None: x_mask = x_mask.unsqueeze(-1).repeat(1, 1, 1, n_sqz).view(b, 1, t * n_sqz) else: x_mask = torch.ones(b, 1, t * n_sqz).to(device=x.device, dtype=x.dtype) return x_unsqz * x_mask, x_mask
null
8,312
import torch import torch.nn.functional as F from torch.autograd import Variable import numpy as np from math import exp def create_window(window_size, channel): _1D_window = gaussian(window_size, 1.5).unsqueeze(1) _2D_window = _1D_window.mm(_1D_window.t()).float().unsqueeze(0).unsqueeze(0) window = Variable(_2D_window.expand(channel, 1, window_size, window_size).contiguous()) return window def _ssim(img1, img2, window, window_size, channel, size_average=True): mu1 = F.conv2d(img1, window, padding=window_size // 2, groups=channel) mu2 = F.conv2d(img2, window, padding=window_size // 2, groups=channel) mu1_sq = mu1.pow(2) mu2_sq = mu2.pow(2) mu1_mu2 = mu1 * mu2 sigma1_sq = F.conv2d(img1 * img1, window, padding=window_size // 2, groups=channel) - mu1_sq sigma2_sq = F.conv2d(img2 * img2, window, padding=window_size // 2, groups=channel) - mu2_sq sigma12 = F.conv2d(img1 * img2, window, padding=window_size // 2, groups=channel) - mu1_mu2 C1 = 0.01 ** 2 C2 = 0.03 ** 2 ssim_map = ((2 * mu1_mu2 + C1) * (2 * sigma12 + C2)) / ((mu1_sq + mu2_sq + C1) * (sigma1_sq + sigma2_sq + C2)) if size_average: return ssim_map.mean() else: return ssim_map.mean(1) window = None def ssim(img1, img2, window_size=11, size_average=True): (_, channel, _, _) = img1.size() global window if window is None: window = create_window(window_size, channel) if img1.is_cuda: window = window.cuda(img1.get_device()) window = window.type_as(img1) return _ssim(img1, img2, window, window_size, channel, size_average)
null
8,313
import math import torch import torch.nn as nn import torch.nn.functional as F from modules.commons.common_layers import Embedding from modules.fastspeech.tts_modules import LayerNorm def init_weights_func(m): classname = m.__class__.__name__ if classname.find("Conv1d") != -1: torch.nn.init.xavier_uniform_(m.weight)
null
8,314
import math import torch from torch import nn from torch.nn import Parameter import torch.onnx.operators import torch.nn.functional as F import utils def Embedding(num_embeddings, embedding_dim, padding_idx=None): m = nn.Embedding(num_embeddings, embedding_dim, padding_idx=padding_idx) nn.init.normal_(m.weight, mean=0, std=embedding_dim ** -0.5) if padding_idx is not None: nn.init.constant_(m.weight[padding_idx], 0) return m
null
8,315
import math import torch from torch import nn from torch.nn import Parameter import torch.onnx.operators import torch.nn.functional as F import utils def LayerNorm(normalized_shape, eps=1e-5, elementwise_affine=True, export=False): if not export and torch.cuda.is_available(): try: from apex.normalization import FusedLayerNorm return FusedLayerNorm(normalized_shape, eps, elementwise_affine) except ImportError: pass return torch.nn.LayerNorm(normalized_shape, eps, elementwise_affine)
null
8,316
import math import torch from torch import nn from torch.nn import Parameter import torch.onnx.operators import torch.nn.functional as F import utils def Linear(in_features, out_features, bias=True): m = nn.Linear(in_features, out_features, bias) nn.init.xavier_uniform_(m.weight) if bias: nn.init.constant_(m.bias, 0.) return m
null
8,317
import math import torch from torch import nn from torch.nn import functional as F from utils.hparams import hparams from modules.commons.common_layers import Embedding from utils.tts_utils import group_hidden_by_segs, expand_word2ph import transformers def convert_pad_shape(pad_shape): l = pad_shape[::-1] pad_shape = [item for sublist in l for item in sublist] return pad_shape def shift_1d(x): x = F.pad(x, convert_pad_shape([[0, 0], [0, 0], [1, 0]]))[:, :, :-1] return x
null
8,318
import math import torch from torch import nn from torch.nn import functional as F from utils.hparams import hparams from modules.commons.common_layers import Embedding from utils.tts_utils import group_hidden_by_segs, expand_word2ph import transformers def sequence_mask(length, max_length=None): if max_length is None: max_length = length.max() x = torch.arange(max_length, dtype=length.dtype, device=length.device) return x.unsqueeze(0) < length.unsqueeze(1)
null
8,319
import logging import math import torch import torch.nn as nn from torch.nn import functional as F from modules.commons.espnet_positional_embedding import RelPositionalEncoding from modules.commons.common_layers import SinusoidalPositionalEmbedding, Linear, EncSALayer, DecSALayer, BatchNorm1dTBC from utils.hparams import hparams def mel2ph_to_dur(mel2ph, T_txt, max_dur=None): B, _ = mel2ph.shape dur = mel2ph.new_zeros(B, T_txt + 1).scatter_add(1, mel2ph, torch.ones_like(mel2ph)) dur = dur[:, 1:] if max_dur is not None: dur = dur.clamp(max=max_dur) return dur
null
8,320
from modules.commons.common_layers import * def fused_add_tanh_sigmoid_multiply(input_a, input_b, n_channels): n_channels_int = n_channels[0] in_act = input_a + input_b t_act = torch.tanh(in_act[:, :n_channels_int, :]) s_act = torch.sigmoid(in_act[:, n_channels_int:, :]) acts = t_act * s_act return acts
null
8,321
import scipy from torch.nn import functional as F import torch from torch import nn import numpy as np from modules.commons.common_layers import Permute from modules.fastspeech.tts_modules import FFTBlocks from modules.GenerSpeech.model.wavenet import fused_add_tanh_sigmoid_multiply, WN def squeeze(x, x_mask=None, n_sqz=2): b, c, t = x.size() t = (t // n_sqz) * n_sqz x = x[:, :, :t] x_sqz = x.view(b, c, t // n_sqz, n_sqz) x_sqz = x_sqz.permute(0, 3, 1, 2).contiguous().view(b, c * n_sqz, t // n_sqz) if x_mask is not None: x_mask = x_mask[:, :, n_sqz - 1::n_sqz] else: x_mask = torch.ones(b, 1, t // n_sqz).to(device=x.device, dtype=x.dtype) return x_sqz * x_mask, x_mask
null
8,322
import scipy from torch.nn import functional as F import torch from torch import nn import numpy as np from modules.commons.common_layers import Permute from modules.fastspeech.tts_modules import FFTBlocks from modules.GenerSpeech.model.wavenet import fused_add_tanh_sigmoid_multiply, WN def unsqueeze(x, x_mask=None, n_sqz=2): b, c, t = x.size() x_unsqz = x.view(b, n_sqz, c // n_sqz, t) x_unsqz = x_unsqz.permute(0, 2, 3, 1).contiguous().view(b, c // n_sqz, t * n_sqz) if x_mask is not None: x_mask = x_mask.unsqueeze(-1).repeat(1, 1, 1, n_sqz).view(b, 1, t * n_sqz) else: x_mask = torch.ones(b, 1, t * n_sqz).to(device=x.device, dtype=x.dtype) return x_unsqz * x_mask, x_mask
null
8,323
from torch import nn import copy import torch from utils.hparams import hparams from modules.GenerSpeech.model.wavenet import WN import math from modules.fastspeech.tts_modules import LayerNorm import torch.nn.functional as F from utils.tts_utils import group_hidden_by_segs, sequence_mask from scipy.cluster.vq import kmeans2 from torch.nn import functional as F def _make_guided_attention_mask(ilen, rilen, olen, rolen, sigma): grid_x, grid_y = torch.meshgrid(torch.arange(ilen, device=rilen.device), torch.arange(olen, device=rolen.device)) grid_x = grid_x.unsqueeze(0).expand(rilen.size(0), -1, -1) grid_y = grid_y.unsqueeze(0).expand(rolen.size(0), -1, -1) rilen = rilen.unsqueeze(1).unsqueeze(1) rolen = rolen.unsqueeze(1).unsqueeze(1) return 1.0 - torch.exp( -((grid_y.float() / rolen - grid_x.float() / rilen) ** 2) / (2 * (sigma ** 2)) )
null
8,324
from torch import nn import copy import torch from utils.hparams import hparams from modules.GenerSpeech.model.wavenet import WN import math from modules.fastspeech.tts_modules import LayerNorm import torch.nn.functional as F from utils.tts_utils import group_hidden_by_segs, sequence_mask from scipy.cluster.vq import kmeans2 from torch.nn import functional as F def init_weights_func(m): classname = m.__class__.__name__ if classname.find("Conv1d") != -1: torch.nn.init.xavier_uniform_(m.weight)
null
8,325
import torch import torch.nn.functional as F import torch.nn as nn from torch.nn import Conv1d, ConvTranspose1d, AvgPool1d, Conv2d from torch.nn.utils import weight_norm, remove_weight_norm, spectral_norm from modules.parallel_wavegan.layers import UpsampleNetwork, ConvInUpsampleNetwork from modules.parallel_wavegan.models.source import SourceModuleHnNSF import numpy as np def init_weights(m, mean=0.0, std=0.01): classname = m.__class__.__name__ if classname.find("Conv") != -1: m.weight.data.normal_(mean, std)
null
8,326
import torch import torch.nn.functional as F import torch.nn as nn from torch.nn import Conv1d, ConvTranspose1d, AvgPool1d, Conv2d from torch.nn.utils import weight_norm, remove_weight_norm, spectral_norm from modules.parallel_wavegan.layers import UpsampleNetwork, ConvInUpsampleNetwork from modules.parallel_wavegan.models.source import SourceModuleHnNSF import numpy as np def apply_weight_norm(m): classname = m.__class__.__name__ if classname.find("Conv") != -1: weight_norm(m)
null
8,327
import torch import torch.nn.functional as F import torch.nn as nn from torch.nn import Conv1d, ConvTranspose1d, AvgPool1d, Conv2d from torch.nn.utils import weight_norm, remove_weight_norm, spectral_norm from modules.parallel_wavegan.layers import UpsampleNetwork, ConvInUpsampleNetwork from modules.parallel_wavegan.models.source import SourceModuleHnNSF import numpy as np def get_padding(kernel_size, dilation=1): return int((kernel_size * dilation - dilation) / 2)
null
8,328
import torch import torch.nn.functional as F import torch.nn as nn from torch.nn import Conv1d, ConvTranspose1d, AvgPool1d, Conv2d from torch.nn.utils import weight_norm, remove_weight_norm, spectral_norm from modules.parallel_wavegan.layers import UpsampleNetwork, ConvInUpsampleNetwork from modules.parallel_wavegan.models.source import SourceModuleHnNSF import numpy as np def feature_loss(fmap_r, fmap_g): loss = 0 for dr, dg in zip(fmap_r, fmap_g): for rl, gl in zip(dr, dg): loss += torch.mean(torch.abs(rl - gl)) return loss * 2
null
8,329
import torch import torch.nn.functional as F import torch.nn as nn from torch.nn import Conv1d, ConvTranspose1d, AvgPool1d, Conv2d from torch.nn.utils import weight_norm, remove_weight_norm, spectral_norm from modules.parallel_wavegan.layers import UpsampleNetwork, ConvInUpsampleNetwork from modules.parallel_wavegan.models.source import SourceModuleHnNSF import numpy as np def discriminator_loss(disc_real_outputs, disc_generated_outputs): r_losses = 0 g_losses = 0 for dr, dg in zip(disc_real_outputs, disc_generated_outputs): r_loss = torch.mean((1 - dr) ** 2) g_loss = torch.mean(dg ** 2) r_losses += r_loss g_losses += g_loss r_losses = r_losses / len(disc_real_outputs) g_losses = g_losses / len(disc_real_outputs) return r_losses, g_losses
null
8,330
import torch import torch.nn.functional as F import torch.nn as nn from torch.nn import Conv1d, ConvTranspose1d, AvgPool1d, Conv2d from torch.nn.utils import weight_norm, remove_weight_norm, spectral_norm from modules.parallel_wavegan.layers import UpsampleNetwork, ConvInUpsampleNetwork from modules.parallel_wavegan.models.source import SourceModuleHnNSF import numpy as np def cond_discriminator_loss(outputs): loss = 0 for dg in outputs: g_loss = torch.mean(dg ** 2) loss += g_loss loss = loss / len(outputs) return loss
null
8,331
import torch import torch.nn.functional as F import torch.nn as nn from torch.nn import Conv1d, ConvTranspose1d, AvgPool1d, Conv2d from torch.nn.utils import weight_norm, remove_weight_norm, spectral_norm from modules.parallel_wavegan.layers import UpsampleNetwork, ConvInUpsampleNetwork from modules.parallel_wavegan.models.source import SourceModuleHnNSF import numpy as np def generator_loss(disc_outputs): loss = 0 for dg in disc_outputs: l = torch.mean((1 - dg) ** 2) loss += l loss = loss / len(disc_outputs) return loss
null
8,332
import numpy as np import torch import torch.utils.data from librosa.filters import mel as librosa_mel_fn from scipy.io.wavfile import read def load_wav(full_path): sampling_rate, data = read(full_path) return data, sampling_rate
null
8,333
import numpy as np import torch import torch.utils.data from librosa.filters import mel as librosa_mel_fn from scipy.io.wavfile import read def dynamic_range_compression(x, C=1, clip_val=1e-5): return np.log(np.clip(x, a_min=clip_val, a_max=None) * C)
null
8,334
import numpy as np import torch import torch.utils.data from librosa.filters import mel as librosa_mel_fn from scipy.io.wavfile import read def dynamic_range_decompression(x, C=1): return np.exp(x) / C
null
8,335
import numpy as np import torch import torch.utils.data from librosa.filters import mel as librosa_mel_fn from scipy.io.wavfile import read def dynamic_range_decompression_torch(x, C=1): return torch.exp(x) / C def spectral_de_normalize_torch(magnitudes): output = dynamic_range_decompression_torch(magnitudes) return output
null
8,336
import numpy as np import torch import torch.utils.data from librosa.filters import mel as librosa_mel_fn from scipy.io.wavfile import read def spectral_normalize_torch(magnitudes): output = dynamic_range_compression_torch(magnitudes) return output mel_basis = {} hann_window = {} def mel_spectrogram(y, hparams, center=False, complex=False): # hop_size: 512 # For 22050Hz, 275 ~= 12.5 ms (0.0125 * sample_rate) # win_size: 2048 # For 22050Hz, 1100 ~= 50 ms (If None, win_size: fft_size) (0.05 * sample_rate) # fmin: 55 # Set this to 55 if your speaker is male! if female, 95 should help taking off noise. (To test depending on dataset. Pitch info: male~[65, 260], female~[100, 525]) # fmax: 10000 # To be increased/reduced depending on data. # fft_size: 2048 # Extra window size is filled with 0 paddings to match this parameter # n_fft, num_mels, sampling_rate, hop_size, win_size, fmin, fmax, n_fft = hparams['fft_size'] num_mels = hparams['audio_num_mel_bins'] sampling_rate = hparams['audio_sample_rate'] hop_size = hparams['hop_size'] win_size = hparams['win_size'] fmin = hparams['fmin'] fmax = hparams['fmax'] y = y.clamp(min=-1., max=1.) global mel_basis, hann_window if fmax not in mel_basis: mel = librosa_mel_fn(sampling_rate, n_fft, num_mels, fmin, fmax) mel_basis[str(fmax) + '_' + str(y.device)] = torch.from_numpy(mel).float().to(y.device) hann_window[str(y.device)] = torch.hann_window(win_size).to(y.device) y = torch.nn.functional.pad(y.unsqueeze(1), (int((n_fft - hop_size) / 2), int((n_fft - hop_size) / 2)), mode='reflect') y = y.squeeze(1) spec = torch.stft(y, n_fft, hop_length=hop_size, win_length=win_size, window=hann_window[str(y.device)], center=center, pad_mode='reflect', normalized=False, onesided=True) if not complex: spec = torch.sqrt(spec.pow(2).sum(-1) + (1e-9)) spec = torch.matmul(mel_basis[str(fmax) + '_' + str(y.device)], spec) spec = spectral_normalize_torch(spec) else: B, C, T, _ = spec.shape spec = spec.transpose(1, 2) # [B, T, n_fft, 2] return spec
null
8,337
import torch import torch.nn.functional as F The provided code snippet includes necessary dependencies for implementing the `stft` function. Write a Python function `def stft(x, fft_size, hop_size, win_length, window)` to solve the following problem: Perform STFT and convert to magnitude spectrogram. Args: x (Tensor): Input signal tensor (B, T). fft_size (int): FFT size. hop_size (int): Hop size. win_length (int): Window length. window (str): Window function type. Returns: Tensor: Magnitude spectrogram (B, #frames, fft_size // 2 + 1). Here is the function: def stft(x, fft_size, hop_size, win_length, window): """Perform STFT and convert to magnitude spectrogram. Args: x (Tensor): Input signal tensor (B, T). fft_size (int): FFT size. hop_size (int): Hop size. win_length (int): Window length. window (str): Window function type. Returns: Tensor: Magnitude spectrogram (B, #frames, fft_size // 2 + 1). """ x_stft = torch.stft(x, fft_size, hop_size, win_length, window) real = x_stft[..., 0] imag = x_stft[..., 1] # NOTE(kan-bayashi): clamp is needed to avoid nan or inf return torch.sqrt(torch.clamp(real ** 2 + imag ** 2, min=1e-7)).transpose(2, 1)
Perform STFT and convert to magnitude spectrogram. Args: x (Tensor): Input signal tensor (B, T). fft_size (int): FFT size. hop_size (int): Hop size. win_length (int): Window length. window (str): Window function type. Returns: Tensor: Magnitude spectrogram (B, #frames, fft_size // 2 + 1).
8,338
import numpy as np import torch import torch.nn.functional as F from scipy.signal import kaiser The provided code snippet includes necessary dependencies for implementing the `design_prototype_filter` function. Write a Python function `def design_prototype_filter(taps=62, cutoff_ratio=0.15, beta=9.0)` to solve the following problem: Design prototype filter for PQMF. This method is based on `A Kaiser window approach for the design of prototype filters of cosine modulated filterbanks`_. Args: taps (int): The number of filter taps. cutoff_ratio (float): Cut-off frequency ratio. beta (float): Beta coefficient for kaiser window. Returns: ndarray: Impluse response of prototype filter (taps + 1,). .. _`A Kaiser window approach for the design of prototype filters of cosine modulated filterbanks`: https://ieeexplore.ieee.org/abstract/document/681427 Here is the function: def design_prototype_filter(taps=62, cutoff_ratio=0.15, beta=9.0): """Design prototype filter for PQMF. This method is based on `A Kaiser window approach for the design of prototype filters of cosine modulated filterbanks`_. Args: taps (int): The number of filter taps. cutoff_ratio (float): Cut-off frequency ratio. beta (float): Beta coefficient for kaiser window. Returns: ndarray: Impluse response of prototype filter (taps + 1,). .. _`A Kaiser window approach for the design of prototype filters of cosine modulated filterbanks`: https://ieeexplore.ieee.org/abstract/document/681427 """ # check the arguments are valid assert taps % 2 == 0, "The number of taps mush be even number." assert 0.0 < cutoff_ratio < 1.0, "Cutoff ratio must be > 0.0 and < 1.0." # make initial filter omega_c = np.pi * cutoff_ratio with np.errstate(invalid='ignore'): h_i = np.sin(omega_c * (np.arange(taps + 1) - 0.5 * taps)) \ / (np.pi * (np.arange(taps + 1) - 0.5 * taps)) h_i[taps // 2] = np.cos(0) * cutoff_ratio # fix nan due to indeterminate form # apply kaiser window w = kaiser(taps + 1, beta) h = h_i * w return h
Design prototype filter for PQMF. This method is based on `A Kaiser window approach for the design of prototype filters of cosine modulated filterbanks`_. Args: taps (int): The number of filter taps. cutoff_ratio (float): Cut-off frequency ratio. beta (float): Beta coefficient for kaiser window. Returns: ndarray: Impluse response of prototype filter (taps + 1,). .. _`A Kaiser window approach for the design of prototype filters of cosine modulated filterbanks`: https://ieeexplore.ieee.org/abstract/document/681427
8,339
import fnmatch import logging import os import sys import h5py import numpy as np The provided code snippet includes necessary dependencies for implementing the `find_files` function. Write a Python function `def find_files(root_dir, query="*.wav", include_root_dir=True)` to solve the following problem: Find files recursively. Args: root_dir (str): Root root_dir to find. query (str): Query to find. include_root_dir (bool): If False, root_dir name is not included. Returns: list: List of found filenames. Here is the function: def find_files(root_dir, query="*.wav", include_root_dir=True): """Find files recursively. Args: root_dir (str): Root root_dir to find. query (str): Query to find. include_root_dir (bool): If False, root_dir name is not included. Returns: list: List of found filenames. """ files = [] for root, dirnames, filenames in os.walk(root_dir, followlinks=True): for filename in fnmatch.filter(filenames, query): files.append(os.path.join(root, filename)) if not include_root_dir: files = [file_.replace(root_dir + "/", "") for file_ in files] return files
Find files recursively. Args: root_dir (str): Root root_dir to find. query (str): Query to find. include_root_dir (bool): If False, root_dir name is not included. Returns: list: List of found filenames.
8,340
import fnmatch import logging import os import sys import h5py import numpy as np The provided code snippet includes necessary dependencies for implementing the `read_hdf5` function. Write a Python function `def read_hdf5(hdf5_name, hdf5_path)` to solve the following problem: Read hdf5 dataset. Args: hdf5_name (str): Filename of hdf5 file. hdf5_path (str): Dataset name in hdf5 file. Return: any: Dataset values. Here is the function: def read_hdf5(hdf5_name, hdf5_path): """Read hdf5 dataset. Args: hdf5_name (str): Filename of hdf5 file. hdf5_path (str): Dataset name in hdf5 file. Return: any: Dataset values. """ if not os.path.exists(hdf5_name): logging.error(f"There is no such a hdf5 file ({hdf5_name}).") sys.exit(1) hdf5_file = h5py.File(hdf5_name, "r") if hdf5_path not in hdf5_file: logging.error(f"There is no such a data in hdf5 file. ({hdf5_path})") sys.exit(1) hdf5_data = hdf5_file[hdf5_path][()] hdf5_file.close() return hdf5_data
Read hdf5 dataset. Args: hdf5_name (str): Filename of hdf5 file. hdf5_path (str): Dataset name in hdf5 file. Return: any: Dataset values.
8,341
import fnmatch import logging import os import sys import h5py import numpy as np The provided code snippet includes necessary dependencies for implementing the `write_hdf5` function. Write a Python function `def write_hdf5(hdf5_name, hdf5_path, write_data, is_overwrite=True)` to solve the following problem: Write dataset to hdf5. Args: hdf5_name (str): Hdf5 dataset filename. hdf5_path (str): Dataset path in hdf5. write_data (ndarray): Data to write. is_overwrite (bool): Whether to overwrite dataset. Here is the function: def write_hdf5(hdf5_name, hdf5_path, write_data, is_overwrite=True): """Write dataset to hdf5. Args: hdf5_name (str): Hdf5 dataset filename. hdf5_path (str): Dataset path in hdf5. write_data (ndarray): Data to write. is_overwrite (bool): Whether to overwrite dataset. """ # convert to numpy array write_data = np.array(write_data) # check folder existence folder_name, _ = os.path.split(hdf5_name) if not os.path.exists(folder_name) and len(folder_name) != 0: os.makedirs(folder_name) # check hdf5 existence if os.path.exists(hdf5_name): # if already exists, open with r+ mode hdf5_file = h5py.File(hdf5_name, "r+") # check dataset existence if hdf5_path in hdf5_file: if is_overwrite: logging.warning("Dataset in hdf5 file already exists. " "recreate dataset in hdf5.") hdf5_file.__delitem__(hdf5_path) else: logging.error("Dataset in hdf5 file already exists. " "if you want to overwrite, please set is_overwrite = True.") hdf5_file.close() sys.exit(1) else: # if not exists, open with w mode hdf5_file = h5py.File(hdf5_name, "w") # write data to hdf5 hdf5_file.create_dataset(hdf5_path, data=write_data) hdf5_file.flush() hdf5_file.close()
Write dataset to hdf5. Args: hdf5_name (str): Hdf5 dataset filename. hdf5_path (str): Dataset path in hdf5. write_data (ndarray): Data to write. is_overwrite (bool): Whether to overwrite dataset.
8,342
def cpop_pinyin2ph_func(): # In the README file of opencpop dataset, they defined a "pinyin to phoneme mapping table" pinyin2phs = {'AP': 'AP', 'SP': 'SP'} with open('NeuralSeq/inference/svs/opencpop/cpop_pinyin2ph.txt') as rf: for line in rf.readlines(): elements = [x.strip() for x in line.split('|') if x.strip() != ''] pinyin2phs[elements[0]] = elements[1] return pinyin2phs
null
8,343
import importlib VOCODERS = {} def register_vocoder(cls): VOCODERS[cls.__name__.lower()] = cls VOCODERS[cls.__name__] = cls return cls
null
8,344
import importlib VOCODERS = {} def get_vocoder_cls(hparams): if hparams['vocoder'] in VOCODERS: return VOCODERS[hparams['vocoder']] else: vocoder_cls = hparams['vocoder'] pkg = ".".join(vocoder_cls.split(".")[:-1]) cls_name = vocoder_cls.split(".")[-1] vocoder_cls = getattr(importlib.import_module(pkg), cls_name) return vocoder_cls
null
8,345
import librosa from utils.hparams import hparams import numpy as np hparams = {} def denoise(wav, v=0.1): spec = librosa.stft(y=wav, n_fft=hparams['fft_size'], hop_length=hparams['hop_size'], win_length=hparams['win_size'], pad_mode='constant') spec_m = np.abs(spec) spec_m = np.clip(spec_m - v, a_min=0, a_max=None) spec_a = np.angle(spec) return librosa.istft(spec_m * np.exp(1j * spec_a), hop_length=hparams['hop_size'], win_length=hparams['win_size'])
null
8,346
import glob import json import os import re import librosa import torch import utils from modules.hifigan.hifigan import HifiGanGenerator from utils.hparams import hparams, set_hparams from vocoders.base_vocoder import register_vocoder from vocoders.pwg import PWG from vocoders.vocoder_utils import denoise class HifiGanGenerator(torch.nn.Module): def __init__(self, h, c_out=1): super(HifiGanGenerator, self).__init__() self.h = h self.num_kernels = len(h['resblock_kernel_sizes']) self.num_upsamples = len(h['upsample_rates']) if h['use_pitch_embed']: self.harmonic_num = 8 self.f0_upsamp = torch.nn.Upsample(scale_factor=np.prod(h['upsample_rates'])) self.m_source = SourceModuleHnNSF( sampling_rate=h['audio_sample_rate'], harmonic_num=self.harmonic_num) self.noise_convs = nn.ModuleList() self.conv_pre = weight_norm(Conv1d(80, h['upsample_initial_channel'], 7, 1, padding=3)) resblock = ResBlock1 if h['resblock'] == '1' else ResBlock2 self.ups = nn.ModuleList() for i, (u, k) in enumerate(zip(h['upsample_rates'], h['upsample_kernel_sizes'])): c_cur = h['upsample_initial_channel'] // (2 ** (i + 1)) self.ups.append(weight_norm( ConvTranspose1d(c_cur * 2, c_cur, k, u, padding=(k - u) // 2))) if h['use_pitch_embed']: if i + 1 < len(h['upsample_rates']): stride_f0 = np.prod(h['upsample_rates'][i + 1:]) self.noise_convs.append(Conv1d( 1, c_cur, kernel_size=stride_f0 * 2, stride=stride_f0, padding=stride_f0 // 2)) else: self.noise_convs.append(Conv1d(1, c_cur, kernel_size=1)) self.resblocks = nn.ModuleList() for i in range(len(self.ups)): ch = h['upsample_initial_channel'] // (2 ** (i + 1)) for j, (k, d) in enumerate(zip(h['resblock_kernel_sizes'], h['resblock_dilation_sizes'])): self.resblocks.append(resblock(h, ch, k, d)) self.conv_post = weight_norm(Conv1d(ch, c_out, 7, 1, padding=3)) self.ups.apply(init_weights) self.conv_post.apply(init_weights) def forward(self, x, f0=None): if f0 is not None: # harmonic-source signal, noise-source signal, uv flag f0 = self.f0_upsamp(f0[:, None]).transpose(1, 2) har_source, noi_source, uv = self.m_source(f0) har_source = har_source.transpose(1, 2) x = self.conv_pre(x) for i in range(self.num_upsamples): x = F.leaky_relu(x, LRELU_SLOPE) x = self.ups[i](x) if f0 is not None: x_source = self.noise_convs[i](har_source) x = x + x_source xs = None for j in range(self.num_kernels): if xs is None: xs = self.resblocks[i * self.num_kernels + j](x) else: xs += self.resblocks[i * self.num_kernels + j](x) x = xs / self.num_kernels x = F.leaky_relu(x) x = self.conv_post(x) x = torch.tanh(x) return x def remove_weight_norm(self): print('Removing weight norm...') for l in self.ups: remove_weight_norm(l) for l in self.resblocks: l.remove_weight_norm() remove_weight_norm(self.conv_pre) remove_weight_norm(self.conv_post) def set_hparams(config='', exp_name='', hparams_str='', print_hparams=True, global_hparams=True): if config == '' and exp_name == '': parser = argparse.ArgumentParser(description='') parser.add_argument('--config', type=str, default='', help='location of the data corpus') parser.add_argument('--exp_name', type=str, default='', help='exp_name') parser.add_argument('-hp', '--hparams', type=str, default='', help='location of the data corpus') parser.add_argument('--infer', action='store_true', help='infer') parser.add_argument('--validate', action='store_true', help='validate') parser.add_argument('--reset', action='store_true', help='reset hparams') parser.add_argument('--remove', action='store_true', help='remove old ckpt') parser.add_argument('--debug', action='store_true', help='debug') args, unknown = parser.parse_known_args() print("| Unknow hparams: ", unknown) else: args = Args(config=config, exp_name=exp_name, hparams=hparams_str, infer=False, validate=False, reset=False, debug=False, remove=False) global hparams assert args.config != '' or args.exp_name != '' if args.config != '': assert os.path.exists(args.config) config_chains = [] loaded_config = set() def load_config(config_fn): # deep first inheritance and avoid the second visit of one node if not os.path.exists(config_fn): return {} with open(config_fn) as f: hparams_ = yaml.safe_load(f) loaded_config.add(config_fn) if 'base_config' in hparams_: ret_hparams = {} if not isinstance(hparams_['base_config'], list): hparams_['base_config'] = [hparams_['base_config']] for c in hparams_['base_config']: if c.startswith('.'): c = f'{os.path.dirname(config_fn)}/{c}' c = os.path.normpath(c) if c not in loaded_config: override_config(ret_hparams, load_config(c)) override_config(ret_hparams, hparams_) else: ret_hparams = hparams_ config_chains.append(config_fn) return ret_hparams saved_hparams = {} args_work_dir = '' if args.exp_name != '': args_work_dir = f'{args.exp_name}' # modified ckpt_config_path = f'{args_work_dir}/config.yaml' if os.path.exists(ckpt_config_path): with open(ckpt_config_path) as f: saved_hparams_ = yaml.safe_load(f) if saved_hparams_ is not None: saved_hparams.update(saved_hparams_) hparams_ = {} if args.config != '': hparams_.update(load_config(args.config)) if not args.reset: hparams_.update(saved_hparams) hparams_['work_dir'] = args_work_dir # Support config overriding in command line. Support list type config overriding. # Examples: --hparams="a=1,b.c=2,d=[1 1 1]" if args.hparams != "": for new_hparam in args.hparams.split(","): k, v = new_hparam.split("=") v = v.strip("\'\" ") config_node = hparams_ for k_ in k.split(".")[:-1]: config_node = config_node[k_] k = k.split(".")[-1] if v in ['True', 'False'] or type(config_node[k]) in [bool, list, dict]: if type(config_node[k]) == list: v = v.replace(" ", ",") config_node[k] = eval(v) else: config_node[k] = type(config_node[k])(v) if args_work_dir != '' and args.remove: answer = input("REMOVE old checkpoint? Y/N [Default: N]: ") if answer.lower() == "y": remove_file(args_work_dir) if args_work_dir != '' and (not os.path.exists(ckpt_config_path) or args.reset) and not args.infer: os.makedirs(hparams_['work_dir'], exist_ok=True) with open(ckpt_config_path, 'w') as f: yaml.safe_dump(hparams_, f) hparams_['infer'] = args.infer hparams_['debug'] = args.debug hparams_['validate'] = args.validate hparams_['exp_name'] = args.exp_name global global_print_hparams if global_hparams: hparams.clear() hparams.update(hparams_) if print_hparams and global_print_hparams and global_hparams: print('| Hparams chains: ', config_chains) print('| Hparams: ') for i, (k, v) in enumerate(sorted(hparams_.items())): print(f"\033[;33;m{k}\033[0m: {v}, ", end="\n" if i % 5 == 4 else "") print("") global_print_hparams = False return hparams_ def load_model(config_path, checkpoint_path): device = torch.device("cuda" if torch.cuda.is_available() else "cpu") ckpt_dict = torch.load(checkpoint_path, map_location="cpu") if '.yaml' in config_path: config = set_hparams(config_path, global_hparams=False) state = ckpt_dict["state_dict"]["model_gen"] elif '.json' in config_path: config = json.load(open(config_path, 'r')) state = ckpt_dict["generator"] model = HifiGanGenerator(config) model.load_state_dict(state, strict=True) model.remove_weight_norm() model = model.eval().to(device) print(f"| Loaded model parameters from {checkpoint_path}.") print(f"| HifiGAN device: {device}.") return model, config, device
null
8,347
import glob import re import librosa import torch import yaml from sklearn.preprocessing import StandardScaler from torch import nn from modules.parallel_wavegan.models import ParallelWaveGANGenerator from modules.parallel_wavegan.utils import read_hdf5 from utils.hparams import hparams from utils.pitch_utils import f0_to_coarse from vocoders.base_vocoder import BaseVocoder, register_vocoder import numpy as np def load_pwg_model(config_path, checkpoint_path, stats_path): # load config with open(config_path) as f: config = yaml.load(f, Loader=yaml.Loader) # setup if torch.cuda.is_available(): device = torch.device("cuda") else: device = torch.device("cpu") model = ParallelWaveGANGenerator(**config["generator_params"]) ckpt_dict = torch.load(checkpoint_path, map_location="cpu") if 'state_dict' not in ckpt_dict: # official vocoder model.load_state_dict(torch.load(checkpoint_path, map_location="cpu")["model"]["generator"]) scaler = StandardScaler() if config["format"] == "hdf5": scaler.mean_ = read_hdf5(stats_path, "mean") scaler.scale_ = read_hdf5(stats_path, "scale") elif config["format"] == "npy": scaler.mean_ = np.load(stats_path)[0] scaler.scale_ = np.load(stats_path)[1] else: raise ValueError("support only hdf5 or npy format.") else: # custom PWG vocoder fake_task = nn.Module() fake_task.model_gen = model fake_task.load_state_dict(torch.load(checkpoint_path, map_location="cpu")["state_dict"], strict=False) scaler = None model.remove_weight_norm() model = model.eval().to(device) print(f"| Loaded model parameters from {checkpoint_path}.") print(f"| PWG device: {device}.") return model, scaler, config, device
null
8,348
import importlib from utils.hparams import set_hparams, hparams hparams = {} def run_task(): assert hparams['task_cls'] != '' pkg = ".".join(hparams["task_cls"].split(".")[:-1]) cls_name = hparams["task_cls"].split(".")[-1] task_cls = getattr(importlib.import_module(pkg), cls_name) task_cls.start()
null
8,349
import importlib from data_gen.tts.base_binarizer import BaseBinarizer from data_gen.tts.base_preprocess import BasePreprocessor from data_gen.tts.txt_processors.base_text_processor import get_txt_processor_cls from utils.hparams import hparams hparams = {} def parse_dataset_configs(): max_tokens = hparams['max_tokens'] max_sentences = hparams['max_sentences'] max_valid_tokens = hparams['max_valid_tokens'] if max_valid_tokens == -1: hparams['max_valid_tokens'] = max_valid_tokens = max_tokens max_valid_sentences = hparams['max_valid_sentences'] if max_valid_sentences == -1: hparams['max_valid_sentences'] = max_valid_sentences = max_sentences return max_tokens, max_sentences, max_valid_tokens, max_valid_sentences
null
8,350
import importlib from data_gen.tts.base_binarizer import BaseBinarizer from data_gen.tts.base_preprocess import BasePreprocessor from data_gen.tts.txt_processors.base_text_processor import get_txt_processor_cls from utils.hparams import hparams hparams = {} def parse_mel_losses(): mel_losses = hparams['mel_losses'].split("|") loss_and_lambda = {} for i, l in enumerate(mel_losses): if l == '': continue if ':' in l: l, lbd = l.split(":") lbd = float(lbd) else: lbd = 1.0 loss_and_lambda[l] = lbd print("| Mel losses:", loss_and_lambda) return loss_and_lambda
null
8,351
import importlib from data_gen.tts.base_binarizer import BaseBinarizer from data_gen.tts.base_preprocess import BasePreprocessor from data_gen.tts.txt_processors.base_text_processor import get_txt_processor_cls from utils.hparams import hparams class BasePreprocessor: def __init__(self): self.preprocess_args = hparams['preprocess_args'] txt_processor = self.preprocess_args['txt_processor'] self.txt_processor = get_txt_processor_cls(txt_processor) self.raw_data_dir = hparams['raw_data_dir'] self.processed_dir = hparams['processed_data_dir'] self.spk_map_fn = f"{self.processed_dir}/spk_map.json" def meta_data(self): """ :return: {'item_name': Str, 'wav_fn': Str, 'txt': Str, 'spk_name': Str, 'txt_loader': None or Func} """ raise NotImplementedError def process(self): processed_dir = self.processed_dir wav_processed_tmp_dir = f'{processed_dir}/processed_tmp' remove_file(wav_processed_tmp_dir) os.makedirs(wav_processed_tmp_dir, exist_ok=True) wav_processed_dir = f'{processed_dir}/{self.wav_processed_dirname}' remove_file(wav_processed_dir) os.makedirs(wav_processed_dir, exist_ok=True) meta_data = list(tqdm(self.meta_data(), desc='Load meta data')) item_names = [d['item_name'] for d in meta_data] assert len(item_names) == len(set(item_names)), 'Key `item_name` should be Unique.' # preprocess data phone_list = [] word_list = [] spk_names = set() process_item = partial(self.preprocess_first_pass, txt_processor=self.txt_processor, wav_processed_dir=wav_processed_dir, wav_processed_tmp=wav_processed_tmp_dir, preprocess_args=self.preprocess_args) items = [] args = [{ 'item_name': item_raw['item_name'], 'txt_raw': item_raw['txt'], 'wav_fn': item_raw['wav_fn'], 'txt_loader': item_raw.get('txt_loader'), 'others': item_raw.get('others', None) } for item_raw in meta_data] for item_, (item_id, item) in zip(meta_data, multiprocess_run_tqdm(process_item, args, desc='Preprocess')): if item is not None: item_.update(item) item = item_ if 'txt_loader' in item: del item['txt_loader'] item['id'] = item_id item['spk_name'] = item.get('spk_name', '<SINGLE_SPK>') item['others'] = item.get('others', None) phone_list += item['ph'].split(" ") word_list += item['word'].split(" ") spk_names.add(item['spk_name']) items.append(item) # add encoded tokens ph_encoder, word_encoder = self._phone_encoder(phone_list), self._word_encoder(word_list) spk_map = self.build_spk_map(spk_names) args = [{ 'ph': item['ph'], 'word': item['word'], 'spk_name': item['spk_name'], 'word_encoder': word_encoder, 'ph_encoder': ph_encoder, 'spk_map': spk_map } for item in items] for idx, item_new_kv in multiprocess_run_tqdm(self.preprocess_second_pass, args, desc='Add encoded tokens'): items[idx].update(item_new_kv) # build mfa data if self.preprocess_args['use_mfa']: mfa_dict = set() mfa_input_dir = f'{processed_dir}/mfa_inputs' remove_file(mfa_input_dir) # group MFA inputs for better parallelism mfa_groups = [i // self.preprocess_args['nsample_per_mfa_group'] for i in range(len(items))] if self.preprocess_args['mfa_group_shuffle']: random.seed(hparams['seed']) random.shuffle(mfa_groups) args = [{ 'item': item, 'mfa_input_dir': mfa_input_dir, 'mfa_group': mfa_group, 'wav_processed_tmp': wav_processed_tmp_dir, 'preprocess_args': self.preprocess_args } for item, mfa_group in zip(items, mfa_groups)] for i, (ph_gb_word_nosil, new_wav_align_fn) in multiprocess_run_tqdm( self.build_mfa_inputs, args, desc='Build MFA data'): items[i]['wav_align_fn'] = new_wav_align_fn for w in ph_gb_word_nosil.split(" "): mfa_dict.add(f"{w} {w.replace('_', ' ')}") mfa_dict = sorted(mfa_dict) with open(f'{processed_dir}/mfa_dict.txt', 'w') as f: f.writelines([f'{l}\n' for l in mfa_dict]) with open(f"{processed_dir}/{self.meta_csv_filename}.json", 'w') as f: f.write(re.sub(r'\n\s+([\d+\]])', r'\1', json.dumps(items, ensure_ascii=False, sort_keys=False, indent=1))) remove_file(wav_processed_tmp_dir) def preprocess_first_pass(cls, item_name, txt_raw, txt_processor, wav_fn, wav_processed_dir, wav_processed_tmp, preprocess_args, txt_loader=None, others=None): try: if txt_loader is not None: txt_raw = txt_loader(txt_raw) ph, txt, word, ph2word, ph_gb_word = cls.txt_to_ph(txt_processor, txt_raw, preprocess_args) wav_fn, wav_align_fn = cls.process_wav( item_name, wav_fn, hparams['processed_data_dir'], wav_processed_tmp, preprocess_args) # wav for binarization ext = os.path.splitext(wav_fn)[1] os.makedirs(wav_processed_dir, exist_ok=True) new_wav_fn = f"{wav_processed_dir}/{item_name}{ext}" move_link_func = move_file if os.path.dirname(wav_fn) == wav_processed_tmp else link_file move_link_func(wav_fn, new_wav_fn) return { 'txt': txt, 'txt_raw': txt_raw, 'ph': ph, 'word': word, 'ph2word': ph2word, 'ph_gb_word': ph_gb_word, 'wav_fn': new_wav_fn, 'wav_align_fn': wav_align_fn, 'others': others } except: traceback.print_exc() print(f"| Error is caught. item_name: {item_name}.") return None def txt_to_ph(txt_processor, txt_raw, preprocess_args): txt_struct, txt = txt_processor.process(txt_raw, preprocess_args) ph = [p for w in txt_struct for p in w[1]] ph_gb_word = ["_".join(w[1]) for w in txt_struct] words = [w[0] for w in txt_struct] # word_id=0 is reserved for padding ph2word = [w_id + 1 for w_id, w in enumerate(txt_struct) for _ in range(len(w[1]))] return " ".join(ph), txt, " ".join(words), ph2word, " ".join(ph_gb_word) def process_wav(item_name, wav_fn, processed_dir, wav_processed_tmp, preprocess_args): processors = [get_wav_processor_cls(v) for v in preprocess_args['wav_processors']] processors = [k() for k in processors if k is not None] if len(processors) >= 1: sr_file = librosa.core.get_samplerate(wav_fn) output_fn_for_align = None ext = os.path.splitext(wav_fn)[1] input_fn = f"{wav_processed_tmp}/{item_name}{ext}" link_file(wav_fn, input_fn) for p in processors: outputs = p.process(input_fn, sr_file, wav_processed_tmp, processed_dir, item_name, preprocess_args) if len(outputs) == 3: input_fn, sr, output_fn_for_align = outputs else: input_fn, sr = outputs if output_fn_for_align is None: return input_fn, input_fn else: return input_fn, output_fn_for_align else: return wav_fn, wav_fn def _phone_encoder(self, ph_set): ph_set_fn = f"{self.processed_dir}/phone_set.json" if self.preprocess_args['reset_phone_dict'] or not os.path.exists(ph_set_fn): ph_set = sorted(set(ph_set)) json.dump(ph_set, open(ph_set_fn, 'w'), ensure_ascii=False) print("| Build phone set: ", ph_set) else: ph_set = json.load(open(ph_set_fn, 'r')) print("| Load phone set: ", ph_set) return build_token_encoder(ph_set_fn) def _word_encoder(self, word_set): word_set_fn = f"{self.processed_dir}/word_set.json" if self.preprocess_args['reset_word_dict']: word_set = Counter(word_set) total_words = sum(word_set.values()) word_set = word_set.most_common(hparams['word_dict_size']) num_unk_words = total_words - sum([x[1] for x in word_set]) word_set = ['<BOS>', '<EOS>'] + [x[0] for x in word_set] word_set = sorted(set(word_set)) json.dump(word_set, open(word_set_fn, 'w'), ensure_ascii=False) print(f"| Build word set. Size: {len(word_set)}, #total words: {total_words}," f" #unk_words: {num_unk_words}, word_set[:10]:, {word_set[:10]}.") else: word_set = json.load(open(word_set_fn, 'r')) print("| Load word set. Size: ", len(word_set), word_set[:10]) return build_token_encoder(word_set_fn) def preprocess_second_pass(cls, word, ph, spk_name, word_encoder, ph_encoder, spk_map): word_token = word_encoder.encode(word) ph_token = ph_encoder.encode(ph) spk_id = spk_map[spk_name] return {'word_token': word_token, 'ph_token': ph_token, 'spk_id': spk_id} def build_spk_map(self, spk_names): spk_map = {x: i for i, x in enumerate(sorted(list(spk_names)))} assert len(spk_map) == 0 or len(spk_map) <= hparams['num_spk'], len(spk_map) print(f"| Number of spks: {len(spk_map)}, spk_map: {spk_map}") json.dump(spk_map, open(self.spk_map_fn, 'w'), ensure_ascii=False) return spk_map def build_mfa_inputs(cls, item, mfa_input_dir, mfa_group, wav_processed_tmp, preprocess_args): item_name = item['item_name'] wav_align_fn = item['wav_align_fn'] ph_gb_word = item['ph_gb_word'] ext = os.path.splitext(wav_align_fn)[1] mfa_input_group_dir = f'{mfa_input_dir}/{mfa_group}' os.makedirs(mfa_input_group_dir, exist_ok=True) new_wav_align_fn = f"{mfa_input_group_dir}/{item_name}{ext}" move_link_func = move_file if os.path.dirname(wav_align_fn) == wav_processed_tmp else link_file move_link_func(wav_align_fn, new_wav_align_fn) ph_gb_word_nosil = " ".join(["_".join([p for p in w.split("_") if not is_sil_phoneme(p)]) for w in ph_gb_word.split(" ") if not is_sil_phoneme(w)]) with open(f'{mfa_input_group_dir}/{item_name}.lab', 'w') as f_txt: f_txt.write(ph_gb_word_nosil) return ph_gb_word_nosil, new_wav_align_fn def load_spk_map(self, base_dir): spk_map_fn = f"{base_dir}/spk_map.json" spk_map = json.load(open(spk_map_fn, 'r')) return spk_map def load_dict(self, base_dir): ph_encoder = build_token_encoder(f'{base_dir}/phone_set.json') word_encoder = build_token_encoder(f'{base_dir}/word_set.json') return ph_encoder, word_encoder def meta_csv_filename(self): return 'metadata' def wav_processed_dirname(self): return 'wav_processed' hparams = {} def load_data_preprocessor(): preprocess_cls = hparams["preprocess_cls"] pkg = ".".join(preprocess_cls.split(".")[:-1]) cls_name = preprocess_cls.split(".")[-1] preprocessor: BasePreprocessor = getattr(importlib.import_module(pkg), cls_name)() preprocess_args = {} preprocess_args.update(hparams['preprocess_args']) return preprocessor, preprocess_args
null
8,352
import importlib from data_gen.tts.base_binarizer import BaseBinarizer from data_gen.tts.base_preprocess import BasePreprocessor from data_gen.tts.txt_processors.base_text_processor import get_txt_processor_cls from utils.hparams import hparams class BaseBinarizer: def __init__(self, processed_data_dir=None): def train_item_names(self): def valid_item_names(self): def test_item_names(self): def build_spk_map(self): def item_name2spk_id(self, item_name): def _phone_encoder(self): def meta_data(self, prefix): def process(self): def process_data(self, prefix): def process_item(cls, item_name, ph, txt, tg_fn, wav_fn, spk_id, encoder, binarization_args): def get_align(tg_fn, ph, mel, phone_encoded, res): def get_pitch(wav, mel, res): def get_f0cwt(f0, res): hparams = {} def load_data_binarizer(): binarizer_cls = hparams['binarizer_cls'] pkg = ".".join(binarizer_cls.split(".")[:-1]) cls_name = binarizer_cls.split(".")[-1] binarizer: BaseBinarizer = getattr(importlib.import_module(pkg), cls_name)() binarization_args = {} binarization_args.update(hparams['binarization_args']) return binarizer, binarization_args
null
8,353
REGISTERED_WAV_PROCESSORS = {} def register_wav_processors(name): def _f(cls): REGISTERED_WAV_PROCESSORS[name] = cls return cls return _f
null
8,354
REGISTERED_WAV_PROCESSORS = {} def get_wav_processor_cls(name): return REGISTERED_WAV_PROCESSORS.get(name, None)
null
8,355
import warnings import parselmouth import os import torch from skimage.transform import resize from utils.text_encoder import TokenTextEncoder from utils.pitch_utils import f0_to_coarse import struct import webrtcvad from scipy.ndimage.morphology import binary_dilation import librosa import numpy as np from utils import audio import pyloudnorm as pyln import re import json from collections import OrderedDict def trim_long_silences(path, sr=None, return_raw_wav=False, norm=True, vad_max_silence_length=12): """ Ensures that segments without voice in the waveform remain no longer than a threshold determined by the VAD parameters in params.py. :param wav: the raw waveform as a numpy array of floats :param vad_max_silence_length: Maximum number of consecutive silent frames a segment can have. :return: the same waveform with silences trimmed away (length <= original wav length) """ ## Voice Activation Detection # Window size of the VAD. Must be either 10, 20 or 30 milliseconds. # This sets the granularity of the VAD. Should not need to be changed. sampling_rate = 16000 wav_raw, sr = librosa.core.load(path, sr=sr) if norm: meter = pyln.Meter(sr) # create BS.1770 meter loudness = meter.integrated_loudness(wav_raw) wav_raw = pyln.normalize.loudness(wav_raw, loudness, -20.0) if np.abs(wav_raw).max() > 1.0: wav_raw = wav_raw / np.abs(wav_raw).max() wav = librosa.resample(wav_raw, sr, sampling_rate, res_type='kaiser_best') vad_window_length = 30 # In milliseconds # Number of frames to average together when performing the moving average smoothing. # The larger this value, the larger the VAD variations must be to not get smoothed out. vad_moving_average_width = 8 # Compute the voice detection window size samples_per_window = (vad_window_length * sampling_rate) // 1000 # Trim the end of the audio to have a multiple of the window size wav = wav[:len(wav) - (len(wav) % samples_per_window)] # Convert the float waveform to 16-bit mono PCM pcm_wave = struct.pack("%dh" % len(wav), *(np.round(wav * int16_max)).astype(np.int16)) # Perform voice activation detection voice_flags = [] vad = webrtcvad.Vad(mode=3) for window_start in range(0, len(wav), samples_per_window): window_end = window_start + samples_per_window voice_flags.append(vad.is_speech(pcm_wave[window_start * 2:window_end * 2], sample_rate=sampling_rate)) voice_flags = np.array(voice_flags) # Smooth the voice detection with a moving average def moving_average(array, width): array_padded = np.concatenate((np.zeros((width - 1) // 2), array, np.zeros(width // 2))) ret = np.cumsum(array_padded, dtype=float) ret[width:] = ret[width:] - ret[:-width] return ret[width - 1:] / width audio_mask = moving_average(voice_flags, vad_moving_average_width) audio_mask = np.round(audio_mask).astype(np.bool) # Dilate the voiced regions audio_mask = binary_dilation(audio_mask, np.ones(vad_max_silence_length + 1)) audio_mask = np.repeat(audio_mask, samples_per_window) audio_mask = resize(audio_mask, (len(wav_raw),)) > 0 if return_raw_wav: return wav_raw, audio_mask, sr return wav_raw[audio_mask], audio_mask, sr def process_utterance(wav_path, fft_size=1024, hop_size=256, win_length=1024, window="hann", num_mels=80, fmin=80, fmax=7600, eps=1e-6, sample_rate=22050, loud_norm=False, min_level_db=-100, return_linear=False, trim_long_sil=False, vocoder='pwg'): if isinstance(wav_path, str): if trim_long_sil: wav, _, _ = trim_long_silences(wav_path, sample_rate) else: wav, _ = librosa.core.load(wav_path, sr=sample_rate) else: wav = wav_path if loud_norm: meter = pyln.Meter(sample_rate) # create BS.1770 meter loudness = meter.integrated_loudness(wav) wav = pyln.normalize.loudness(wav, loudness, -22.0) if np.abs(wav).max() > 1: wav = wav / np.abs(wav).max() # get amplitude spectrogram x_stft = librosa.stft(wav, n_fft=fft_size, hop_length=hop_size, win_length=win_length, window=window, pad_mode="constant") spc = np.abs(x_stft) # (n_bins, T) # get mel basis fmin = 0 if fmin == -1 else fmin fmax = sample_rate / 2 if fmax == -1 else fmax mel_basis = librosa.filters.mel(sample_rate, fft_size, num_mels, fmin, fmax) mel = mel_basis @ spc if vocoder == 'pwg': mel = np.log10(np.maximum(eps, mel)) # (n_mel_bins, T) else: assert False, f'"{vocoder}" is not in ["pwg"].' l_pad, r_pad = audio.librosa_pad_lr(wav, fft_size, hop_size, 1) wav = np.pad(wav, (l_pad, r_pad), mode='constant', constant_values=0.0) wav = wav[:mel.shape[1] * hop_size] if not return_linear: return wav, mel else: spc = audio.amp_to_db(spc) spc = audio.normalize(spc, {'min_level_db': min_level_db}) return wav, mel, spc
null
8,356
import warnings import parselmouth import os import torch from skimage.transform import resize from utils.text_encoder import TokenTextEncoder from utils.pitch_utils import f0_to_coarse import struct import webrtcvad from scipy.ndimage.morphology import binary_dilation import librosa import numpy as np from utils import audio import pyloudnorm as pyln import re import json from collections import OrderedDict def f0_to_coarse(f0): is_torch = isinstance(f0, torch.Tensor) f0_mel = 1127 * (1 + f0 / 700).log() if is_torch else 1127 * np.log(1 + f0 / 700) f0_mel[f0_mel > 0] = (f0_mel[f0_mel > 0] - f0_mel_min) * (f0_bin - 2) / (f0_mel_max - f0_mel_min) + 1 f0_mel[f0_mel <= 1] = 1 f0_mel[f0_mel > f0_bin - 1] = f0_bin - 1 f0_coarse = (f0_mel + 0.5).long() if is_torch else np.rint(f0_mel).astype(np.int) assert f0_coarse.max() <= 255 and f0_coarse.min() >= 1, (f0_coarse.max(), f0_coarse.min()) return f0_coarse The provided code snippet includes necessary dependencies for implementing the `get_pitch` function. Write a Python function `def get_pitch(wav_data, mel, hparams)` to solve the following problem: :param wav_data: [T] :param mel: [T, 80] :param hparams: :return: Here is the function: def get_pitch(wav_data, mel, hparams): """ :param wav_data: [T] :param mel: [T, 80] :param hparams: :return: """ time_step = hparams['hop_size'] / hparams['audio_sample_rate'] * 1000 f0_min = 80 f0_max = 750 if hparams['hop_size'] == 128: pad_size = 4 elif hparams['hop_size'] == 256: pad_size = 2 else: assert False f0 = parselmouth.Sound(wav_data, hparams['audio_sample_rate']).to_pitch_ac( time_step=time_step / 1000, voicing_threshold=0.6, pitch_floor=f0_min, pitch_ceiling=f0_max).selected_array['frequency'] lpad = pad_size * 2 rpad = len(mel) - len(f0) - lpad f0 = np.pad(f0, [[lpad, rpad]], mode='constant') # mel and f0 are extracted by 2 different libraries. we should force them to have the same length. # Attention: we find that new version of some libraries could cause ``rpad'' to be a negetive value... # Just to be sure, we recommend users to set up the same environments as them in requirements_auto.txt (by Anaconda) delta_l = len(mel) - len(f0) assert np.abs(delta_l) <= 8 if delta_l > 0: f0 = np.concatenate([f0, [f0[-1]] * delta_l], 0) f0 = f0[:len(mel)] pitch_coarse = f0_to_coarse(f0) return f0, pitch_coarse
:param wav_data: [T] :param mel: [T, 80] :param hparams: :return:
8,357
import warnings import parselmouth import os import torch from skimage.transform import resize from utils.text_encoder import TokenTextEncoder from utils.pitch_utils import f0_to_coarse import struct import webrtcvad from scipy.ndimage.morphology import binary_dilation import librosa import numpy as np from utils import audio import pyloudnorm as pyln import re import json from collections import OrderedDict def remove_empty_lines(text): class TextGrid(object): def __init__(self, text): def _extract_pattern(self, pattern, inc): def _get_type(self): def _get_time_intval(self): def _get_size(self): def _get_item_list(self): def toJson(self): def is_sil_phoneme(p): def get_mel2ph(tg_fn, ph, mel, hparams): ph_list = ph.split(" ") with open(tg_fn, "r") as f: tg = f.readlines() tg = remove_empty_lines(tg) tg = TextGrid(tg) tg = json.loads(tg.toJson()) split = np.ones(len(ph_list) + 1, np.float) * -1 tg_idx = 0 ph_idx = 0 tg_align = [x for x in tg['tiers'][-1]['items']] tg_align_ = [] for x in tg_align: x['xmin'] = float(x['xmin']) x['xmax'] = float(x['xmax']) if x['text'] in ['sil', 'sp', '', 'SIL', 'PUNC']: x['text'] = '' if len(tg_align_) > 0 and tg_align_[-1]['text'] == '': tg_align_[-1]['xmax'] = x['xmax'] continue tg_align_.append(x) tg_align = tg_align_ tg_len = len([x for x in tg_align if x['text'] != '']) ph_len = len([x for x in ph_list if not is_sil_phoneme(x)]) assert tg_len == ph_len, (tg_len, ph_len, tg_align, ph_list, tg_fn) while tg_idx < len(tg_align) or ph_idx < len(ph_list): if tg_idx == len(tg_align) and is_sil_phoneme(ph_list[ph_idx]): split[ph_idx] = 1e8 ph_idx += 1 continue x = tg_align[tg_idx] if x['text'] == '' and ph_idx == len(ph_list): tg_idx += 1 continue assert ph_idx < len(ph_list), (tg_len, ph_len, tg_align, ph_list, tg_fn) ph = ph_list[ph_idx] if x['text'] == '' and not is_sil_phoneme(ph): assert False, (ph_list, tg_align) if x['text'] != '' and is_sil_phoneme(ph): ph_idx += 1 else: assert (x['text'] == '' and is_sil_phoneme(ph)) \ or x['text'].lower() == ph.lower() \ or x['text'].lower() == 'sil', (x['text'], ph) split[ph_idx] = x['xmin'] if ph_idx > 0 and split[ph_idx - 1] == -1 and is_sil_phoneme(ph_list[ph_idx - 1]): split[ph_idx - 1] = split[ph_idx] ph_idx += 1 tg_idx += 1 assert tg_idx == len(tg_align), (tg_idx, [x['text'] for x in tg_align]) assert ph_idx >= len(ph_list) - 1, (ph_idx, ph_list, len(ph_list), [x['text'] for x in tg_align], tg_fn) mel2ph = np.zeros([mel.shape[0]], np.int) split[0] = 0 split[-1] = 1e8 for i in range(len(split) - 1): assert split[i] != -1 and split[i] <= split[i + 1], (split[:-1],) split = [int(s * hparams['audio_sample_rate'] / hparams['hop_size'] + 0.5) for s in split] for ph_idx in range(len(ph_list)): mel2ph[split[ph_idx]:split[ph_idx + 1]] = ph_idx + 1 mel2ph_torch = torch.from_numpy(mel2ph) T_t = len(ph_list) dur = mel2ph_torch.new_zeros([T_t + 1]).scatter_add(0, mel2ph_torch, torch.ones_like(mel2ph_torch)) dur = dur[1:].numpy() return mel2ph, dur
null
8,358
import warnings import parselmouth import os import torch from skimage.transform import resize from utils.text_encoder import TokenTextEncoder from utils.pitch_utils import f0_to_coarse import struct import webrtcvad from scipy.ndimage.morphology import binary_dilation import librosa import numpy as np from utils import audio import pyloudnorm as pyln import re import json from collections import OrderedDict class TokenTextEncoder(TextEncoder): """Encoder based on a user-supplied vocabulary (file or list).""" def __init__(self, vocab_filename, reverse=False, vocab_list=None, replace_oov=None, num_reserved_ids=NUM_RESERVED_TOKENS): """Initialize from a file or list, one token per line. Handling of reserved tokens works as follows: - When initializing from a list, we add reserved tokens to the vocab. - When initializing from a file, we do not add reserved tokens to the vocab. - When saving vocab files, we save reserved tokens to the file. Args: vocab_filename: If not None, the full filename to read vocab from. If this is not None, then vocab_list should be None. reverse: Boolean indicating if tokens should be reversed during encoding and decoding. vocab_list: If not None, a list of elements of the vocabulary. If this is not None, then vocab_filename should be None. replace_oov: If not None, every out-of-vocabulary token seen when encoding will be replaced by this string (which must be in vocab). num_reserved_ids: Number of IDs to save for reserved tokens like <EOS>. """ super(TokenTextEncoder, self).__init__(num_reserved_ids=num_reserved_ids) self._reverse = reverse self._replace_oov = replace_oov if vocab_filename: self._init_vocab_from_file(vocab_filename) else: assert vocab_list is not None self._init_vocab_from_list(vocab_list) self.pad_index = self._token_to_id[PAD] self.eos_index = self._token_to_id[EOS] self.unk_index = self._token_to_id[UNK] self.seg_index = self._token_to_id[SEG] if SEG in self._token_to_id else self.eos_index def encode(self, s): """Converts a space-separated string of tokens to a list of ids.""" sentence = s tokens = sentence.strip().split() if self._replace_oov is not None: tokens = [t if t in self._token_to_id else self._replace_oov for t in tokens] ret = [self._token_to_id[tok] for tok in tokens] return ret[::-1] if self._reverse else ret def decode(self, ids, strip_eos=False, strip_padding=False): if strip_padding and self.pad() in list(ids): pad_pos = list(ids).index(self.pad()) ids = ids[:pad_pos] if strip_eos and self.eos() in list(ids): eos_pos = list(ids).index(self.eos()) ids = ids[:eos_pos] return " ".join(self.decode_list(ids)) def decode_list(self, ids): seq = reversed(ids) if self._reverse else ids return [self._safe_id_to_token(i) for i in seq] def vocab_size(self): return len(self._id_to_token) def __len__(self): return self.vocab_size def _safe_id_to_token(self, idx): return self._id_to_token.get(idx, "ID_%d" % idx) def _init_vocab_from_file(self, filename): """Load vocab from a file. Args: filename: The file to load vocabulary from. """ with open(filename) as f: tokens = [token.strip() for token in f.readlines()] def token_gen(): for token in tokens: yield token self._init_vocab(token_gen(), add_reserved_tokens=False) def _init_vocab_from_list(self, vocab_list): """Initialize tokens from a list of tokens. It is ok if reserved tokens appear in the vocab list. They will be removed. The set of tokens in vocab_list should be unique. Args: vocab_list: A list of tokens. """ def token_gen(): for token in vocab_list: if token not in RESERVED_TOKENS: yield token self._init_vocab(token_gen()) def _init_vocab(self, token_generator, add_reserved_tokens=True): """Initialize vocabulary with tokens from token_generator.""" self._id_to_token = {} non_reserved_start_index = 0 if add_reserved_tokens: self._id_to_token.update(enumerate(RESERVED_TOKENS)) non_reserved_start_index = len(RESERVED_TOKENS) self._id_to_token.update( enumerate(token_generator, start=non_reserved_start_index)) # _token_to_id is the reverse of _id_to_token self._token_to_id = dict((v, k) for k, v in six.iteritems(self._id_to_token)) def pad(self): return self.pad_index def eos(self): return self.eos_index def unk(self): return self.unk_index def seg(self): return self.seg_index def store_to_file(self, filename): """Write vocab file to disk. Vocab files have one token per line. The file ends in a newline. Reserved tokens are written to the vocab file as well. Args: filename: Full path of the file to store the vocab to. """ with open(filename, "w") as f: for i in range(len(self._id_to_token)): f.write(self._id_to_token[i] + "\n") def sil_phonemes(self): return [p for p in self._id_to_token.values() if not p[0].isalpha()] def build_phone_encoder(data_dir): phone_list_file = os.path.join(data_dir, 'phone_set.json') phone_list = json.load(open(phone_list_file)) return TokenTextEncoder(None, vocab_list=phone_list, replace_oov=',')
null
8,359
import warnings import parselmouth import os import torch from skimage.transform import resize from utils.text_encoder import TokenTextEncoder from utils.pitch_utils import f0_to_coarse import struct import webrtcvad from scipy.ndimage.morphology import binary_dilation import librosa import numpy as np from utils import audio import pyloudnorm as pyln import re import json from collections import OrderedDict class TokenTextEncoder(TextEncoder): """Encoder based on a user-supplied vocabulary (file or list).""" def __init__(self, vocab_filename, reverse=False, vocab_list=None, replace_oov=None, num_reserved_ids=NUM_RESERVED_TOKENS): """Initialize from a file or list, one token per line. Handling of reserved tokens works as follows: - When initializing from a list, we add reserved tokens to the vocab. - When initializing from a file, we do not add reserved tokens to the vocab. - When saving vocab files, we save reserved tokens to the file. Args: vocab_filename: If not None, the full filename to read vocab from. If this is not None, then vocab_list should be None. reverse: Boolean indicating if tokens should be reversed during encoding and decoding. vocab_list: If not None, a list of elements of the vocabulary. If this is not None, then vocab_filename should be None. replace_oov: If not None, every out-of-vocabulary token seen when encoding will be replaced by this string (which must be in vocab). num_reserved_ids: Number of IDs to save for reserved tokens like <EOS>. """ super(TokenTextEncoder, self).__init__(num_reserved_ids=num_reserved_ids) self._reverse = reverse self._replace_oov = replace_oov if vocab_filename: self._init_vocab_from_file(vocab_filename) else: assert vocab_list is not None self._init_vocab_from_list(vocab_list) self.pad_index = self._token_to_id[PAD] self.eos_index = self._token_to_id[EOS] self.unk_index = self._token_to_id[UNK] self.seg_index = self._token_to_id[SEG] if SEG in self._token_to_id else self.eos_index def encode(self, s): """Converts a space-separated string of tokens to a list of ids.""" sentence = s tokens = sentence.strip().split() if self._replace_oov is not None: tokens = [t if t in self._token_to_id else self._replace_oov for t in tokens] ret = [self._token_to_id[tok] for tok in tokens] return ret[::-1] if self._reverse else ret def decode(self, ids, strip_eos=False, strip_padding=False): if strip_padding and self.pad() in list(ids): pad_pos = list(ids).index(self.pad()) ids = ids[:pad_pos] if strip_eos and self.eos() in list(ids): eos_pos = list(ids).index(self.eos()) ids = ids[:eos_pos] return " ".join(self.decode_list(ids)) def decode_list(self, ids): seq = reversed(ids) if self._reverse else ids return [self._safe_id_to_token(i) for i in seq] def vocab_size(self): return len(self._id_to_token) def __len__(self): return self.vocab_size def _safe_id_to_token(self, idx): return self._id_to_token.get(idx, "ID_%d" % idx) def _init_vocab_from_file(self, filename): """Load vocab from a file. Args: filename: The file to load vocabulary from. """ with open(filename) as f: tokens = [token.strip() for token in f.readlines()] def token_gen(): for token in tokens: yield token self._init_vocab(token_gen(), add_reserved_tokens=False) def _init_vocab_from_list(self, vocab_list): """Initialize tokens from a list of tokens. It is ok if reserved tokens appear in the vocab list. They will be removed. The set of tokens in vocab_list should be unique. Args: vocab_list: A list of tokens. """ def token_gen(): for token in vocab_list: if token not in RESERVED_TOKENS: yield token self._init_vocab(token_gen()) def _init_vocab(self, token_generator, add_reserved_tokens=True): """Initialize vocabulary with tokens from token_generator.""" self._id_to_token = {} non_reserved_start_index = 0 if add_reserved_tokens: self._id_to_token.update(enumerate(RESERVED_TOKENS)) non_reserved_start_index = len(RESERVED_TOKENS) self._id_to_token.update( enumerate(token_generator, start=non_reserved_start_index)) # _token_to_id is the reverse of _id_to_token self._token_to_id = dict((v, k) for k, v in six.iteritems(self._id_to_token)) def pad(self): return self.pad_index def eos(self): return self.eos_index def unk(self): return self.unk_index def seg(self): return self.seg_index def store_to_file(self, filename): """Write vocab file to disk. Vocab files have one token per line. The file ends in a newline. Reserved tokens are written to the vocab file as well. Args: filename: Full path of the file to store the vocab to. """ with open(filename, "w") as f: for i in range(len(self._id_to_token)): f.write(self._id_to_token[i] + "\n") def sil_phonemes(self): return [p for p in self._id_to_token.values() if not p[0].isalpha()] def build_word_encoder(data_dir): word_list_file = os.path.join(data_dir, 'word_set.json') word_list = json.load(open(word_list_file)) return TokenTextEncoder(None, vocab_list=word_list, replace_oov=',')
null
8,360
import warnings import parselmouth import os import torch from skimage.transform import resize from utils.text_encoder import TokenTextEncoder from utils.pitch_utils import f0_to_coarse import struct import webrtcvad from scipy.ndimage.morphology import binary_dilation import librosa import numpy as np from utils import audio import pyloudnorm as pyln import re import json from collections import OrderedDict class TokenTextEncoder(TextEncoder): """Encoder based on a user-supplied vocabulary (file or list).""" def __init__(self, vocab_filename, reverse=False, vocab_list=None, replace_oov=None, num_reserved_ids=NUM_RESERVED_TOKENS): """Initialize from a file or list, one token per line. Handling of reserved tokens works as follows: - When initializing from a list, we add reserved tokens to the vocab. - When initializing from a file, we do not add reserved tokens to the vocab. - When saving vocab files, we save reserved tokens to the file. Args: vocab_filename: If not None, the full filename to read vocab from. If this is not None, then vocab_list should be None. reverse: Boolean indicating if tokens should be reversed during encoding and decoding. vocab_list: If not None, a list of elements of the vocabulary. If this is not None, then vocab_filename should be None. replace_oov: If not None, every out-of-vocabulary token seen when encoding will be replaced by this string (which must be in vocab). num_reserved_ids: Number of IDs to save for reserved tokens like <EOS>. """ super(TokenTextEncoder, self).__init__(num_reserved_ids=num_reserved_ids) self._reverse = reverse self._replace_oov = replace_oov if vocab_filename: self._init_vocab_from_file(vocab_filename) else: assert vocab_list is not None self._init_vocab_from_list(vocab_list) self.pad_index = self._token_to_id[PAD] self.eos_index = self._token_to_id[EOS] self.unk_index = self._token_to_id[UNK] self.seg_index = self._token_to_id[SEG] if SEG in self._token_to_id else self.eos_index def encode(self, s): """Converts a space-separated string of tokens to a list of ids.""" sentence = s tokens = sentence.strip().split() if self._replace_oov is not None: tokens = [t if t in self._token_to_id else self._replace_oov for t in tokens] ret = [self._token_to_id[tok] for tok in tokens] return ret[::-1] if self._reverse else ret def decode(self, ids, strip_eos=False, strip_padding=False): if strip_padding and self.pad() in list(ids): pad_pos = list(ids).index(self.pad()) ids = ids[:pad_pos] if strip_eos and self.eos() in list(ids): eos_pos = list(ids).index(self.eos()) ids = ids[:eos_pos] return " ".join(self.decode_list(ids)) def decode_list(self, ids): seq = reversed(ids) if self._reverse else ids return [self._safe_id_to_token(i) for i in seq] def vocab_size(self): return len(self._id_to_token) def __len__(self): return self.vocab_size def _safe_id_to_token(self, idx): return self._id_to_token.get(idx, "ID_%d" % idx) def _init_vocab_from_file(self, filename): """Load vocab from a file. Args: filename: The file to load vocabulary from. """ with open(filename) as f: tokens = [token.strip() for token in f.readlines()] def token_gen(): for token in tokens: yield token self._init_vocab(token_gen(), add_reserved_tokens=False) def _init_vocab_from_list(self, vocab_list): """Initialize tokens from a list of tokens. It is ok if reserved tokens appear in the vocab list. They will be removed. The set of tokens in vocab_list should be unique. Args: vocab_list: A list of tokens. """ def token_gen(): for token in vocab_list: if token not in RESERVED_TOKENS: yield token self._init_vocab(token_gen()) def _init_vocab(self, token_generator, add_reserved_tokens=True): """Initialize vocabulary with tokens from token_generator.""" self._id_to_token = {} non_reserved_start_index = 0 if add_reserved_tokens: self._id_to_token.update(enumerate(RESERVED_TOKENS)) non_reserved_start_index = len(RESERVED_TOKENS) self._id_to_token.update( enumerate(token_generator, start=non_reserved_start_index)) # _token_to_id is the reverse of _id_to_token self._token_to_id = dict((v, k) for k, v in six.iteritems(self._id_to_token)) def pad(self): return self.pad_index def eos(self): return self.eos_index def unk(self): return self.unk_index def seg(self): return self.seg_index def store_to_file(self, filename): """Write vocab file to disk. Vocab files have one token per line. The file ends in a newline. Reserved tokens are written to the vocab file as well. Args: filename: Full path of the file to store the vocab to. """ with open(filename, "w") as f: for i in range(len(self._id_to_token)): f.write(self._id_to_token[i] + "\n") def sil_phonemes(self): return [p for p in self._id_to_token.values() if not p[0].isalpha()] def build_token_encoder(token_list_file): token_list = json.load(open(token_list_file)) return TokenTextEncoder(None, vocab_list=token_list, replace_oov='<UNK>')
null
8,361
from data_gen.tts.emotion.params_data import * from data_gen.tts.emotion.model import EmotionEncoder from data_gen.tts.emotion.audio import preprocess_wav from matplotlib import cm from data_gen.tts.emotion import audio from pathlib import Path import matplotlib.pyplot as plt import numpy as np import torch _model = None _device = None class EmotionEncoder(nn.Module): def __init__(self, device, loss_device): super().__init__() self.loss_device = loss_device # Network defition self.lstm = nn.LSTM(input_size=mel_n_channels, hidden_size=model_hidden_size, num_layers=model_num_layers, batch_first=True).to(device) self.linear = nn.Linear(in_features=model_hidden_size, out_features=model_embedding_size).to(device) self.relu = torch.nn.ReLU().to(device) # Cosine similarity scaling (with fixed initial parameter values) self.similarity_weight = nn.Parameter(torch.tensor([10.])).to(loss_device) self.similarity_bias = nn.Parameter(torch.tensor([-5.])).to(loss_device) # Loss self.loss_fn = nn.CrossEntropyLoss().to(loss_device) def do_gradient_ops(self): # Gradient scale self.similarity_weight.grad *= 0.01 self.similarity_bias.grad *= 0.01 # Gradient clipping clip_grad_norm_(self.parameters(), 3, norm_type=2) def forward(self, utterances, hidden_init=None): """ Computes the embeddings of a batch of utterance spectrograms. :param utterances: batch of mel-scale filterbanks of same duration as a tensor of shape (batch_size, n_frames, n_channels) :param hidden_init: initial hidden state of the LSTM as a tensor of shape (num_layers, batch_size, hidden_size). Will default to a tensor of zeros if None. :return: the embeddings as a tensor of shape (batch_size, embedding_size) """ # Pass the input through the LSTM layers and retrieve all outputs, the final hidden state # and the final cell state. out, (hidden, cell) = self.lstm(utterances, hidden_init) # We take only the hidden state of the last layer embeds_raw = self.relu(self.linear(hidden[-1])) # L2-normalize it embeds = embeds_raw / torch.norm(embeds_raw, dim=1, keepdim=True) return embeds def inference(self, utterances, hidden_init=None): """ Computes the embeddings of a batch of utterance spectrograms. :param utterances: batch of mel-scale filterbanks of same duration as a tensor of shape (batch_size, n_frames, n_channels) :param hidden_init: initial hidden state of the LSTM as a tensor of shape (num_layers, batch_size, hidden_size). Will default to a tensor of zeros if None. :return: the embeddings as a tensor of shape (batch_size, embedding_size) """ # Pass the input through the LSTM layers and retrieve all outputs, the final hidden state # and the final cell state. out, (hidden, cell) = self.lstm(utterances, hidden_init) return hidden[-1] The provided code snippet includes necessary dependencies for implementing the `load_model` function. Write a Python function `def load_model(weights_fpath: Path, device=None)` to solve the following problem: Loads the model in memory. If this function is not explicitely called, it will be run on the first call to embed_frames() with the default weights file. :param weights_fpath: the path to saved model weights. :param device: either a torch device or the name of a torch device (e.g. "cpu", "cuda"). The model will be loaded and will run on this device. Outputs will however always be on the cpu. If None, will default to your GPU if it"s available, otherwise your CPU. Here is the function: def load_model(weights_fpath: Path, device=None): """ Loads the model in memory. If this function is not explicitely called, it will be run on the first call to embed_frames() with the default weights file. :param weights_fpath: the path to saved model weights. :param device: either a torch device or the name of a torch device (e.g. "cpu", "cuda"). The model will be loaded and will run on this device. Outputs will however always be on the cpu. If None, will default to your GPU if it"s available, otherwise your CPU. """ # TODO: I think the slow loading of the encoder might have something to do with the device it # was saved on. Worth investigating. global _model, _device if device is None: _device = torch.device("cuda" if torch.cuda.is_available() else "cpu") elif isinstance(device, str): _device = torch.device(device) _model = EmotionEncoder(_device, torch.device("cpu")) checkpoint = torch.load(weights_fpath) _model.load_state_dict(checkpoint["model_state"]) _model.eval() print("Loaded encoder trained to step %d" % (checkpoint["step"]))
Loads the model in memory. If this function is not explicitely called, it will be run on the first call to embed_frames() with the default weights file. :param weights_fpath: the path to saved model weights. :param device: either a torch device or the name of a torch device (e.g. "cpu", "cuda"). The model will be loaded and will run on this device. Outputs will however always be on the cpu. If None, will default to your GPU if it"s available, otherwise your CPU.
8,362
from data_gen.tts.emotion.params_data import * from data_gen.tts.emotion.model import EmotionEncoder from data_gen.tts.emotion.audio import preprocess_wav from matplotlib import cm from data_gen.tts.emotion import audio from pathlib import Path import matplotlib.pyplot as plt import numpy as np import torch _model = None def is_loaded(): return _model is not None
null
8,363
from data_gen.tts.emotion.params_data import * from data_gen.tts.emotion.model import EmotionEncoder from data_gen.tts.emotion.audio import preprocess_wav from matplotlib import cm from data_gen.tts.emotion import audio from pathlib import Path import matplotlib.pyplot as plt import numpy as np import torch def embed_frames_batch(frames_batch): """ Computes embeddings for a batch of mel spectrogram. :param frames_batch: a batch mel of spectrogram as a numpy array of float32 of shape (batch_size, n_frames, n_channels) :return: the embeddings as a numpy array of float32 of shape (batch_size, model_embedding_size) """ if _model is None: raise Exception("Model was not loaded. Call load_model() before inference.") frames = torch.from_numpy(frames_batch).to(_device) embed = _model.inference(frames).detach().cpu().numpy() return embed def compute_partial_slices(n_samples, partial_utterance_n_frames=partials_n_frames, min_pad_coverage=0.75, overlap=0.5): """ Computes where to split an utterance waveform and its corresponding mel spectrogram to obtain partial utterances of <partial_utterance_n_frames> each. Both the waveform and the mel spectrogram slices are returned, so as to make each partial utterance waveform correspond to its spectrogram. This function assumes that the mel spectrogram parameters used are those defined in params_data.py. The returned ranges may be indexing further than the length of the waveform. It is recommended that you pad the waveform with zeros up to wave_slices[-1].stop. :param n_samples: the number of samples in the waveform :param partial_utterance_n_frames: the number of mel spectrogram frames in each partial utterance :param min_pad_coverage: when reaching the last partial utterance, it may or may not have enough frames. If at least <min_pad_coverage> of <partial_utterance_n_frames> are present, then the last partial utterance will be considered, as if we padded the audio. Otherwise, it will be discarded, as if we trimmed the audio. If there aren't enough frames for 1 partial utterance, this parameter is ignored so that the function always returns at least 1 slice. :param overlap: by how much the partial utterance should overlap. If set to 0, the partial utterances are entirely disjoint. :return: the waveform slices and mel spectrogram slices as lists of array slices. Index respectively the waveform and the mel spectrogram with these slices to obtain the partial utterances. """ assert 0 <= overlap < 1 assert 0 < min_pad_coverage <= 1 samples_per_frame = int((sampling_rate * mel_window_step / 1000)) n_frames = int(np.ceil((n_samples + 1) / samples_per_frame)) frame_step = max(int(np.round(partial_utterance_n_frames * (1 - overlap))), 1) # Compute the slices wav_slices, mel_slices = [], [] steps = max(1, n_frames - partial_utterance_n_frames + frame_step + 1) for i in range(0, steps, frame_step): mel_range = np.array([i, i + partial_utterance_n_frames]) wav_range = mel_range * samples_per_frame mel_slices.append(slice(*mel_range)) wav_slices.append(slice(*wav_range)) # Evaluate whether extra padding is warranted or not last_wav_range = wav_slices[-1] coverage = (n_samples - last_wav_range.start) / (last_wav_range.stop - last_wav_range.start) if coverage < min_pad_coverage and len(mel_slices) > 1: mel_slices = mel_slices[:-1] wav_slices = wav_slices[:-1] return wav_slices, mel_slices The provided code snippet includes necessary dependencies for implementing the `embed_utterance` function. Write a Python function `def embed_utterance(wav, using_partials=True, return_partials=False, **kwargs)` to solve the following problem: Computes an embedding for a single utterance. # TODO: handle multiple wavs to benefit from batching on GPU :param wav: a preprocessed (see audio.py) utterance waveform as a numpy array of float32 :param using_partials: if True, then the utterance is split in partial utterances of <partial_utterance_n_frames> frames and the utterance embedding is computed from their normalized average. If False, the utterance is instead computed from feeding the entire spectogram to the network. :param return_partials: if True, the partial embeddings will also be returned along with the wav slices that correspond to the partial embeddings. :param kwargs: additional arguments to compute_partial_splits() :return: the embedding as a numpy array of float32 of shape (model_embedding_size,). If <return_partials> is True, the partial utterances as a numpy array of float32 of shape (n_partials, model_embedding_size) and the wav partials as a list of slices will also be returned. If <using_partials> is simultaneously set to False, both these values will be None instead. Here is the function: def embed_utterance(wav, using_partials=True, return_partials=False, **kwargs): """ Computes an embedding for a single utterance. # TODO: handle multiple wavs to benefit from batching on GPU :param wav: a preprocessed (see audio.py) utterance waveform as a numpy array of float32 :param using_partials: if True, then the utterance is split in partial utterances of <partial_utterance_n_frames> frames and the utterance embedding is computed from their normalized average. If False, the utterance is instead computed from feeding the entire spectogram to the network. :param return_partials: if True, the partial embeddings will also be returned along with the wav slices that correspond to the partial embeddings. :param kwargs: additional arguments to compute_partial_splits() :return: the embedding as a numpy array of float32 of shape (model_embedding_size,). If <return_partials> is True, the partial utterances as a numpy array of float32 of shape (n_partials, model_embedding_size) and the wav partials as a list of slices will also be returned. If <using_partials> is simultaneously set to False, both these values will be None instead. """ # Process the entire utterance if not using partials if not using_partials: frames = audio.wav_to_mel_spectrogram(wav) embed = embed_frames_batch(frames[None, ...])[0] if return_partials: return embed, None, None return embed # Compute where to split the utterance into partials and pad if necessary wave_slices, mel_slices = compute_partial_slices(len(wav), **kwargs) max_wave_length = wave_slices[-1].stop if max_wave_length >= len(wav): wav = np.pad(wav, (0, max_wave_length - len(wav)), "constant") # Split the utterance into partials frames = audio.wav_to_mel_spectrogram(wav) frames_batch = np.array([frames[s] for s in mel_slices]) partial_embeds = embed_frames_batch(frames_batch) # Compute the utterance embedding from the partial embeddings raw_embed = np.mean(partial_embeds, axis=0) embed = raw_embed / np.linalg.norm(raw_embed, 2) if return_partials: return embed, partial_embeds, wave_slices return embed
Computes an embedding for a single utterance. # TODO: handle multiple wavs to benefit from batching on GPU :param wav: a preprocessed (see audio.py) utterance waveform as a numpy array of float32 :param using_partials: if True, then the utterance is split in partial utterances of <partial_utterance_n_frames> frames and the utterance embedding is computed from their normalized average. If False, the utterance is instead computed from feeding the entire spectogram to the network. :param return_partials: if True, the partial embeddings will also be returned along with the wav slices that correspond to the partial embeddings. :param kwargs: additional arguments to compute_partial_splits() :return: the embedding as a numpy array of float32 of shape (model_embedding_size,). If <return_partials> is True, the partial utterances as a numpy array of float32 of shape (n_partials, model_embedding_size) and the wav partials as a list of slices will also be returned. If <using_partials> is simultaneously set to False, both these values will be None instead.
8,364
from data_gen.tts.emotion.params_data import * from data_gen.tts.emotion.model import EmotionEncoder from data_gen.tts.emotion.audio import preprocess_wav from matplotlib import cm from data_gen.tts.emotion import audio from pathlib import Path import matplotlib.pyplot as plt import numpy as np import torch def embed_speaker(wavs, **kwargs): raise NotImplemented()
null
8,365
from data_gen.tts.emotion.params_data import * from data_gen.tts.emotion.model import EmotionEncoder from data_gen.tts.emotion.audio import preprocess_wav from matplotlib import cm from data_gen.tts.emotion import audio from pathlib import Path import matplotlib.pyplot as plt import numpy as np import torch def plot_embedding_as_heatmap(embed, ax=None, title="", shape=None, color_range=(0, 0.30)): if ax is None: ax = plt.gca() if shape is None: height = int(np.sqrt(len(embed))) shape = (height, -1) embed = embed.reshape(shape) cmap = cm.get_cmap() mappable = ax.imshow(embed, cmap=cmap) cbar = plt.colorbar(mappable, ax=ax, fraction=0.046, pad=0.04) cbar.set_clim(*color_range) ax.set_xticks([]), ax.set_yticks([]) ax.set_title(title)
null
8,366
from scipy.ndimage.morphology import binary_dilation from data_gen.tts.emotion.params_data import * from pathlib import Path from typing import Optional, Union import numpy as np import webrtcvad import librosa import struct def trim_long_silences(wav): """ Ensures that segments without voice in the waveform remain no longer than a threshold determined by the VAD parameters in params.py. :param wav: the raw waveform as a numpy array of floats :return: the same waveform with silences trimmed away (length <= original wav length) """ # Compute the voice detection window size samples_per_window = (vad_window_length * sampling_rate) // 1000 # Trim the end of the audio to have a multiple of the window size wav = wav[:len(wav) - (len(wav) % samples_per_window)] # Convert the float waveform to 16-bit mono PCM pcm_wave = struct.pack("%dh" % len(wav), *(np.round(wav * int16_max)).astype(np.int16)) # Perform voice activation detection voice_flags = [] vad = webrtcvad.Vad(mode=3) for window_start in range(0, len(wav), samples_per_window): window_end = window_start + samples_per_window voice_flags.append(vad.is_speech(pcm_wave[window_start * 2:window_end * 2], sample_rate=sampling_rate)) voice_flags = np.array(voice_flags) # Smooth the voice detection with a moving average def moving_average(array, width): array_padded = np.concatenate((np.zeros((width - 1) // 2), array, np.zeros(width // 2))) ret = np.cumsum(array_padded, dtype=float) ret[width:] = ret[width:] - ret[:-width] return ret[width - 1:] / width audio_mask = moving_average(voice_flags, vad_moving_average_width) audio_mask = np.round(audio_mask).astype(np.bool) # Dilate the voiced regions audio_mask = binary_dilation(audio_mask, np.ones(vad_max_silence_length + 1)) audio_mask = np.repeat(audio_mask, samples_per_window) return wav[audio_mask == True] def normalize_volume(wav, target_dBFS, increase_only=False, decrease_only=False): if increase_only and decrease_only: raise ValueError("Both increase only and decrease only are set") dBFS_change = target_dBFS - 10 * np.log10(np.mean(wav ** 2)) if (dBFS_change < 0 and increase_only) or (dBFS_change > 0 and decrease_only): return wav return wav * (10 ** (dBFS_change / 20)) The provided code snippet includes necessary dependencies for implementing the `preprocess_wav` function. Write a Python function `def preprocess_wav(fpath_or_wav: Union[str, Path, np.ndarray], source_sr: Optional[int] = None)` to solve the following problem: Applies the preprocessing operations used in training the Speaker Encoder to a waveform either on disk or in memory. The waveform will be resampled to match the data hyperparameters. :param fpath_or_wav: either a filepath to an audio file (many extensions are supported, not just .wav), either the waveform as a numpy array of floats. :param source_sr: if passing an audio waveform, the sampling rate of the waveform before preprocessing. After preprocessing, the waveform's sampling rate will match the data hyperparameters. If passing a filepath, the sampling rate will be automatically detected and this argument will be ignored. Here is the function: def preprocess_wav(fpath_or_wav: Union[str, Path, np.ndarray], source_sr: Optional[int] = None): """ Applies the preprocessing operations used in training the Speaker Encoder to a waveform either on disk or in memory. The waveform will be resampled to match the data hyperparameters. :param fpath_or_wav: either a filepath to an audio file (many extensions are supported, not just .wav), either the waveform as a numpy array of floats. :param source_sr: if passing an audio waveform, the sampling rate of the waveform before preprocessing. After preprocessing, the waveform's sampling rate will match the data hyperparameters. If passing a filepath, the sampling rate will be automatically detected and this argument will be ignored. """ # Load the wav from disk if needed if isinstance(fpath_or_wav, str) or isinstance(fpath_or_wav, Path): wav, source_sr = librosa.load(str(fpath_or_wav), sr=None) else: wav = fpath_or_wav # Resample the wav if needed if source_sr is not None and source_sr != sampling_rate: wav = librosa.resample(wav, source_sr, sampling_rate) # Apply the preprocessing: normalize volume and shorten long silences wav = normalize_volume(wav, audio_norm_target_dBFS, increase_only=True) wav = trim_long_silences(wav) return wav
Applies the preprocessing operations used in training the Speaker Encoder to a waveform either on disk or in memory. The waveform will be resampled to match the data hyperparameters. :param fpath_or_wav: either a filepath to an audio file (many extensions are supported, not just .wav), either the waveform as a numpy array of floats. :param source_sr: if passing an audio waveform, the sampling rate of the waveform before preprocessing. After preprocessing, the waveform's sampling rate will match the data hyperparameters. If passing a filepath, the sampling rate will be automatically detected and this argument will be ignored.
8,367
from data_gen.tts.data_gen_utils import is_sil_phoneme REGISTERED_TEXT_PROCESSORS = {} def register_txt_processors(name): def _f(cls): REGISTERED_TEXT_PROCESSORS[name] = cls return cls return _f
null
8,368
from data_gen.tts.data_gen_utils import is_sil_phoneme REGISTERED_TEXT_PROCESSORS = {} def get_txt_processor_cls(name): return REGISTERED_TEXT_PROCESSORS.get(name, None)
null
8,369
import matplotlib from torch.nn import DataParallel from torch.nn.parallel import DistributedDataParallel import glob import itertools import subprocess import threading import traceback from pytorch_lightning.callbacks import GradientAccumulationScheduler from pytorch_lightning.callbacks import ModelCheckpoint from functools import wraps from torch.cuda._utils import _get_device_index import numpy as np import torch.optim import torch.utils.data import copy import logging import os import re import sys import torch import torch.distributed as dist import torch.multiprocessing as mp import tqdm from torch.optim.optimizer import Optimizer The provided code snippet includes necessary dependencies for implementing the `data_loader` function. Write a Python function `def data_loader(fn)` to solve the following problem: Decorator to make any fx with this use the lazy property :param fn: :return: Here is the function: def data_loader(fn): """ Decorator to make any fx with this use the lazy property :param fn: :return: """ wraps(fn) attr_name = '_lazy_' + fn.__name__ def _get_data_loader(self): try: value = getattr(self, attr_name) except AttributeError: try: value = fn(self) # Lazy evaluation, done only once. if ( value is not None and not isinstance(value, list) and fn.__name__ in ['test_dataloader', 'val_dataloader'] ): value = [value] except AttributeError as e: # Guard against AttributeError suppression. (Issue #142) traceback.print_exc() error = f'{fn.__name__}: An AttributeError was encountered: ' + str(e) raise RuntimeError(error) from e setattr(self, attr_name, value) # Memoize evaluation. return value return _get_data_loader
Decorator to make any fx with this use the lazy property :param fn: :return:
8,370
import matplotlib from torch.nn import DataParallel from torch.nn.parallel import DistributedDataParallel import glob import itertools import subprocess import threading import traceback from pytorch_lightning.callbacks import GradientAccumulationScheduler from pytorch_lightning.callbacks import ModelCheckpoint from functools import wraps from torch.cuda._utils import _get_device_index import numpy as np import torch.optim import torch.utils.data import copy import logging import os import re import sys import torch import torch.distributed as dist import torch.multiprocessing as mp import tqdm from torch.optim.optimizer import Optimizer def get_a_var(obj): # pragma: no cover if isinstance(obj, torch.Tensor): return obj if isinstance(obj, list) or isinstance(obj, tuple): for result in map(get_a_var, obj): if isinstance(result, torch.Tensor): return result if isinstance(obj, dict): for result in map(get_a_var, obj.items()): if isinstance(result, torch.Tensor): return result return None The provided code snippet includes necessary dependencies for implementing the `parallel_apply` function. Write a Python function `def parallel_apply(modules, inputs, kwargs_tup=None, devices=None)` to solve the following problem: r"""Applies each `module` in :attr:`modules` in parallel on arguments contained in :attr:`inputs` (positional) and :attr:`kwargs_tup` (keyword) on each of :attr:`devices`. Args: modules (Module): modules to be parallelized inputs (tensor): inputs to the modules devices (list of int or torch.device): CUDA devices :attr:`modules`, :attr:`inputs`, :attr:`kwargs_tup` (if given), and :attr:`devices` (if given) should all have same length. Moreover, each element of :attr:`inputs` can either be a single object as the only argument to a module, or a collection of positional arguments. Here is the function: def parallel_apply(modules, inputs, kwargs_tup=None, devices=None): # pragma: no cover r"""Applies each `module` in :attr:`modules` in parallel on arguments contained in :attr:`inputs` (positional) and :attr:`kwargs_tup` (keyword) on each of :attr:`devices`. Args: modules (Module): modules to be parallelized inputs (tensor): inputs to the modules devices (list of int or torch.device): CUDA devices :attr:`modules`, :attr:`inputs`, :attr:`kwargs_tup` (if given), and :attr:`devices` (if given) should all have same length. Moreover, each element of :attr:`inputs` can either be a single object as the only argument to a module, or a collection of positional arguments. """ assert len(modules) == len(inputs) if kwargs_tup is not None: assert len(modules) == len(kwargs_tup) else: kwargs_tup = ({},) * len(modules) if devices is not None: assert len(modules) == len(devices) else: devices = [None] * len(modules) devices = list(map(lambda x: _get_device_index(x, True), devices)) lock = threading.Lock() results = {} grad_enabled = torch.is_grad_enabled() def _worker(i, module, input, kwargs, device=None): torch.set_grad_enabled(grad_enabled) if device is None: device = get_a_var(input).get_device() try: with torch.cuda.device(device): # this also avoids accidental slicing of `input` if it is a Tensor if not isinstance(input, (list, tuple)): input = (input,) # --------------- # CHANGE if module.training: output = module.training_step(*input, **kwargs) elif module.testing: output = module.test_step(*input, **kwargs) else: output = module.validation_step(*input, **kwargs) # --------------- with lock: results[i] = output except Exception as e: with lock: results[i] = e # make sure each module knows what training state it's in... # fixes weird bug where copies are out of sync root_m = modules[0] for m in modules[1:]: m.training = root_m.training m.testing = root_m.testing if len(modules) > 1: threads = [threading.Thread(target=_worker, args=(i, module, input, kwargs, device)) for i, (module, input, kwargs, device) in enumerate(zip(modules, inputs, kwargs_tup, devices))] for thread in threads: thread.start() for thread in threads: thread.join() else: _worker(0, modules[0], inputs[0], kwargs_tup[0], devices[0]) outputs = [] for i in range(len(inputs)): output = results[i] if isinstance(output, Exception): raise output outputs.append(output) return outputs
r"""Applies each `module` in :attr:`modules` in parallel on arguments contained in :attr:`inputs` (positional) and :attr:`kwargs_tup` (keyword) on each of :attr:`devices`. Args: modules (Module): modules to be parallelized inputs (tensor): inputs to the modules devices (list of int or torch.device): CUDA devices :attr:`modules`, :attr:`inputs`, :attr:`kwargs_tup` (if given), and :attr:`devices` (if given) should all have same length. Moreover, each element of :attr:`inputs` can either be a single object as the only argument to a module, or a collection of positional arguments.
8,371
import matplotlib from torch.nn import DataParallel from torch.nn.parallel import DistributedDataParallel import glob import itertools import subprocess import threading import traceback from pytorch_lightning.callbacks import GradientAccumulationScheduler from pytorch_lightning.callbacks import ModelCheckpoint from functools import wraps from torch.cuda._utils import _get_device_index import numpy as np import torch.optim import torch.utils.data import copy import logging import os import re import sys import torch import torch.distributed as dist import torch.multiprocessing as mp import tqdm from torch.optim.optimizer import Optimizer The provided code snippet includes necessary dependencies for implementing the `_find_tensors` function. Write a Python function `def _find_tensors(obj)` to solve the following problem: r""" Recursively find all tensors contained in the specified object. Here is the function: def _find_tensors(obj): # pragma: no cover r""" Recursively find all tensors contained in the specified object. """ if isinstance(obj, torch.Tensor): return [obj] if isinstance(obj, (list, tuple)): return itertools.chain(*map(_find_tensors, obj)) if isinstance(obj, dict): return itertools.chain(*map(_find_tensors, obj.values())) return []
r""" Recursively find all tensors contained in the specified object.
8,372
import sys, os, argparse, codecs, string, re CHINESE_DIGIS = u'零一二三四五六七八九' BIG_CHINESE_DIGIS_SIMPLIFIED = u'零壹贰叁肆伍陆柒捌玖' BIG_CHINESE_DIGIS_TRADITIONAL = u'零壹貳參肆伍陸柒捌玖' SMALLER_BIG_CHINESE_UNITS_SIMPLIFIED = u'十百千万' SMALLER_BIG_CHINESE_UNITS_TRADITIONAL = u'拾佰仟萬' LARGER_CHINESE_NUMERING_UNITS_SIMPLIFIED = u'亿兆京垓秭穰沟涧正载' LARGER_CHINESE_NUMERING_UNITS_TRADITIONAL = u'億兆京垓秭穰溝澗正載' SMALLER_CHINESE_NUMERING_UNITS_SIMPLIFIED = u'十百千万' SMALLER_CHINESE_NUMERING_UNITS_TRADITIONAL = u'拾佰仟萬' ZERO_ALT = u'〇' ONE_ALT = u'幺' TWO_ALTS = [u'两', u'兩'] POSITIVE = [u'正', u'正'] NEGATIVE = [u'负', u'負'] POINT = [u'点', u'點'] = ['low', 'mid', 'high'] CURRENCY_NAMES = '(人民币|美元|日元|英镑|欧元|马克|法郎|加拿大元|澳元|港币|先令|芬兰马克|爱尔兰镑|' \ '里拉|荷兰盾|埃斯库多|比塞塔|印尼盾|林吉特|新西兰元|比索|卢布|新加坡元|韩元|泰铢)' CURRENCY_UNITS = '((亿|千万|百万|万|千|百)|(亿|千万|百万|万|千|百|)元|(亿|千万|百万|万|千|百|)块|角|毛|分)' COM_QUANTIFIERS = '(匹|张|座|回|场|尾|条|个|首|阙|阵|网|炮|顶|丘|棵|只|支|袭|辆|挑|担|颗|壳|窠|曲|墙|群|腔|' \ '砣|座|客|贯|扎|捆|刀|令|打|手|罗|坡|山|岭|江|溪|钟|队|单|双|对|出|口|头|脚|板|跳|枝|件|贴|' \ '针|线|管|名|位|身|堂|课|本|页|家|户|层|丝|毫|厘|分|钱|两|斤|担|铢|石|钧|锱|忽|(千|毫|微)克|' \ '毫|厘|分|寸|尺|丈|里|寻|常|铺|程|(千|分|厘|毫|微)米|撮|勺|合|升|斗|石|盘|碗|碟|叠|桶|笼|盆|' \ '盒|杯|钟|斛|锅|簋|篮|盘|桶|罐|瓶|壶|卮|盏|箩|箱|煲|啖|袋|钵|年|月|日|季|刻|时|周|天|秒|分|旬|' \ '纪|岁|世|更|夜|春|夏|秋|冬|代|伏|辈|丸|泡|粒|颗|幢|堆|条|根|支|道|面|片|张|颗|块)' CHINESE_PUNC_STOP = '!?。。' CHINESE_PUNC_NON_STOP = '"#$%&'()*+,-/:;<=>@[\]^_`{|}~⦅⦆「」、、〃《》「」『』【】〔〕〖〗〘〙〚〛〜〝〞〟〰〾〿–—‘’‛“”„‟…‧﹏' CHINESE_PUNC_LIST = CHINESE_PUNC_STOP + CHINESE_PUNC_NON_STOP CC, CNU, CND, CM = ChineseChar, ChineseNumberUnit, ChineseNumberDigit, ChineseMath def create_system(numbering_type=NUMBERING_TYPES[1]): """ 根据数字系统类型返回创建相应的数字系统,默认为 mid NUMBERING_TYPES = ['low', 'mid', 'high']: 中文数字系统类型 low: '兆' = '亿' * '十' = $10^{9}$, '京' = '兆' * '十', etc. mid: '兆' = '亿' * '万' = $10^{12}$, '京' = '兆' * '万', etc. high: '兆' = '亿' * '亿' = $10^{16}$, '京' = '兆' * '兆', etc. 返回对应的数字系统 """ # chinese number units of '亿' and larger all_larger_units = zip( LARGER_CHINESE_NUMERING_UNITS_SIMPLIFIED, LARGER_CHINESE_NUMERING_UNITS_TRADITIONAL) larger_units = [CNU.create(i, v, numbering_type, False) for i, v in enumerate(all_larger_units)] # chinese number units of '十, 百, 千, 万' all_smaller_units = zip( SMALLER_CHINESE_NUMERING_UNITS_SIMPLIFIED, SMALLER_CHINESE_NUMERING_UNITS_TRADITIONAL) smaller_units = [CNU.create(i, v, small_unit=True) for i, v in enumerate(all_smaller_units)] # digis chinese_digis = zip(CHINESE_DIGIS, CHINESE_DIGIS, BIG_CHINESE_DIGIS_SIMPLIFIED, BIG_CHINESE_DIGIS_TRADITIONAL) digits = [CND.create(i, v) for i, v in enumerate(chinese_digis)] digits[0].alt_s, digits[0].alt_t = ZERO_ALT, ZERO_ALT digits[1].alt_s, digits[1].alt_t = ONE_ALT, ONE_ALT digits[2].alt_s, digits[2].alt_t = TWO_ALTS[0], TWO_ALTS[1] # symbols positive_cn = CM(POSITIVE[0], POSITIVE[1], '+', lambda x: x) negative_cn = CM(NEGATIVE[0], NEGATIVE[1], '-', lambda x: -x) point_cn = CM(POINT[0], POINT[1], '.', lambda x, y: float(str(x) + '.' + str(y))) # sil_cn = CM(SIL[0], SIL[1], '-', lambda x, y: float(str(x) + '-' + str(y))) system = NumberSystem() system.units = smaller_units + larger_units system.digits = digits system.math = MathSymbol(positive_cn, negative_cn, point_cn) # system.symbols = OtherSymbol(sil_cn) return system me__ == '__main__': # nsw_test() p = argparse.ArgumentParser() p.add_argument('ifile', help='input filename, assume utf-8 encoding') p.add_argument('ofile', help='output filename') p.add_argument('--to_upper', action='store_true', help='convert to upper case') p.add_argument('--to_lower', action='store_true', help='convert to lower case') p.add_argument('--has_key', action='store_true', help="input text has Kaldi's key as first field.") p.add_argument('--log_interval', type=int, default=10000, help='log interval in number of processed lines') args = p.parse_args() ifile = codecs.open(args.ifile, 'r', 'utf8') ofile = codecs.open(args.ofile, 'w+', 'utf8') n = 0 for l in ifile: key = '' text = '' if args.has_key: cols = l.split(maxsplit=1) key = cols[0] if len(cols) == 2: text = cols[1] else: text = '' else: text = l # cases if args.to_upper and args.to_lower: sys.stderr.write('text norm: to_upper OR to_lower?') exit(1) if args.to_upper: text = text.upper() if args.to_lower: text = text.lower() # NSW(Non-Standard-Word) normalization text = NSWNormalizer(text).normalize() # if args.has_key: ofile.write(key + '\t' + text) else: ofile.write(text) n += 1 if n % args.log_interval == 0: sys.stderr.write("text norm: {} lines done.\n".format(n)) sys.stderr.write("text norm: {} lines done in total.\n".format(n)) ifile.close() ofile.close() def chn2num(chinese_string, numbering_type=NUMBERING_TYPES[1]): def get_symbol(char, system): for u in system.units: if char in [u.traditional, u.simplified, u.big_s, u.big_t]: return u for d in system.digits: if char in [d.traditional, d.simplified, d.big_s, d.big_t, d.alt_s, d.alt_t]: return d for m in system.math: if char in [m.traditional, m.simplified]: return m def string2symbols(chinese_string, system): int_string, dec_string = chinese_string, '' for p in [system.math.point.simplified, system.math.point.traditional]: if p in chinese_string: int_string, dec_string = chinese_string.split(p) break return [get_symbol(c, system) for c in int_string], \ [get_symbol(c, system) for c in dec_string] def correct_symbols(integer_symbols, system): """ 一百八 to 一百八十 一亿一千三百万 to 一亿 一千万 三百万 """ if integer_symbols and isinstance(integer_symbols[0], CNU): if integer_symbols[0].power == 1: integer_symbols = [system.digits[1]] + integer_symbols if len(integer_symbols) > 1: if isinstance(integer_symbols[-1], CND) and isinstance(integer_symbols[-2], CNU): integer_symbols.append( CNU(integer_symbols[-2].power - 1, None, None, None, None)) result = [] unit_count = 0 for s in integer_symbols: if isinstance(s, CND): result.append(s) unit_count = 0 elif isinstance(s, CNU): current_unit = CNU(s.power, None, None, None, None) unit_count += 1 if unit_count == 1: result.append(current_unit) elif unit_count > 1: for i in range(len(result)): if isinstance(result[-i - 1], CNU) and result[-i - 1].power < current_unit.power: result[-i - 1] = CNU(result[-i - 1].power + current_unit.power, None, None, None, None) return result def compute_value(integer_symbols): """ Compute the value. When current unit is larger than previous unit, current unit * all previous units will be used as all previous units. e.g. '两千万' = 2000 * 10000 not 2000 + 10000 """ value = [0] last_power = 0 for s in integer_symbols: if isinstance(s, CND): value[-1] = s.value elif isinstance(s, CNU): value[-1] *= pow(10, s.power) if s.power > last_power: value[:-1] = list(map(lambda v: v * pow(10, s.power), value[:-1])) last_power = s.power value.append(0) return sum(value) system = create_system(numbering_type) int_part, dec_part = string2symbols(chinese_string, system) int_part = correct_symbols(int_part, system) int_str = str(compute_value(int_part)) dec_str = ''.join([str(d.value) for d in dec_part]) if dec_part: return '{0}.{1}'.format(int_str, dec_str) else: return int_str
null
8,373
import sys, os, argparse, codecs, string, re CHINESE_DIGIS = u'零一二三四五六七八九' BIG_CHINESE_DIGIS_SIMPLIFIED = u'零壹贰叁肆伍陆柒捌玖' BIG_CHINESE_DIGIS_TRADITIONAL = u'零壹貳參肆伍陸柒捌玖' SMALLER_BIG_CHINESE_UNITS_SIMPLIFIED = u'十百千万' SMALLER_BIG_CHINESE_UNITS_TRADITIONAL = u'拾佰仟萬' LARGER_CHINESE_NUMERING_UNITS_SIMPLIFIED = u'亿兆京垓秭穰沟涧正载' LARGER_CHINESE_NUMERING_UNITS_TRADITIONAL = u'億兆京垓秭穰溝澗正載' SMALLER_CHINESE_NUMERING_UNITS_SIMPLIFIED = u'十百千万' SMALLER_CHINESE_NUMERING_UNITS_TRADITIONAL = u'拾佰仟萬' ZERO_ALT = u'〇' ONE_ALT = u'幺' TWO_ALTS = [u'两', u'兩'] POSITIVE = [u'正', u'正'] NEGATIVE = [u'负', u'負'] POINT = [u'点', u'點'] = ['low', 'mid', 'high'] CURRENCY_NAMES = '(人民币|美元|日元|英镑|欧元|马克|法郎|加拿大元|澳元|港币|先令|芬兰马克|爱尔兰镑|' \ '里拉|荷兰盾|埃斯库多|比塞塔|印尼盾|林吉特|新西兰元|比索|卢布|新加坡元|韩元|泰铢)' CURRENCY_UNITS = '((亿|千万|百万|万|千|百)|(亿|千万|百万|万|千|百|)元|(亿|千万|百万|万|千|百|)块|角|毛|分)' COM_QUANTIFIERS = '(匹|张|座|回|场|尾|条|个|首|阙|阵|网|炮|顶|丘|棵|只|支|袭|辆|挑|担|颗|壳|窠|曲|墙|群|腔|' \ '砣|座|客|贯|扎|捆|刀|令|打|手|罗|坡|山|岭|江|溪|钟|队|单|双|对|出|口|头|脚|板|跳|枝|件|贴|' \ '针|线|管|名|位|身|堂|课|本|页|家|户|层|丝|毫|厘|分|钱|两|斤|担|铢|石|钧|锱|忽|(千|毫|微)克|' \ '毫|厘|分|寸|尺|丈|里|寻|常|铺|程|(千|分|厘|毫|微)米|撮|勺|合|升|斗|石|盘|碗|碟|叠|桶|笼|盆|' \ '盒|杯|钟|斛|锅|簋|篮|盘|桶|罐|瓶|壶|卮|盏|箩|箱|煲|啖|袋|钵|年|月|日|季|刻|时|周|天|秒|分|旬|' \ '纪|岁|世|更|夜|春|夏|秋|冬|代|伏|辈|丸|泡|粒|颗|幢|堆|条|根|支|道|面|片|张|颗|块)' CHINESE_PUNC_STOP = '!?。。' CHINESE_PUNC_NON_STOP = '"#$%&'()*+,-/:;<=>@[\]^_`{|}~⦅⦆「」、、〃《》「」『』【】〔〕〖〗〘〙〚〛〜〝〞〟〰〾〿–—‘’‛“”„‟…‧﹏' CHINESE_PUNC_LIST = CHINESE_PUNC_STOP + CHINESE_PUNC_NON_STOP CC, CNU, CND, CM = ChineseChar, ChineseNumberUnit, ChineseNumberDigit, ChineseMath def create_system(numbering_type=NUMBERING_TYPES[1]): """ 根据数字系统类型返回创建相应的数字系统,默认为 mid NUMBERING_TYPES = ['low', 'mid', 'high']: 中文数字系统类型 low: '兆' = '亿' * '十' = $10^{9}$, '京' = '兆' * '十', etc. mid: '兆' = '亿' * '万' = $10^{12}$, '京' = '兆' * '万', etc. high: '兆' = '亿' * '亿' = $10^{16}$, '京' = '兆' * '兆', etc. 返回对应的数字系统 """ # chinese number units of '亿' and larger all_larger_units = zip( LARGER_CHINESE_NUMERING_UNITS_SIMPLIFIED, LARGER_CHINESE_NUMERING_UNITS_TRADITIONAL) larger_units = [CNU.create(i, v, numbering_type, False) for i, v in enumerate(all_larger_units)] # chinese number units of '十, 百, 千, 万' all_smaller_units = zip( SMALLER_CHINESE_NUMERING_UNITS_SIMPLIFIED, SMALLER_CHINESE_NUMERING_UNITS_TRADITIONAL) smaller_units = [CNU.create(i, v, small_unit=True) for i, v in enumerate(all_smaller_units)] # digis chinese_digis = zip(CHINESE_DIGIS, CHINESE_DIGIS, BIG_CHINESE_DIGIS_SIMPLIFIED, BIG_CHINESE_DIGIS_TRADITIONAL) digits = [CND.create(i, v) for i, v in enumerate(chinese_digis)] digits[0].alt_s, digits[0].alt_t = ZERO_ALT, ZERO_ALT digits[1].alt_s, digits[1].alt_t = ONE_ALT, ONE_ALT digits[2].alt_s, digits[2].alt_t = TWO_ALTS[0], TWO_ALTS[1] # symbols positive_cn = CM(POSITIVE[0], POSITIVE[1], '+', lambda x: x) negative_cn = CM(NEGATIVE[0], NEGATIVE[1], '-', lambda x: -x) point_cn = CM(POINT[0], POINT[1], '.', lambda x, y: float(str(x) + '.' + str(y))) # sil_cn = CM(SIL[0], SIL[1], '-', lambda x, y: float(str(x) + '-' + str(y))) system = NumberSystem() system.units = smaller_units + larger_units system.digits = digits system.math = MathSymbol(positive_cn, negative_cn, point_cn) # system.symbols = OtherSymbol(sil_cn) return system me__ == '__main__': # nsw_test() p = argparse.ArgumentParser() p.add_argument('ifile', help='input filename, assume utf-8 encoding') p.add_argument('ofile', help='output filename') p.add_argument('--to_upper', action='store_true', help='convert to upper case') p.add_argument('--to_lower', action='store_true', help='convert to lower case') p.add_argument('--has_key', action='store_true', help="input text has Kaldi's key as first field.") p.add_argument('--log_interval', type=int, default=10000, help='log interval in number of processed lines') args = p.parse_args() ifile = codecs.open(args.ifile, 'r', 'utf8') ofile = codecs.open(args.ofile, 'w+', 'utf8') n = 0 for l in ifile: key = '' text = '' if args.has_key: cols = l.split(maxsplit=1) key = cols[0] if len(cols) == 2: text = cols[1] else: text = '' else: text = l # cases if args.to_upper and args.to_lower: sys.stderr.write('text norm: to_upper OR to_lower?') exit(1) if args.to_upper: text = text.upper() if args.to_lower: text = text.lower() # NSW(Non-Standard-Word) normalization text = NSWNormalizer(text).normalize() # if args.has_key: ofile.write(key + '\t' + text) else: ofile.write(text) n += 1 if n % args.log_interval == 0: sys.stderr.write("text norm: {} lines done.\n".format(n)) sys.stderr.write("text norm: {} lines done in total.\n".format(n)) ifile.close() ofile.close() def num2chn(number_string, numbering_type=NUMBERING_TYPES[1], big=False, traditional=False, alt_zero=False, alt_one=False, alt_two=True, use_zeros=True, use_units=True): def get_value(value_string, use_zeros=True): striped_string = value_string.lstrip('0') # record nothing if all zeros if not striped_string: return [] # record one digits elif len(striped_string) == 1: if use_zeros and len(value_string) != len(striped_string): return [system.digits[0], system.digits[int(striped_string)]] else: return [system.digits[int(striped_string)]] # recursively record multiple digits else: result_unit = next(u for u in reversed( system.units) if u.power < len(striped_string)) result_string = value_string[:-result_unit.power] return get_value(result_string) + [result_unit] + get_value(striped_string[-result_unit.power:]) system = create_system(numbering_type) int_dec = number_string.split('.') if len(int_dec) == 1: int_string = int_dec[0] dec_string = "" elif len(int_dec) == 2: int_string = int_dec[0] dec_string = int_dec[1] else: raise ValueError( "invalid input num string with more than one dot: {}".format(number_string)) if use_units and len(int_string) > 1: result_symbols = get_value(int_string) else: result_symbols = [system.digits[int(c)] for c in int_string] dec_symbols = [system.digits[int(c)] for c in dec_string] if dec_string: result_symbols += [system.math.point] + dec_symbols if alt_two: liang = CND(2, system.digits[2].alt_s, system.digits[2].alt_t, system.digits[2].big_s, system.digits[2].big_t) for i, v in enumerate(result_symbols): if isinstance(v, CND) and v.value == 2: next_symbol = result_symbols[i + 1] if i < len(result_symbols) - 1 else None previous_symbol = result_symbols[i - 1] if i > 0 else None if isinstance(next_symbol, CNU) and isinstance(previous_symbol, (CNU, type(None))): if next_symbol.power != 1 and ((previous_symbol is None) or (previous_symbol.power != 1)): result_symbols[i] = liang # if big is True, '两' will not be used and `alt_two` has no impact on output if big: attr_name = 'big_' if traditional: attr_name += 't' else: attr_name += 's' else: if traditional: attr_name = 'traditional' else: attr_name = 'simplified' result = ''.join([getattr(s, attr_name) for s in result_symbols]) # if not use_zeros: # result = result.strip(getattr(system.digits[0], attr_name)) if alt_zero: result = result.replace( getattr(system.digits[0], attr_name), system.digits[0].alt_s) if alt_one: result = result.replace( getattr(system.digits[1], attr_name), system.digits[1].alt_s) for i, p in enumerate(POINT): if result.startswith(p): return CHINESE_DIGIS[0] + result # ^10, 11, .., 19 if len(result) >= 2 and result[1] in [SMALLER_CHINESE_NUMERING_UNITS_SIMPLIFIED[0], SMALLER_CHINESE_NUMERING_UNITS_TRADITIONAL[0]] and \ result[0] in [CHINESE_DIGIS[1], BIG_CHINESE_DIGIS_SIMPLIFIED[1], BIG_CHINESE_DIGIS_TRADITIONAL[1]]: result = result[1:] return result
null
8,374
import sys, os, argparse, codecs, string, re CHINESE_DIGIS = u'零一二三四五六七八九' BIG_CHINESE_DIGIS_SIMPLIFIED = u'零壹贰叁肆伍陆柒捌玖' BIG_CHINESE_DIGIS_TRADITIONAL = u'零壹貳參肆伍陸柒捌玖' SMALLER_BIG_CHINESE_UNITS_SIMPLIFIED = u'十百千万' SMALLER_BIG_CHINESE_UNITS_TRADITIONAL = u'拾佰仟萬' LARGER_CHINESE_NUMERING_UNITS_SIMPLIFIED = u'亿兆京垓秭穰沟涧正载' LARGER_CHINESE_NUMERING_UNITS_TRADITIONAL = u'億兆京垓秭穰溝澗正載' SMALLER_CHINESE_NUMERING_UNITS_SIMPLIFIED = u'十百千万' SMALLER_CHINESE_NUMERING_UNITS_TRADITIONAL = u'拾佰仟萬' ZERO_ALT = u'〇' ONE_ALT = u'幺' TWO_ALTS = [u'两', u'兩'] POSITIVE = [u'正', u'正'] NEGATIVE = [u'负', u'負'] POINT = [u'点', u'點'] = ['low', 'mid', 'high'] CURRENCY_NAMES = '(人民币|美元|日元|英镑|欧元|马克|法郎|加拿大元|澳元|港币|先令|芬兰马克|爱尔兰镑|' \ '里拉|荷兰盾|埃斯库多|比塞塔|印尼盾|林吉特|新西兰元|比索|卢布|新加坡元|韩元|泰铢)' CURRENCY_UNITS = '((亿|千万|百万|万|千|百)|(亿|千万|百万|万|千|百|)元|(亿|千万|百万|万|千|百|)块|角|毛|分)' COM_QUANTIFIERS = '(匹|张|座|回|场|尾|条|个|首|阙|阵|网|炮|顶|丘|棵|只|支|袭|辆|挑|担|颗|壳|窠|曲|墙|群|腔|' \ '砣|座|客|贯|扎|捆|刀|令|打|手|罗|坡|山|岭|江|溪|钟|队|单|双|对|出|口|头|脚|板|跳|枝|件|贴|' \ '针|线|管|名|位|身|堂|课|本|页|家|户|层|丝|毫|厘|分|钱|两|斤|担|铢|石|钧|锱|忽|(千|毫|微)克|' \ '毫|厘|分|寸|尺|丈|里|寻|常|铺|程|(千|分|厘|毫|微)米|撮|勺|合|升|斗|石|盘|碗|碟|叠|桶|笼|盆|' \ '盒|杯|钟|斛|锅|簋|篮|盘|桶|罐|瓶|壶|卮|盏|箩|箱|煲|啖|袋|钵|年|月|日|季|刻|时|周|天|秒|分|旬|' \ '纪|岁|世|更|夜|春|夏|秋|冬|代|伏|辈|丸|泡|粒|颗|幢|堆|条|根|支|道|面|片|张|颗|块)' CHINESE_PUNC_STOP = '!?。。' CHINESE_PUNC_NON_STOP = '"#$%&'()*+,-/:;<=>@[\]^_`{|}~⦅⦆「」、、〃《》「」『』【】〔〕〖〗〘〙〚〛〜〝〞〟〰〾〿–—‘’‛“”„‟…‧﹏' CHINESE_PUNC_LIST = CHINESE_PUNC_STOP + CHINESE_PUNC_NON_STOP CC, CNU, CND, CM = ChineseChar, ChineseNumberUnit, ChineseNumberDigit, ChineseMath def nsw_test_case(raw_text): print('I:' + raw_text) print('O:' + NSWNormalizer(raw_text).normalize()) print('') me__ == '__main__': # nsw_test() p = argparse.ArgumentParser() p.add_argument('ifile', help='input filename, assume utf-8 encoding') p.add_argument('ofile', help='output filename') p.add_argument('--to_upper', action='store_true', help='convert to upper case') p.add_argument('--to_lower', action='store_true', help='convert to lower case') p.add_argument('--has_key', action='store_true', help="input text has Kaldi's key as first field.") p.add_argument('--log_interval', type=int, default=10000, help='log interval in number of processed lines') args = p.parse_args() ifile = codecs.open(args.ifile, 'r', 'utf8') ofile = codecs.open(args.ofile, 'w+', 'utf8') n = 0 for l in ifile: key = '' text = '' if args.has_key: cols = l.split(maxsplit=1) key = cols[0] if len(cols) == 2: text = cols[1] else: text = '' else: text = l # cases if args.to_upper and args.to_lower: sys.stderr.write('text norm: to_upper OR to_lower?') exit(1) if args.to_upper: text = text.upper() if args.to_lower: text = text.lower() # NSW(Non-Standard-Word) normalization text = NSWNormalizer(text).normalize() # if args.has_key: ofile.write(key + '\t' + text) else: ofile.write(text) n += 1 if n % args.log_interval == 0: sys.stderr.write("text norm: {} lines done.\n".format(n)) sys.stderr.write("text norm: {} lines done in total.\n".format(n)) ifile.close() ofile.close() def nsw_test(): nsw_test_case('固话:0595-23865596或23880880。') nsw_test_case('固话:0595-23865596或23880880。') nsw_test_case('手机:+86 19859213959或15659451527。') nsw_test_case('分数:32477/76391。') nsw_test_case('百分数:80.03%。') nsw_test_case('编号:31520181154418。') nsw_test_case('纯数:2983.07克或12345.60米。') nsw_test_case('日期:1999年2月20日或09年3月15号。') nsw_test_case('金钱:12块5,34.5元,20.1万') nsw_test_case('特殊:O2O或B2C。') nsw_test_case('3456万吨') nsw_test_case('2938个') nsw_test_case('938') nsw_test_case('今天吃了115个小笼包231个馒头') nsw_test_case('有62%的概率')
null
8,375
import glob import logging import os import re import torch def get_last_checkpoint(work_dir, steps=None): checkpoint = None last_ckpt_path = None ckpt_paths = get_all_ckpts(work_dir, steps) if len(ckpt_paths) > 0: last_ckpt_path = ckpt_paths[0] checkpoint = torch.load(last_ckpt_path, map_location='cpu') logging.info(f'load module from checkpoint: {last_ckpt_path}') return checkpoint, last_ckpt_path def load_ckpt(cur_model, ckpt_base_dir, model_name='model', force=True, strict=True): if os.path.isfile(ckpt_base_dir): base_dir = os.path.dirname(ckpt_base_dir) ckpt_path = ckpt_base_dir checkpoint = torch.load(ckpt_base_dir, map_location='cpu') else: base_dir = ckpt_base_dir checkpoint, ckpt_path = get_last_checkpoint(ckpt_base_dir) if checkpoint is not None: state_dict = checkpoint["state_dict"] if len([k for k in state_dict.keys() if '.' in k]) > 0: state_dict = {k[len(model_name) + 1:]: v for k, v in state_dict.items() if k.startswith(f'{model_name}.')} else: if '.' not in model_name: state_dict = state_dict[model_name] else: base_model_name = model_name.split('.')[0] rest_model_name = model_name[len(base_model_name) + 1:] state_dict = { k[len(rest_model_name) + 1:]: v for k, v in state_dict[base_model_name].items() if k.startswith(f'{rest_model_name}.')} if not strict: cur_model_state_dict = cur_model.state_dict() unmatched_keys = [] for key, param in state_dict.items(): if key in cur_model_state_dict: new_param = cur_model_state_dict[key] if new_param.shape != param.shape: unmatched_keys.append(key) print("| Unmatched keys: ", key, new_param.shape, param.shape) for key in unmatched_keys: del state_dict[key] cur_model.load_state_dict(state_dict, strict=strict) print(f"| load '{model_name}' from '{ckpt_path}'.") else: e_msg = f"| ckpt not found in {base_dir}." if force: assert False, e_msg else: print(e_msg)
null
8,376
from numpy import array, zeros, full, argmin, inf, ndim from scipy.spatial.distance import cdist from math import isinf def _traceback(D): i, j = array(D.shape) - 2 p, q = [i], [j] while (i > 0) or (j > 0): tb = argmin((D[i, j], D[i, j + 1], D[i + 1, j])) if tb == 0: i -= 1 j -= 1 elif tb == 1: i -= 1 else: # (tb == 2): j -= 1 p.insert(0, i) q.insert(0, j) return array(p), array(q) The provided code snippet includes necessary dependencies for implementing the `dtw` function. Write a Python function `def dtw(x, y, dist, warp=1, w=inf, s=1.0)` to solve the following problem: Computes Dynamic Time Warping (DTW) of two sequences. :param array x: N1*M array :param array y: N2*M array :param func dist: distance used as cost measure :param int warp: how many shifts are computed. :param int w: window size limiting the maximal distance between indices of matched entries |i,j|. :param float s: weight applied on off-diagonal moves of the path. As s gets larger, the warping path is increasingly biased towards the diagonal Returns the minimum distance, the cost matrix, the accumulated cost matrix, and the wrap path. Here is the function: def dtw(x, y, dist, warp=1, w=inf, s=1.0): """ Computes Dynamic Time Warping (DTW) of two sequences. :param array x: N1*M array :param array y: N2*M array :param func dist: distance used as cost measure :param int warp: how many shifts are computed. :param int w: window size limiting the maximal distance between indices of matched entries |i,j|. :param float s: weight applied on off-diagonal moves of the path. As s gets larger, the warping path is increasingly biased towards the diagonal Returns the minimum distance, the cost matrix, the accumulated cost matrix, and the wrap path. """ assert len(x) assert len(y) assert isinf(w) or (w >= abs(len(x) - len(y))) assert s > 0 r, c = len(x), len(y) if not isinf(w): D0 = full((r + 1, c + 1), inf) for i in range(1, r + 1): D0[i, max(1, i - w):min(c + 1, i + w + 1)] = 0 D0[0, 0] = 0 else: D0 = zeros((r + 1, c + 1)) D0[0, 1:] = inf D0[1:, 0] = inf D1 = D0[1:, 1:] # view for i in range(r): for j in range(c): if (isinf(w) or (max(0, i - w) <= j <= min(c, i + w))): D1[i, j] = dist(x[i], y[j]) C = D1.copy() jrange = range(c) for i in range(r): if not isinf(w): jrange = range(max(0, i - w), min(c, i + w + 1)) for j in jrange: min_list = [D0[i, j]] for k in range(1, warp + 1): i_k = min(i + k, r) j_k = min(j + k, c) min_list += [D0[i_k, j] * s, D0[i, j_k] * s] D1[i, j] += min(min_list) if len(x) == 1: path = zeros(len(y)), range(len(y)) elif len(y) == 1: path = range(len(x)), zeros(len(x)) else: path = _traceback(D0) return D1[-1, -1], C, D1, path
Computes Dynamic Time Warping (DTW) of two sequences. :param array x: N1*M array :param array y: N2*M array :param func dist: distance used as cost measure :param int warp: how many shifts are computed. :param int w: window size limiting the maximal distance between indices of matched entries |i,j|. :param float s: weight applied on off-diagonal moves of the path. As s gets larger, the warping path is increasingly biased towards the diagonal Returns the minimum distance, the cost matrix, the accumulated cost matrix, and the wrap path.
8,377
from numpy import array, zeros, full, argmin, inf, ndim from scipy.spatial.distance import cdist from math import isinf def _traceback(D): i, j = array(D.shape) - 2 p, q = [i], [j] while (i > 0) or (j > 0): tb = argmin((D[i, j], D[i, j + 1], D[i + 1, j])) if tb == 0: i -= 1 j -= 1 elif tb == 1: i -= 1 else: # (tb == 2): j -= 1 p.insert(0, i) q.insert(0, j) return array(p), array(q) The provided code snippet includes necessary dependencies for implementing the `accelerated_dtw` function. Write a Python function `def accelerated_dtw(x, y, dist, warp=1)` to solve the following problem: Computes Dynamic Time Warping (DTW) of two sequences in a faster way. Instead of iterating through each element and calculating each distance, this uses the cdist function from scipy (https://docs.scipy.org/doc/scipy/reference/generated/scipy.spatial.distance.cdist.html) :param array x: N1*M array :param array y: N2*M array :param string or func dist: distance parameter for cdist. When string is given, cdist uses optimized functions for the distance metrics. If a string is passed, the distance function can be 'braycurtis', 'canberra', 'chebyshev', 'cityblock', 'correlation', 'cosine', 'dice', 'euclidean', 'hamming', 'jaccard', 'kulsinski', 'mahalanobis', 'matching', 'minkowski', 'rogerstanimoto', 'russellrao', 'seuclidean', 'sokalmichener', 'sokalsneath', 'sqeuclidean', 'wminkowski', 'yule'. :param int warp: how many shifts are computed. Returns the minimum distance, the cost matrix, the accumulated cost matrix, and the wrap path. Here is the function: def accelerated_dtw(x, y, dist, warp=1): """ Computes Dynamic Time Warping (DTW) of two sequences in a faster way. Instead of iterating through each element and calculating each distance, this uses the cdist function from scipy (https://docs.scipy.org/doc/scipy/reference/generated/scipy.spatial.distance.cdist.html) :param array x: N1*M array :param array y: N2*M array :param string or func dist: distance parameter for cdist. When string is given, cdist uses optimized functions for the distance metrics. If a string is passed, the distance function can be 'braycurtis', 'canberra', 'chebyshev', 'cityblock', 'correlation', 'cosine', 'dice', 'euclidean', 'hamming', 'jaccard', 'kulsinski', 'mahalanobis', 'matching', 'minkowski', 'rogerstanimoto', 'russellrao', 'seuclidean', 'sokalmichener', 'sokalsneath', 'sqeuclidean', 'wminkowski', 'yule'. :param int warp: how many shifts are computed. Returns the minimum distance, the cost matrix, the accumulated cost matrix, and the wrap path. """ assert len(x) assert len(y) if ndim(x) == 1: x = x.reshape(-1, 1) if ndim(y) == 1: y = y.reshape(-1, 1) r, c = len(x), len(y) D0 = zeros((r + 1, c + 1)) D0[0, 1:] = inf D0[1:, 0] = inf D1 = D0[1:, 1:] D0[1:, 1:] = cdist(x, y, dist) C = D1.copy() for i in range(r): for j in range(c): min_list = [D0[i, j]] for k in range(1, warp + 1): min_list += [D0[min(i + k, r), j], D0[i, min(j + k, c)]] D1[i, j] += min(min_list) if len(x) == 1: path = zeros(len(y)), range(len(y)) elif len(y) == 1: path = range(len(x)), zeros(len(x)) else: path = _traceback(D0) return D1[-1, -1], C, D1, path
Computes Dynamic Time Warping (DTW) of two sequences in a faster way. Instead of iterating through each element and calculating each distance, this uses the cdist function from scipy (https://docs.scipy.org/doc/scipy/reference/generated/scipy.spatial.distance.cdist.html) :param array x: N1*M array :param array y: N2*M array :param string or func dist: distance parameter for cdist. When string is given, cdist uses optimized functions for the distance metrics. If a string is passed, the distance function can be 'braycurtis', 'canberra', 'chebyshev', 'cityblock', 'correlation', 'cosine', 'dice', 'euclidean', 'hamming', 'jaccard', 'kulsinski', 'mahalanobis', 'matching', 'minkowski', 'rogerstanimoto', 'russellrao', 'seuclidean', 'sokalmichener', 'sokalsneath', 'sqeuclidean', 'wminkowski', 'yule'. :param int warp: how many shifts are computed. Returns the minimum distance, the cost matrix, the accumulated cost matrix, and the wrap path.
8,378
import re import six from six.moves import range The provided code snippet includes necessary dependencies for implementing the `strip_ids` function. Write a Python function `def strip_ids(ids, ids_to_strip)` to solve the following problem: Strip ids_to_strip from the end ids. Here is the function: def strip_ids(ids, ids_to_strip): """Strip ids_to_strip from the end ids.""" ids = list(ids) while ids and ids[-1] in ids_to_strip: ids.pop() return ids
Strip ids_to_strip from the end ids.
8,379
import subprocess import matplotlib import os import librosa import librosa.filters import numpy as np from scipy import signal from scipy.io import wavfile def save_wav(wav, path, sr, norm=False): if norm: wav = wav / np.abs(wav).max() wav *= 32767 # proposed by @dsmiller wavfile.write(path, sr, wav.astype(np.int16))
null
8,380
import subprocess import matplotlib import os import librosa import librosa.filters import numpy as np from scipy import signal from scipy.io import wavfile def get_hop_size(hparams): def _stft(y, hparams): return librosa.stft(y=y, n_fft=hparams['fft_size'], hop_length=get_hop_size(hparams), win_length=hparams['win_size'], pad_mode='constant')
null
8,381
import subprocess import matplotlib import os import librosa import librosa.filters import numpy as np from scipy import signal from scipy.io import wavfile def get_hop_size(hparams): hop_size = hparams['hop_size'] if hop_size is None: assert hparams['frame_shift_ms'] is not None hop_size = int(hparams['frame_shift_ms'] / 1000 * hparams['audio_sample_rate']) return hop_size def _istft(y, hparams): return librosa.istft(y, hop_length=get_hop_size(hparams), win_length=hparams['win_size'])
null
8,382
import subprocess import matplotlib import os import librosa import librosa.filters import numpy as np from scipy import signal from scipy.io import wavfile def denormalize(D, hparams): return (D * -hparams['min_level_db']) + hparams['min_level_db']
null
8,383
import subprocess import matplotlib import os import librosa import librosa.filters import numpy as np from scipy import signal from scipy.io import wavfile def rnnoise(filename, out_fn=None, verbose=False, out_sample_rate=22050): assert os.path.exists('./rnnoise/examples/rnnoise_demo'), INSTALL_STR if out_fn is None: out_fn = f"{filename[:-4]}.denoised.wav" out_48k_fn = f"{out_fn}.48000.wav" tmp0_fn = f"{out_fn}.0.wav" tmp1_fn = f"{out_fn}.1.wav" tmp2_fn = f"{out_fn}.2.raw" tmp3_fn = f"{out_fn}.3.raw" if verbose: print("Pre-processing audio...") # wav to pcm raw subprocess.check_call( f'sox "{filename}" -G -r48000 "{tmp0_fn}"', shell=True, stdin=subprocess.PIPE) # convert to raw subprocess.check_call( f'sox -v 0.95 "{tmp0_fn}" "{tmp1_fn}"', shell=True, stdin=subprocess.PIPE) # convert to raw subprocess.check_call( f'ffmpeg -y -i "{tmp1_fn}" -loglevel quiet -f s16le -ac 1 -ar 48000 "{tmp2_fn}"', shell=True, stdin=subprocess.PIPE) # convert to raw if verbose: print("Applying rnnoise algorithm to audio...") # rnnoise subprocess.check_call( f'./rnnoise/examples/rnnoise_demo "{tmp2_fn}" "{tmp3_fn}"', shell=True) if verbose: print("Post-processing audio...") # pcm raw to wav if filename == out_fn: subprocess.check_call(f'rm -f "{out_fn}"', shell=True) subprocess.check_call( f'sox -t raw -r 48000 -b 16 -e signed-integer -c 1 "{tmp3_fn}" "{out_48k_fn}"', shell=True) subprocess.check_call(f'sox "{out_48k_fn}" -G -r{out_sample_rate} "{out_fn}"', shell=True) subprocess.check_call(f'rm -f "{tmp0_fn}" "{tmp1_fn}" "{tmp2_fn}" "{tmp3_fn}" "{out_48k_fn}"', shell=True) if verbose: print("Audio-filtering completed!")
null
8,384
import librosa import numpy as np from pycwt import wavelet from scipy.interpolate import interp1d def load_wav(wav_file, sr): wav, _ = librosa.load(wav_file, sr=sr, mono=True) return wav
null