python_code
stringlengths
0
229k
import torch from torchaudio_unittest import common_utils from .kaldi_compatibility_impl import Kaldi class TestKaldiFloat32(Kaldi, common_utils.PytorchTestCase): dtype = torch.float32 device = torch.device('cpu') class TestKaldiFloat64(Kaldi, common_utils.PytorchTestCase): dtype = torch.float64 de...
import torch from torchaudio_unittest.common_utils import PytorchTestCase, skipIfNoCuda from .librosa_compatibility_test_impl import TransformsTestBase @skipIfNoCuda class TestTransforms(TransformsTestBase, PytorchTestCase): dtype = torch.float64 device = torch.device('cuda')
import torch from torchaudio_unittest.common_utils import PytorchTestCase from .librosa_compatibility_test_impl import TransformsTestBase class TestTransforms(TransformsTestBase, PytorchTestCase): dtype = torch.float64 device = torch.device('cpu')
"""Generate opus file for testing load functions""" import argparse import subprocess import scipy.io.wavfile import torch def _parse_args(): parser = argparse.ArgumentParser( description='Generate opus files for test' ) parser.add_argument('--num-channels', required=True, type=int) parser.a...
#!/usr/bin/env python3 """Generate the conf JSON from fairseq pretrained weight file, that is consumed by unit tests Usage: 1. Download pretrained parameters from https://github.com/pytorch/fairseq/tree/main/examples/wav2vec 2. Download the dict from https://dl.fbaipublicfiles.com/fairseq/wav2vec/dict.ltr.txt and p...
#!/usr/bin/env python3 """Generate the conf JSONs from fairseq pretrained weight file, consumed by unit tests Note: The current configuration files were generated on fairseq e47a4c84 Usage: 1. Download pretrained parameters from https://github.com/pytorch/fairseq/tree/main/examples/hubert 2. Run this script and s...
import os import json from transformers import Wav2Vec2Model _THIS_DIR = os.path.dirname(os.path.abspath(__file__)) def _main(): keys = [ # pretrained "facebook/wav2vec2-base", "facebook/wav2vec2-large", "facebook/wav2vec2-large-lv60", "facebook/wav2vec2-base-10k-voxpopul...
from typing import Callable, Tuple from functools import partial import torch from parameterized import parameterized from torch import Tensor import torchaudio.functional as F from torch.autograd import gradcheck, gradgradcheck from torchaudio_unittest.common_utils import ( TestBaseMixin, get_whitenoise, r...
import torch import unittest from torchaudio_unittest.common_utils import PytorchTestCase, skipIfNoCuda from .functional_impl import Functional @skipIfNoCuda class TestFunctionalFloat32(Functional, PytorchTestCase): dtype = torch.float32 device = torch.device('cuda') @unittest.expectedFailure def te...
import torch import torchaudio.functional as F from torchaudio_unittest.common_utils import ( skipIfNoSox, skipIfNoExec, TempDirMixin, TorchaudioTestCase, get_asset_path, sox_utils, load_wav, save_wav, get_whitenoise, ) @skipIfNoSox @skipIfNoExec('sox') class TestFunctionalFilteri...
import torch import torchaudio.functional as F import unittest from parameterized import parameterized from torchaudio_unittest.common_utils import PytorchTestCase, TorchaudioTestCase, skipIfNoSox from .functional_impl import Functional, FunctionalCPUOnly class TestFunctionalFloat32(Functional, FunctionalCPUOnly, Py...
import torch from .autograd_impl import Autograd, AutogradFloat32 from torchaudio_unittest import common_utils @common_utils.skipIfNoCuda class TestAutogradLfilterCUDA(Autograd, common_utils.PytorchTestCase): dtype = torch.float64 device = torch.device('cuda') @common_utils.skipIfNoCuda class TestAutogradRN...
"""Test suites for jit-ability and its numerical compatibility""" import unittest import torch import torchaudio.functional as F from torchaudio_unittest import common_utils from torchaudio_unittest.common_utils import ( TempDirMixin, TestBaseMixin, skipIfRocm, torch_script, ) class Functional(TempD...
from parameterized import parameterized import torch import torchaudio.functional as F from torchaudio_unittest.common_utils import ( get_sinusoid, load_params, save_wav, skipIfNoExec, TempDirMixin, TestBaseMixin, ) from torchaudio_unittest.common_utils.kaldi_utils import ( convert_args, ...
import torch from .autograd_impl import Autograd, AutogradFloat32 from torchaudio_unittest import common_utils class TestAutogradLfilterCPU(Autograd, common_utils.PytorchTestCase): dtype = torch.float64 device = torch.device('cpu') class TestAutogradRNNTCPU(AutogradFloat32, common_utils.PytorchTestCase): ...
"""Test numerical consistency among single input and batched input.""" import itertools import math from parameterized import parameterized, parameterized_class import torch import torchaudio.functional as F from torchaudio_unittest import common_utils def _name_from_args(func, _, params): """Return a parameter...
import unittest from distutils.version import StrictVersion import torch from parameterized import param import torchaudio.functional as F from torchaudio._internal.module_utils import is_module_available LIBROSA_AVAILABLE = is_module_available('librosa') if LIBROSA_AVAILABLE: import numpy as np import libr...
import torch from torchaudio_unittest.common_utils import skipIfNoCuda, PytorchTestCase from .torchscript_consistency_impl import Functional, FunctionalFloat32Only @skipIfNoCuda class TestFunctionalFloat32(Functional, FunctionalFloat32Only, PytorchTestCase): dtype = torch.float32 device = torch.device('cuda'...
import torch from torchaudio_unittest.common_utils import PytorchTestCase from .torchscript_consistency_impl import Functional, FunctionalFloat32Only class TestFunctionalFloat32(Functional, FunctionalFloat32Only, PytorchTestCase): dtype = torch.float32 device = torch.device('cpu') class TestFunctionalFloat...
import torch from torchaudio_unittest.common_utils import PytorchTestCase, skipIfNoCuda from .kaldi_compatibility_test_impl import Kaldi @skipIfNoCuda class TestKaldiFloat32(Kaldi, PytorchTestCase): dtype = torch.float32 device = torch.device('cuda') @skipIfNoCuda class TestKaldiFloat64(Kaldi, PytorchTestC...
import torch from torchaudio_unittest.common_utils import PytorchTestCase from .kaldi_compatibility_test_impl import Kaldi, KaldiCPUOnly class TestKaldiCPUOnly(KaldiCPUOnly, PytorchTestCase): dtype = torch.float32 device = torch.device('cpu') class TestKaldiFloat32(Kaldi, PytorchTestCase): dtype = torc...
from torchaudio_unittest.common_utils import PytorchTestCase, skipIfNoCuda from .librosa_compatibility_test_impl import Functional, FunctionalComplex @skipIfNoCuda class TestFunctionalCUDA(Functional, PytorchTestCase): device = 'cuda' @skipIfNoCuda class TestFunctionalComplexCUDA(FunctionalComplex, PytorchTestC...
from torchaudio_unittest.common_utils import PytorchTestCase from .librosa_compatibility_test_impl import Functional, FunctionalComplex class TestFunctionalCPU(Functional, PytorchTestCase): device = 'cpu' class TestFunctionalComplexCPU(FunctionalComplex, PytorchTestCase): device = 'cpu'
"""Test definition common to CPU and CUDA""" import math import itertools import warnings import numpy as np import torch import torchaudio.functional as F from parameterized import parameterized from scipy import signal from torchaudio_unittest.common_utils import ( TestBaseMixin, get_sinusoid, nested_pa...
#!/usr/bin/env python3 # -*- coding: utf-8 -*- # # PyTorch documentation build configuration file, created by # sphinx-quickstart on Fri Dec 23 13:31:47 2016. # # This file is execfile()d with the current directory set to its # containing dir. # # Note that not all possible configuration values are present in this # au...
#!/usr/bin/env python3 """ Create a data preprocess pipeline that can be run with libtorchaudio """ import os import argparse import torch import torchaudio class Pipeline(torch.nn.Module): """Example audio process pipeline. This example load waveform from a file then apply effects and save it to a file. ...
#!/usr/bin/env python """Parse a directory contains VoxForge dataset. Recursively search for "PROMPTS" file in the given directory and print out `<ID>\\t<AUDIO_PATH>\\t<TRANSCRIPTION>` example: python parse_voxforge.py voxforge/de/Helge-20150608-aku de5-001\t/datasets/voxforge/de/guenter-20140214-afn/wav/de5-00...
import torch class Decoder(torch.nn.Module): def __init__(self, labels): super().__init__() self.labels = labels def forward(self, logits: torch.Tensor) -> str: """Given a sequence logits over labels, get the best path string Args: logits (Tensor): Logit tensors. ...
#!/usr/bin/evn python3 """Build Speech Recognition pipeline based on fairseq's wav2vec2.0 and dump it to TorchScript file. To use this script, you need `fairseq`. """ import os import argparse import logging from typing import Tuple import torch from torch.utils.mobile_optimizer import optimize_for_mobile import torc...
#!/usr/bin/env python3 """Parse a directory contains Librispeech dataset. Recursively search for "*.trans.txt" file in the given directory and print out `<ID>\\t<AUDIO_PATH>\\t<TRANSCRIPTION>` example: python parse_librispeech.py LibriSpeech/test-clean 1089-134691-0000\t/LibriSpeech/test-clean/1089/134691/1089-...
#!/usr/bin/env python3 import argparse import logging import os from typing import Tuple import torch import torchaudio from torchaudio.models.wav2vec2.utils.import_huggingface import import_huggingface_model from greedy_decoder import Decoder TORCH_VERSION: Tuple[int, ...] = tuple(int(x) for x in torch.__version__.s...
import argparse import logging import os import unittest from interactive_asr.utils import setup_asr, transcribe_file class ASRTest(unittest.TestCase): logger = logging.getLogger(__name__) logger.setLevel(logging.INFO) arguments_dict = { "path": "/scratch/jamarshon/downloads/model.pt", "...
#!/usr/bin/env python3 """This is the preprocessing script for HuBERT model training. The script includes: - File list creation - MFCC/HuBERT feature extraction - KMeans clustering model training - Pseudo-label generation """ import logging from argparse import ArgumentParser, RawTextHelpFormatter from ...
from .common_utils import create_tsv from .feature_utils import dump_features from .kmeans import learn_kmeans, get_km_label __all__ = [ "create_tsv", "dump_features", "learn_kmeans", "get_km_label", ]
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # https://github.com/pytorch/fairseq/blob/265df7144c79446f5ea8d835bda6e727f54dad9d/LICENSE import logging from pathlib import Path from typing import ( Tuple, ) import jobli...
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # https://github.com/pytorch/fairseq/blob/265df7144c79446f5ea8d835bda6e727f54dad9d/LICENSE import logging from pathlib import Path from typing import ( Tuple, Union, ) i...
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # https://github.com/pytorch/fairseq/blob/265df7144c79446f5ea8d835bda6e727f54dad9d/LICENSE """ Data pre-processing: create tsv files for training (and valiation). """ import logg...
# ***************************************************************************** # Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are met: # * Redistributions...
# ***************************************************************************** # Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are met: # * Redistributions...
# ***************************************************************************** # Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are met: # * Redistributions...
# ***************************************************************************** # Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are met: # * Redistributions...
""" Text-to-speech pipeline using Tacotron2. """ from functools import partial import argparse import os import random import sys import torch import torchaudio import numpy as np from torchaudio.models import Tacotron2 from torchaudio.models import tacotron2 as pretrained_tacotron2 from utils import prepare_input_s...
# ***************************************************************************** # Copyright (c) 2017 Keith Ito # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including...
# ***************************************************************************** # Copyright (c) 2017 Keith Ito # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including...
from . import utils, vad __all__ = ['utils', 'vad']
#!/usr/bin/env python3 # Copyright (c) 2017-present, Facebook, Inc. # All rights reserved. # # This source code is licensed under the license found in the LICENSE file in # the root directory of this source tree. An additional grant of patent rights # can be found in the PATENTS file in the same directory. """ Run infe...
#!/usr/bin/env python3 # Copyright (c) 2017-present, Facebook, Inc. # All rights reserved. # # This source code is licensed under the license found in the LICENSE file in # the root directory of this source tree. An additional grant of patent rights # can be found in the PATENTS file in the same directory. import os im...
#!/usr/bin/env python3 # Copyright (c) 2017-present, Facebook, Inc. # All rights reserved. # # This source code is licensed under the license found in the LICENSE file in # the root directory of this source tree. An additional grant of patent rights # can be found in the PATENTS file in the same directory. """ Followin...
#!/usr/bin/env python3 """Launch souce separation training. This script runs training in Distributed Data Parallel (DDP) framework and has two major operation modes. This behavior depends on if `--worker-id` argument is given or not. 1. (`--worker-id` is not given) Launchs worker subprocesses that performs the actual...
#!/usr/bin/env python3 # pyre-strict from pathlib import Path from argparse import ArgumentParser from typing import ( Any, Callable, Dict, Mapping, List, Optional, Tuple, TypedDict, Union, ) import torch import torchaudio from pytorch_lightning import LightningModule, Trainer from...
from argparse import ArgumentParser from pathlib import Path from lightning_train import _get_model, _get_dataloader, sisdri_metric import mir_eval import torch def _eval(model, data_loader, device): results = torch.zeros(4) with torch.no_grad(): for _, batch in enumerate(data_loader): mi...
import math from typing import Optional from itertools import permutations import torch def sdr( estimate: torch.Tensor, reference: torch.Tensor, mask: Optional[torch.Tensor] = None, epsilon: float = 1e-8 ) -> torch.Tensor: """Computes source-to-distortion ratio. 1. scale the...
from . import ( dataset, dist_utils, metrics, ) __all__ = ['dataset', 'dist_utils', 'metrics']
import os import csv import types import logging import torch import torch.distributed as dist def _info_on_master(self, *args, **kwargs): if dist.get_rank() == 0: self.info(*args, **kwargs) def getLogger(name): """Get logging.Logger module with additional ``info_on_master`` method.""" logger =...
from . import utils, wsj0mix __all__ = ['utils', 'wsj0mix']
from typing import List from functools import partial from collections import namedtuple from torchaudio.datasets import LibriMix import torch from . import wsj0mix Batch = namedtuple("Batch", ["mix", "src", "mask"]) def get_dataset(dataset_type, root_dir, num_speakers, sample_rate, task=None, librimix_tr_split=No...
from pathlib import Path from typing import Union, Tuple, List import torch from torch.utils.data import Dataset import torchaudio SampleType = Tuple[int, torch.Tensor, List[torch.Tensor]] class WSJ0Mix(Dataset): """Create a Dataset for wsj0-mix. Args: root (str or Path): Path to the directory whe...
from . import ( train, trainer ) __all__ = ['train', 'trainer']
#!/usr/bin/env python3 """Train Conv-TasNet""" import time import pathlib import argparse import torch import torchaudio import torchaudio.models import conv_tasnet from utils import dist_utils from utils.dataset import utils as dataset_utils _LG = dist_utils.getLogger(__name__) def _parse_args(args): parser =...
import time from typing import Tuple from collections import namedtuple import torch import torch.distributed as dist from utils import dist_utils, metrics _LG = dist_utils.getLogger(__name__) Metric = namedtuple("SNR", ["si_snri", "sdri"]) Metric.__str__ = ( lambda self: f"SI-SNRi: {self.si_snri:10.3e}, SDRi: ...
import torch class Normalize(torch.nn.Module): def forward(self, tensor): return (tensor - tensor.mean(-1, keepdim=True)) / tensor.std(-1, keepdim=True) class UnsqueezeFirst(torch.nn.Module): def forward(self, tensor): return tensor.unsqueeze(0)
import torch from torchaudio.datasets import LIBRISPEECH class MapMemoryCache(torch.utils.data.Dataset): """ Wrap a dataset so that, whenever a new item is returned, it is saved to memory. """ def __init__(self, dataset): self.dataset = dataset self._cache = [None] * len(dataset) ...
import json import logging import os import shutil from collections import defaultdict import torch class MetricLogger(defaultdict): def __init__(self, name, print_freq=1, disable=False): super().__init__(lambda: 0.0) self.disable = disable self.print_freq = print_freq self._iter ...
from torch import topk class GreedyDecoder: def __call__(self, outputs): """Greedy Decoder. Returns highest probability of class labels for each timestep Args: outputs (torch.Tensor): shape (input length, batch size, number of classes (including blank)) Returns: t...
import collections import itertools class LanguageModel: def __init__(self, labels, char_blank, char_space): self.char_space = char_space self.char_blank = char_blank labels = list(labels) self.length = len(labels) enumerated = list(enumerate(labels)) flipped = [(...
import argparse import logging import os import string from datetime import datetime from time import time import torch import torchaudio from torch.optim import SGD, Adadelta, Adam, AdamW from torch.optim.lr_scheduler import ExponentialLR, ReduceLROnPlateau from torch.utils.data import DataLoader from torchaudio.data...
# -*- coding: utf-8 -*- """ Audio Resampling ================ Here, we will walk through resampling audio waveforms using ``torchaudio``. """ # When running this tutorial in Google Colab, install the required packages # with the following. # !pip install torchaudio librosa import torch import torchaudio import torc...
""" Speech Recognition with Wav2Vec2 ================================ **Author**: `Moto Hira <moto@fb.com>`__ This tutorial shows how to perform speech recognition using using pre-trained models from wav2vec 2.0 [`paper <https://arxiv.org/abs/2006.11477>`__]. """ ###################################################...
# -*- coding: utf-8 -*- """ Audio I/O ========= ``torchaudio`` integrates ``libsox`` and provides a rich set of audio I/O. """ # When running this tutorial in Google Colab, install the required packages # with the following. # !pip install torchaudio boto3 import torch import torchaudio print(torch.__version__) pri...
""" Forced Alignment with Wav2Vec2 ============================== **Author** `Moto Hira <moto@fb.com>`__ This tutorial shows how to align transcript to speech with ``torchaudio``, using CTC segmentation algorithm described in `CTC-Segmentation of Large Corpora for German End-to-end Speech Recognition <https://arxiv.o...
# -*- coding: utf-8 -*- """ Audio Feature Augmentation ========================== """ # When running this tutorial in Google Colab, install the required packages # with the following. # !pip install torchaudio librosa import torch import torchaudio import torchaudio.transforms as T print(torch.__version__) print(tor...
""" Text-to-Speech with Tacotron2 ============================= **Author** `Yao-Yuan Yang <https://github.com/yangarbiter>`__, `Moto Hira <moto@fb.com>`__ """ ###################################################################### # Overview # -------- # # This tutorial shows how to build text-to-speech pipeline, usi...
# -*- coding: utf-8 -*- """ Audio Feature Extractions ========================= ``torchaudio`` implements feature extractions commonly used in the audio domain. They are available in ``torchaudio.functional`` and ``torchaudio.transforms``. ``functional`` implements features as standalone functions. They are stateless...
""" MVDR with torchaudio ==================== **Author** `Zhaoheng Ni <zni@fb.com>`__ """ ###################################################################### # Overview # -------- # # This is a tutorial on how to apply MVDR beamforming by using `torchaudio <https://github.com/pytorch/audio>`__. # # Steps # # - Id...
# -*- coding: utf-8 -*- """ Audio Datasets ============== ``torchaudio`` provides easy access to common, publicly accessible datasets. Please refer to the official documentation for the list of available datasets. """ # When running this tutorial in Google Colab, install the required packages # with the following. # ...
# -*- coding: utf-8 -*- """ Audio Data Augmentation ======================= ``torchaudio`` provides a variety of ways to augment audio data. """ # When running this tutorial in Google Colab, install the required packages # with the following. # !pip install torchaudio import torch import torchaudio import torchaudio...
import random import torch from torch.utils.data.dataset import random_split from torchaudio.datasets import LJSPEECH, LIBRITTS from torchaudio.transforms import MuLawEncoding from processing import bits_to_normalized_waveform, normalized_waveform_to_bits class MapMemoryCache(torch.utils.data.Dataset): r"""Wrap...
# ***************************************************************************** # Copyright (c) 2019 fatchord (https://github.com/fatchord) # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software w...
import logging import os import shutil from collections import defaultdict, deque import torch class MetricLogger: r"""Logger for model metrics """ def __init__(self, group, print_freq=1): self.print_freq = print_freq self._iter = 0 self.data = defaultdict(lambda: deque(maxlen=se...
import argparse import torch import torchaudio from torchaudio.transforms import MelSpectrogram from torchaudio.models import wavernn from torchaudio.models.wavernn import _MODEL_CONFIG_AND_URLS from torchaudio.datasets import LJSPEECH from wavernn_inference_wrapper import WaveRNNInferenceWrapper from processing impo...
import math import torch from torch import nn as nn from torch.nn import functional as F class LongCrossEntropyLoss(nn.Module): r""" CrossEntropy loss """ def __init__(self): super(LongCrossEntropyLoss, self).__init__() def forward(self, output, target): output = output.transpose(1,...
import torch import torch.nn as nn class NormalizeDB(nn.Module): r"""Normalize the spectrogram with a minimum db value """ def __init__(self, min_level_db, normalization): super().__init__() self.min_level_db = min_level_db self.normalization = normalization def forward(self,...
import argparse import logging import os from collections import defaultdict from datetime import datetime from time import time from typing import List import torch import torchaudio from torch.optim import Adam from torch.utils.data import DataLoader from torchaudio.datasets.utils import bg_iterator from torchaudio....
""" This script finds the merger responsible for labeling a PR by a commit SHA. It is used by the workflow in '.github/workflows/pr-labels.yml'. If there exists no PR associated with the commit or the PR is properly labeled, this script is a no-op. Note: we ping the merger only, not the reviewers, as the reviewers can ...
# -*- coding: utf-8 -*- import math import warnings from typing import Callable, Optional import torch from torch import Tensor from torchaudio import functional as F from .functional.functional import ( _get_sinc_resample_kernel, _apply_sinc_resample_kernel, ) __all__ = [ 'Spectrogram', 'InverseSpe...
# To use this file, the dependency (https://github.com/vesis84/kaldi-io-for-python) # needs to be installed. This is a light wrapper around kaldi_io that returns # torch.Tensors. from typing import Any, Callable, Iterable, Tuple import torch from torch import Tensor from torchaudio._internal import module_utils as _mo...
from torchaudio import _extension # noqa: F401 from torchaudio import ( compliance, datasets, functional, models, pipelines, kaldi_io, utils, sox_effects, transforms, ) from torchaudio.backend import ( list_audio_backends, get_audio_backend, set_audio_backend, ) try: ...
import os import warnings from pathlib import Path import torch from torchaudio._internal import module_utils as _mod_utils # noqa: F401 def _init_extension(): if not _mod_utils.is_module_available('torchaudio._torchaudio'): warnings.warn('torchaudio C++ extension is not available.') return ...
from torch.hub import load_state_dict_from_url, download_url_to_file __all__ = [ "load_state_dict_from_url", "download_url_to_file", ]
import warnings import importlib.util from typing import Optional from functools import wraps import torch def is_module_available(*modules: str) -> bool: r"""Returns if a top-level module with :attr:`name` exists *without** importing it. This is generally safer than try-catch block around a `import X`. ...
import os from typing import Tuple, Optional, Union from pathlib import Path import torchaudio from torch.utils.data import Dataset from torch import Tensor from torchaudio.datasets.utils import ( download_url, extract_archive, ) FOLDER_IN_ARCHIVE = "SpeechCommands" URL = "speech_commands_v0.02" HASH_DIVIDER ...
from pathlib import Path from typing import Dict, Tuple, Union from torch import Tensor from torch.utils.data import Dataset import torchaudio from torchaudio.datasets.utils import ( download_url, extract_archive, validate_file, ) _URL = "https://datashare.ed.ac.uk/bitstream/handle/10283/3038/DR-VCTK.zi...
import os import re from pathlib import Path from typing import Iterable, Tuple, Union, List from torch.utils.data import Dataset from torchaudio.datasets.utils import download_url _CHECKSUMS = { "http://svn.code.sf.net/p/cmusphinx/code/trunk/cmudict/cmudict-0.7b": "825f4ebd9183f2417df9f067a9cabe86", "htt...
import os from pathlib import Path from typing import Tuple, Optional, Union import torchaudio from torch import Tensor from torch.utils.data import Dataset from torchaudio.datasets.utils import ( download_url, extract_archive, ) # The following lists prefixed with `filtered_` provide a filtered split # that:...
import os import csv from pathlib import Path from typing import Tuple, Union import torchaudio from torch import Tensor from torch.utils.data import Dataset from torchaudio.datasets.utils import ( download_url, extract_archive, ) URL = "aew" FOLDER_IN_ARCHIVE = "ARCTIC" _CHECKSUMS = { "http://festvox.org...
import csv import os from pathlib import Path from typing import List, Dict, Tuple, Union from torch import Tensor from torch.utils.data import Dataset import torchaudio def load_commonvoice_item(line: List[str], header: List[str], path: str, ...
from .commonvoice import COMMONVOICE from .librispeech import LIBRISPEECH from .speechcommands import SPEECHCOMMANDS from .vctk import VCTK_092 from .dr_vctk import DR_VCTK from .gtzan import GTZAN from .yesno import YESNO from .ljspeech import LJSPEECH from .cmuarctic import CMUARCTIC from .cmudict import CMUDict from...