Noise and audio augmentation techniques
The noise and data augmentation techniques were written in an effort to understand how augmenatation can affect model robustness and performance in both clean and noisy settings.
All transforms discussed in this section are subclasses of AudioFeatureTransform, AudioWaveformTransform, or AudioDatasetTransform. Each Audio*Transform has unique interaction with the data. If interested in implemented one's own transforms, it is highly advisable to review the differences (see Adding your own transforms). If only applying the in-built transforms, then one only needs to be mindful that the correct kind of transform is listed in the config (see Using transforms). These transforms can be applied to instances of SpeechToTextDataset.
Contents
In-built transforms
1. Utterance concatenation
Utterance concatenation is a data augmenation technique introduced as ConcatAug in Translatotron 2: High-quality direct speech-to-speech translation
with voice preservation.
With some parameterized probability, samples are concatenated with one other randomly chosen sample from the whole dataset. In the positive (concatenation) case, accessing dataset[i] will return a SpeechToTextDatasetItem where source=source[i]+source[j] and target=target[i]+target[j]. In the negative (skip concatenation) case, accessing dataset[i] will return a SpeechToTextDatasetItem where source=source[i] and target=target[i] as usual.
Usage: concataugment is an AudioDatasetTransform and has three configurable hyperparameters:
rate: probability that any single access will result in the positive (concatenation) case. Defaults to 0.25.max_tokens: maximum number of tokens allowed for concatenated source sequences. This parameter is meant to limit the length of concatenated samples to avoid out-of-memory errors. Defaults to 300.attempts: maximum number of invalid concatenation attempts before defaulting to the negative (skip concatenation) case. This parameter aims to limit excessive time spent trying to find candidate samples that are short enough to concatenate with. Defaults to 5.
Please be wary of OOMs while using this augmentation technique; we used smaller batch sizes as a workaround to avoid OOMs. Batch size is determined by update frequency, batch size hyperparameter, and the number of GPU, so you may want to alter these to this end.
2. Noise augmentation suite
The four noise augmentation methods in this suite adhere to the following principle: with some parameterized probability, samples are overlayed with a noise track. The content of the noise track is specific to the method. Signal-to-noise ratio with which the noise track is overlayed is determined by choosing a value from a random uniform distribution with parameterized endpoints. The first three methods are based off data augmentation methods suggested in Section 3.3 of X-Vectors: Robust DNN Embeddings for Speaker Recognition.
2.1. Music augmentation
For music augmentation, the noise track consists of one file uniformly randomly selected from a corpus of music files. The music file is cut to size, including being repeated to fill the original sample length if necessary.
Usage: musicaugment is an AudioWaveformTransform and has four configurable hyperparameters:
samples_path: path where background music files are saved as audios (.wav files). No default.rate: probability that any single access will result in the positive (background music) case. Defaults to 0.25.snr_min: lower endpoint of the range from which a signal-to-noise ratio is uniformly randomly chosen with which to add background noise to the original source. Defaults to 5.snr_max: higher endpoint of the range from which a signal-to-noise ratio is uniformly randomly chosen with which to add background noise to the original source. Defaults to 15.
2.2. Babble augmentation
For babble augmentation, the noise track consists of multiple audios uniformly randomly selected from a corpus of speech files. The number of speech audios in the background track is chosen randomly with equal probability between 3 and 7 audios.
Usage: babbleaugment is an AudioWaveformTransform and has four configurable hyperparameters:
samples_path: path where background speech files are saved as audios (.wav files). No default.rate: probability that any single access will result in the positive (background speech) case. Defaults to 0.25.snr_min: lower endpoint of the range from which a signal-to-noise ratio is uniformly randomly chosen with which to add background noise to the original source. Defaults to 5.snr_max: higher endpoint of the range from which a signal-to-noise ratio is uniformly randomly chosen with which to add background noise to the original source. Defaults to 15.
2.3. Sporadic noise augmentation
For sporadic noise augmentation, the noise track is mostly silent except for intermittent short clips of noise which are added at roughly a parameterized frequency. These clips are randomly chosen and cut from a corpus of noise files to lengths according to a parameterized Gaussian distribution.
Usage: sporadicnoiseaugment is an AudioWaveformTransform and has seven configurable hyperparameters:
samples_path: path where background noise files are saved as audios (.wav files). No default.rate: probability that any single access will result in the positive (add a sporadic noise track) case. Defaults to 0.25.snr_min: lower endpoint of the range from which a signal-to-noise ratio is uniformly randomly chosen with which to add background noise to the original source. Defaults to 5.snr_max: higher endpoint of the range from which a signal-to-noise ratio is uniformly randomly chosen with which to add background noise to the original source. Defaults to 15.noise_rate: rate in noises per second at which noise clip will be added to the original samplenoise_len_mean: mean of Gaussian normal distribution from which length of noise clip is chosennoise_len_std: standard deviation of Gaussian normal distribution from which length of noise clip is chosen
2.4. Background noise augmentation
For background noise augmentation, the noise track is a single track uniformly randomly selected from a corpus of noise files. The noise file is cut to size, including being repeated to fill the original sample length if necessary.
Usage: backgroundnoiseaugment is an AudioWaveformTransform and has four configurable hyperparameters:
samples_path: path where background noise files are saved as audios (.wav files). No default.rate: probability that any single access will result in the positive (background noise) case. Defaults to 0.25.snr_min: lower endpoint of the range from which a signal-to-noise ratio is uniformly randomly chosen with which to add background noise to the original source. Defaults to 5.snr_max: higher endpoint of the range from which a signal-to-noise ratio is uniformly randomly chosen with which to add background noise to the original source. Defaults to 15.
3. Mixed babble and background noise augmentation with recognizable source speaker
This augmentation technique is based on Algorithm 1 in WavLM: Large-Scale Self-Supervised Pre-Training for Full Stack Speech Processing and is similar to the noise augmentation suite techniques in that it has a background noise track. The noise track consists of either (1) another audio sample from the batch or (2) a background noise track. A key difference is the length of the noise track is chosen from a uniform random distribution between 0 and half of the original sample length.
Usage: noisyoverlapaugment is an AudioDatasetTransform and has seven configurable hyperparameters:
noises_path: path where background noise files are saved as audios (.wav files). No default.rate: probability that any single access will result in the positive (background noise) case. Defaults to 0.25.mixing_noise_rate: probability that in a positive (background noise) case, the noise track will consist of background noise (rather than babble from the batch). Defaults to 0.1.noise_snr_min: lower endpoint of the range from which a signal-to-noise ratio is uniformly randomly chosen with which to add background noise to the original source. Defaults to -5.noise_snr_max: higher endpoint of the range from which a signal-to-noise ratio is uniformly randomly chosen with which to add background noise to the original source. Defaults to 5.utterance_snr_min: lower endpoint of the range from which a signal-to-noise ratio is uniformly randomly chosen with which to add another audio from the batch to the original source. Defaults to -5.utterance_snr_max: higher endpoint of the range from which a signal-to-noise ratio is uniformly randomly chosen with which to add another audio from the batch to the original source. Defaults to 5.
Benchmark studies
Evaluation on clean data
| Augmentation in training data | Hyperparameters | Training loss | BLEU (covost) | BLEU (epst) | BLEU (mtedx) |
|---|---|---|---|---|---|
| None | 3.954 | 24.984 | 23.962 | 24.448 | |
| ConcatAugment | rate = 0.25, max_tokens = 3000, attempts = 5 | 3.940 | 25.322 | 26.124 | 26.19 |
| BabbleAugment | rate = 0.25, MUSAN speech, snr_min = (-5), snr_max = 5 | 3.957 | 24.226 | 23.186 | 22.368 |
| BackgroundNoiseAugment | rate = 0.1, MUSAN noises, snr_min = (-10), snr_max = 10 | 3.955 | 24.745 | 23.513 | 23.819 |
| MusicAugment | rate = 0.25, MUSAN music, snr_min = 0, snr_max = 20 | 3.954 | 25.096 | 24.301 | 23.341 |
| SporadicNoiseAugment | rate = 0.1, noise_rate = 0.25, MUSAN noises, snr_min = 10, snr_max = 35 | 3.954 | 24.924 | 23.951 | 23.484 |
| MusicAugment + BabbleAugment + BackgroundNoiseAugment + SporadicNoiseAugment | as above, except limited rates to sum to 0.25: music (0.074), background (0.029), babble (0.074), sporadic (0.029) | 3.953 | 24.874 | 23.675 | 24.249 |
| NoisyOverlapAugment | rate = 0.25, mixing_noise_rate = 0.5, MUSAN noises, utterance_snr_min = (-10), utterance_snr_max = 0, noise_snr_min = (-5), noise_snr_max = 20 | 3.954 | 24.949 | 24.015 | 23.768 |
Evaluation on data with music noise added at SNR = (-5) - 5
| Augmentation in training data | Training loss | BLEU (covost) | BLEU (epst) | BLEU (mtedx) |
|---|---|---|---|---|
| None | 3.954 | 15.785 | 21.105 | 16.944 |
| ConcatAugment | 3.940 | 17.186 | 23.255 | 18.24 |
| BabbleAugment | 3.957 | 19.158 | 22.064 | 17.116 |
| BackgroundNoiseAugment | 3.955 | 17.777 | 22.0 | 17.535 |
| MusicAugment | 3.954 | 20.345 | 23.126 | 19.433 |
| SporadicNoiseAugment | 3.954 | 15.927 | 21.382 | 14.736 |
| MusicAugment + BabbleAugment + BackgroundNoiseAugment + SporadicNoiseAugment | 3.953 | 19.724 | 22.659 | 17.852 |
| NoisyOverlapAugment | 3.954 | 17.49 | 22.142 | 17.207 |
Evaluation on data with babble noise added at SNR = (-5) - 5
| Augmentation in training data | Training loss | BLEU (covost) | BLEU (epst) | BLEU (mtedx) |
|---|---|---|---|---|
| None | 3.954 | 4.092 | 13.514 | 5.13 |
| ConcatAugment | 3.940 | 5.493 | 15.835 | 6.893 |
| BabbleAugment | 3.957 | 16.12 | 21.097 | 13.996 |
| BackgroundNoiseAugment | 3.955 | 4.691 | 15.784 | 5.982 |
| MusicAugment | 3.954 | 8.06 | 17.764 | 9.008 |
| SporadicNoiseAugment | 3.954 | 4.009 | 13.935 | 4.814 |
| MusicAugment + BabbleAugment + BackgroundNoiseAugment + SporadicNoiseAugment | 3.953 | 14.692 | 20.882 | 14.45 |
| NoisyOverlapAugment | 3.954 | 4.032 | 16.434 | 7.284 |
Evaluation on data with sporadic noise added at SNR = (-5) - 5
| Augmentation in training data | Training loss | BLEU (covost) | BLEU (epst) | BLEU (mtedx) |
|---|---|---|---|---|
| None | 3.954 | 23.778 | 23.745 | 22.748 |
| ConcatAugment | 3.940 | 24.239 | 25.907 | 25.723 |
| BabbleAugment | 3.957 | 23.42 | 23.048 | 21.076 |
| BackgroundNoiseAugment | 3.955 | 23.998 | 23.467 | 22.494 |
| MusicAugment | 3.954 | 24.142 | 24.181 | 19.143 |
| SporadicNoiseAugment | 3.954 | 23.97 | 23.894 | 22.61 |
| MusicAugment + BabbleAugment + BackgroundNoiseAugment + SporadicNoiseAugment | 3.953 | 24.118 | 23.59 | 23.717 |
| NoisyOverlapAugment | 3.954 | 24.265 | 24.103 | 23.167 |
Evaluation on data with background noise added at SNR = (-5) - 5
| Augmentation in training data | Training loss | BLEU (covost) | BLEU (epst) | BLEU (mtedx) |
|---|---|---|---|---|
| None | 3.954 | 20.201 | 22.525 | 19.66 |
| ConcatAugment | 3.940 | 20.904 | 24.706 | 21.353 |
| BabbleAugment | 3.957 | 20.687 | 22.374 | 18.907 |
| BackgroundNoiseAugment | 3.955 | 21.574 | 22.998 | 20.043 |
| MusicAugment | 3.954 | 21.65 | 23.529 | 19.87 |
| SporadicNoiseAugment | 3.954 | 20.578 | 22.577 | 19.096 |
| MusicAugment + BabbleAugment + BackgroundNoiseAugment + SporadicNoiseAugment | 3.953 | 21.811 | 23.144 | 20.986 |
| NoisyOverlapAugment | 3.954 | 21.312 | 23.153 | 20.302 |
Evaluation on data with all four types of noises added at SNR = (-5) - 5, each applied with prob 0.5
| Augmentation in training data | Training loss | BLEU (covost) | BLEU (epst) | BLEU (mtedx) |
|---|---|---|---|---|
| None | 3.954 | 10.895 | 19.319 | 12.748 |
| ConcatAugment | 3.940 | 13.517 | 21.658 | 15.428 |
| BabbleAugment | 3.957 | 18.09 | 21.384 | 16.018 |
| BackgroundNoiseAugment | 3.955 | 12.837 | 20.719 | 13.933 |
| MusicAugment | 3.954 | 16.589 | 21.823 | 15.927 |
| SporadicNoiseAugment | 3.954 | 11.238 | 19.91 | 13.31 |
| MusicAugment + BabbleAugment + BackgroundNoiseAugment + SporadicNoiseAugment | 3.953 | 18.636 | 21.935 | 17.845 |
| NoisyOverlapAugment | 3.954 | 12.829 | 20.856 | 15.048 |
Evaluation on data with noisy overlap augment
| Augmentation in training data | Training loss | BLEU (covost) | BLEU (epst) | BLEU (mtedx) |
|---|---|---|---|---|
| None | 3.954 | 21.245 | 22.24 | 20.994 |
| ConcatAugment | 3.940 | 21.611 | 24.247 | 23.068 |
| BabbleAugment | 3.957 | 21.867 | 21.987 | 20.099 |
| BackgroundNoiseAugment | 3.955 | 21.533 | 21.806 | 19.717 |
| MusicAugment | 3.954 | 21.823 | 22.643 | 20.847 |
| SporadicNoiseAugment | 3.954 | 21.373 | 22.381 | 20.672 |
| MusicAugment + BabbleAugment + BackgroundNoiseAugment + SporadicNoiseAugment | 3.953 | 22.206 | 22.414 | 21.375 |
| NoisyOverlapAugment | 3.954 | 23.371 | 23.396 | 22.627 |
Using transforms
Transforms are configurable.
- Please pay careful attention to the type of transform you are applying.
concataugmentandnoisyoverlapaugmentare instances ofAudioDatasetTransformand should be listed in the config underdataset_transforms.musicaugment,babbleaugment,sporadicnoiseaugment, andbackgroundnoiseaugmentare instances ofAudioWaveformTransformand should be listed underwaveform_transforms.- Instances of
AudioFeatureTransformshould be listed underfeature_transforms.
- Feel free to apply these augmentations in different contexts, e.g., you may use a
_trainor_evalflag to specify when the transform will be applied. If the dataset at hand containstrainin its name, those transforms under the_trainflag will be applied; else, the remaining transforms will be applied.
For example, you would add this to your config to apply the musicaugment transform to a training dataset:
musicaugment:
samples_path: ${MUSIC_PATH}
snr_min: 10
snr_max: 15
rate: 0.25
waveform_transforms:
_train:
- musicaugment
or add this to apply the concataugment transform:
concataugment:
rate: 0.25
max_tokens: 3000
attempts: 5
dataset_transforms:
_train:
- concataugment
You may also want to add multiple of one type of transform; here, we add multiple AudioWaveformTransforms:
musicaugment:
samples_path: ${MUSIC_PATH}
snr_min: 5
snr_max: 20
rate: 0.25
backgroundnoiseaugment:
samples_path: ${NOISES_PATH}
snr_min: 10
snr_max: 20
rate: 0.1
sporadicnoiseaugment:
samples_path: ${NOISES_PATH}
snr_min: 5
snr_max: 15
rate: 0.1
noise_rate: 0.25
waveform_transforms:
_train:
- musicaugment
- backgroundnoiseaugment
- sporadicnoiseaugment
Adding your own transforms
Note: We store transform implementations in fairseq/data/audio/*_transforms directories. You may refer to these as examples while implementing your own transform.
Step 1. Picking the right class for your transform
The integration into SpeechToTextDataset is quite different for each kind of transform, so it is important to understand which one is best suited to your purposes.
Feature transforms
AudioFeatureTransform is a base class which allows some transform to be applied to audio spectrograms in the data loading step. One thing to note is that the source data is either saved as np.ndarrays or as audio files, and is to be returned either as features (spectrogram) or waveform. If and only if the data is to be returned as a spectrogram, then AudioFeatureTransforms will be applied.
Waveform transforms
AudioWaveformTransform is a base class which allows some transform to be applied to waveforms in the data loading step. As mentioned above, there are two source and return types to data loading for this dataset. If and only if the data is saved in audio file format, then AudioWaveformTransforms will be applied, whichever return type is used.
Dataset transforms
AudioDatasetTransform is a base class for transforms based on more than one item in a dataset, ex. concatenation of two random samples in a dataset. Rather than being applied in a consistent way, i.e., to all features or to all waveforms, the integration of a dataset transform is entirely specific. Adding a dataset transform requires actually editing the fairseq/data/audio/speech_to_text_dataset.py file.
Step 2. Setting up your transform (generic to all types of transforms)
Now that you know which kind of transform you would like to use, we are ready to implement it. This step is generic for all transform types, i.e., TRANSFORM_TYPE may be any of feature, waveform, or dataset. We will show how to build utterance concatenation (an AudioDatasetTransform) as an example.
Import the base class and registration function for your transform.
from fairseq.data.audio.dataset_transforms import (
AudioDatasetTransform,
register_audio_dataset_transform
)
Define the class and register the transform. The name passed into the registration function is how your transform should be named in the config.
@register_audio_dataset_transform("concataugment")
class ConcatAugment(AudioDatasetTransform):
We are now ready to add the basic important functions to our new class. In this example, _DEFAULTS refers to a dictionary with the default hyperparameter values that we defined. from_config_dict is called to instantiate the transform given hyperparameters from the config.
@classmethod
def from_config_dict(cls, config=None):
_config = {} if config is None else config
return ConcatAugment(
_config.get("rate", _DEFAULTS["rate"]),
_config.get("max_tokens", _DEFAULTS["max_tokens"]),
_config.get("attempts", _DEFAULTS["attempts"]),
)
We edit the instantiation function __init__ to track hyperparameters and do any setup work.
def __init__(
self,
rate=_DEFAULTS["rate"],
max_tokens=_DEFAULTS["max_tokens"],
attempts=_DEFAULTS["attempts"],
):
self.rate, self.max_tokens, self.attempts = rate, max_tokens, attempts
Lastly __repr__ gives how the transform will be reported in an output log.
def __repr__(self):
return (
self.__class__.__name__
+ "("
+ ", ".join(
[
f"rate={self.rate}",
f"max_tokens={self.max_tokens}",
f"attempts={self.attempts}",
]
)
+ ")"
)
Step 3. Adding the transform logic
At this point, we are ready to implement the actual transform logic. The flow from here is different for each of the three transforms, so follow the path that is relevant to you.
...for feature transforms
The final step is implementing the __call__ function, which applies the transform logic and returns the spectrogram with transform applied. This supports and should take exactly two arguments:
selfx(np.ndarray): the spectrogram for one source sample. (This is a positional argument, so you can use another parameter name likespectrograminstead ofx.)
For example, this is the __call__ function for GlobalCMVN (cepstral mean and variance normalization).
def __call__(self, x):
x = np.subtract(x, self.mean)
x = np.divide(x, self.std)
return x
...for waveform transforms
The final step is implementing the __call__ function, which applies the transform logic. This supports and should take exactly three arguments:
selfsource(numpy.ndarray or torch.Tensor): source audio 2d waveform (channels x length)sample_rate(optional, defaults to None): sample rate ofsource
__call__ returns:
- transformed audio waveform
- sample rate of transformed audio waveform
For example, this is the __call__ function for augmentations in the Noise Augmentation Suite.
def __call__(self, source, sample_rate=None):
if np.random.random() > self.rate:
return source
noise = self._get_noise(
source.shape, always_2d=True, use_sample_rate=sample_rate
)
return self._mix(source, noise, rand_uniform(self.snr_min, self.snr_max)), sample_rate
...for dataset transforms
Dataset transforms are extremely flexible, and implementation involves directly integrating them into fairseq/data/audio/speech_to_text_dataset.py in transform-specific ways.
There are two basic components: (1) check whether or not this transform is part of this dataset instance using self.dataset_transforms.has_transform(TRANSFORM_CLS), and (2) if so, get the transform using self.dataset_transforms.get_transform(TRANSFORM_CLS) & apply it.
Due to the case-by-case specificity, it is easier to demonstrate this by examples.
Example: NoisyOverlapAugment
This transform requires access to multiple items within the same batch at once.
Logic: We still use the transform classes to keep away the transform logic. For example, __call__ of NoisyOverlapAugment class takes a list of source tokens for items in a mini-batch, applies noise/utterance as dictated by the transform, and returns the list of transformed source tokens for items in the mini-batch.
def __call__(self, sources):
for i, source in enumerate(sources):
if np.random.random() > self.rate:
continue
pri = source.numpy()
# ... some transform code omitted
pri[s_source : s_source + l] = np.add(
pri[s_source : s_source + l], np.multiply(scl, sec[s_sec : s_sec + l])
)
sources[i] = torch.from_numpy(pri).float()
return sources
Integration: The collater function for SpeechToTextDataset is responsible for preparing a mini-batch for training, so we integrate NOAug through adding a few lines to the top of this function:
def collater(
self, samples: List[SpeechToTextDatasetItem], return_order: bool = False
) -> Dict:
if len(samples) == 0:
return {}
indices = torch.tensor([x.index for x in samples], dtype=torch.long)
sources = [x.source for x in samples]
# NOAUG INTEGRATION BLOCK
# (1) Check whether or not this transform is part of this dataset instance
has_NOAug = self.dataset_transforms.has_transform(NoisyOverlapAugment)
# (2) If so, get & apply the transform
if has_NOAug and self.cfg.use_audio_input:
NOAug = self.dataset_transforms.get_transform(NoisyOverlapAugment)
sources = NOAug(sources)
frames = _collate_frames(sources, self.cfg.use_audio_input)
# sort samples by descending number of frames
n_frames = torch.tensor([x.size(0) for x in sources], dtype=torch.long)
n_frames, order = n_frames.sort(descending=True)
indices = indices.index_select(0, order)
frames = frames.index_select(0, order)
# ... rest of function
Example: ConcatAugment
This transform requires access to another item within the dataset at once.
Logic: We abstract the logic for picking indices to concatenate by adding a find_indices function to the ConcatAugment class, which takes one index in the dataset and finds a compatible second index to concatenate source and target tokens.
def find_indices(self, index: int, n_frames: List[int], n_samples: int):
# skip conditions: application rate, max_tokens limit exceeded
if np.random.random() > self.rate:
return [index]
if self.max_tokens and n_frames[index] > self.max_tokens:
return [index]
# pick second sample to concatenate
for _ in range(self.attempts):
index2 = np.random.randint(0, n_samples)
if index2 != index and (
not self.max_tokens
or n_frames[index] + n_frames[index2] < self.max_tokens
):
return [index, index2]
return [index]
Integration: SpeechToTextDataset uses a custom __getitem__(self, index) function (called in the background when you write dataset[i]). We edited this function (as well as _get_source_audio and get_tokenized_tgt_text) to achieve the desired transform effect where accessing dataset[i] will return a SpeechToTextDatasetItem where source=source[i]+source[j] and target=target[i]+target[j].
def __getitem__(self, index: int) -> SpeechToTextDatasetItem:
# CONCATAUGMENT INTEGRATION BLOCK
# (1) Check whether or not this transform is part of this dataset instance
has_concat = self.dataset_transforms.has_transform(ConcatAugment)
# (2) If so, get & apply the transform
if has_concat:
concat = self.dataset_transforms.get_transform(ConcatAugment)
indices = concat.find_indices(index, self.n_frames, self.n_samples)
source = self._get_source_audio(indices if has_concat else index)
source = self.pack_frames(source)
target = None
if self.tgt_texts is not None:
tokenized = self.get_tokenized_tgt_text(indices if has_concat else index)
target = self.tgt_dict.encode_line(
# ... rest of function