Buckets:

|
download
raw
82.2 kB

Title: Speaking Style Conversion in the Waveform Domain Using Discrete Self-Supervised Units

URL Source: https://arxiv.org/html/2212.09730

Markdown Content: Gallil Maimon, Yossi Adi

School of Computer Science and Engineering

The Hebrew University of Jerusalem

gallil.maimon@mail.huji.ac.il

Abstract

We introduce DISSC, a novel, lightweight method that converts the rhythm, pitch contour and timbre of a recording to a target speaker in a textless manner. Unlike DISSC, most voice conversion (VC) methods focus primarily on timbre, and ignore people’s unique speaking style (prosody). The proposed approach uses a pretrained, self-supervised model for encoding speech to discrete units, which makes it simple, effective, and fast to train. All conversion modules are only trained on reconstruction like tasks, thus suitable for any-to-many VC with no paired data. We introduce a suite of quantitative and qualitative evaluation metrics for this setup, and empirically demonstrate that DISSC significantly outperforms the evaluated baselines. Code and samples are available at https://pages.cs.huji.ac.il/adiyoss-lab/dissc/.

1 Introduction

Imagine hearing a famous catchphrase of your favourite television character spoken in their voice, but uncharacteristically fast, while atypically emphasising the end. This would immediately raise suspicion that something is “wrong”. As humans we learn to recognise familiar people and voices, not only by their voice texture (timbre), but also by their typical speaking style Williams (1965). Therefore, a true VC method should convert both voice texture and speaking style (rhythm, F0, etc.). Figure1 describes this visually.

Image 1: Refer to caption

Figure 1: Comparing traditional VC to speaking style conversion. White lines on the spectrograms show the pitch contour. While VC methods change the spectral features so the new utterance sounds like the new speaker, they do not change the rhythm and pitch. Conversely, SSC matches the faster speaking style, and the target speaker’s tendency to finish sentences with a pitch increase. This a real sample converted by DISSC. Due to variance within speaking style we do not expect the converted speech to exactly match the target.

Traditional VC methods mainly focused on changing the timbre of a given speaker while leaving the speaking style unchanged Stylianou et al. (1998); Kain and Macon (1998); Nakashika et al. (2013); Chou et al. (2019); Huang et al. (2021). Recent methods propose to additionally convert speaking style Qian et al. (2020); Chen and Duan (2022); Qian et al. (2021); Kuhlmann et al. (2022). However, these mainly use only a single target utterance which does not fully capture speaker prosody.

Another line of work uses text transcription supervision and shows the potential benefit to VC Kim et al. (2022); Liu et al. (2021). However, this limits the application to high resource languages and requires large scale data labelling.

Some recent prosody aware VC methods use spectrogram representations as input and output Qu et al. (2022); Chen et al. (2022), rather than operating in the waveform domain. Thus, they involve another phase of converting from the spectral domain to the time domain using a vocoder. Moreover, using discrete self-supervised speech representations and generating waveforms from these was demonstrated to provide superior performance on plenty of downstream tasks such as speech and audio language modelling(Lakhotia et al., 2021; Borsos et al., 2022; Qian et al., 2022), multi-stream processing(Kharitonov et al., 2022b), speech emotion conversion(Kreuk et al., 2021), spoken dialogue(Nguyen et al., 2022), speech-to-speech translation(Lee et al., 2022a, b; Popuri et al., 2022), and audio generation Kreuk et al. (2022a, b).

Inspired by this line of work, we propose a simple and effective method for Speaking Style Conversion (SSC) based on discrete self-supervised and partially disentangled speech representations. Curating samples of different speakers pronouncing the same utterances is challenging, thus we follow the setup of unpaired data, similarly to Lee et al. (2021); Yuan et al. (2021); Kameoka et al. (2018); Kaneko and Kameoka (2017). We give exact details of our approach in Section3. We formalise the setup of SSC and introduce an extensive evaluation suite in Section4. Results suggest the proposed approach is greatly superior to the evaluated baselines (see Section5), while utilising a simple and effective method which can be trained on a single GPU within a couple of hours.

Our contributions: (i) We introduce a novel, simple, and effective method which uses discrete self-supervised speech representations to perform textless speaking style conversion in the waveform domain; (ii) We formalise the task of speaking style conversion and propose diverse evaluation methods to analyze different speech characteristics; (iii) We empirically demonstrate the efficacy of the proposed approach using extensive empirical evaluation on several benchmarks considering both objective and subjective metrics.

2 Related Work

Unpaired Voice Conversion.

Many existing methods perform VC in the setting of unpaired utterances Kameoka et al. (2018); Kaneko and Kameoka (2017). Most of them contain a vocoder which generates the audio (or an intermediate representation) from several representations, one meant to capture speaker information and others to capture the remaining information Lin et al. (2021b, a). They encourage disentanglement through information bottlenecks by neural network architecture Chen et al. (2021); Qian et al. (2019), mutual information and adversarial losses Lee et al. (2021); Yuan et al. (2021), pretrained models Huang et al. (2022); Polyak et al. (2021) or combinations of these.

Notably, Polyak et al. (2021) used discrete HuBERT Hsu et al. (2021) tokens, pitch representation based on YAAPT Kasi and Zahorian (2002) and a learned speaker representation to re-synthesise the audio waveform. They showed how replacing the speaker representation at inference time performs VC. Inspired by this modeling paradigm, our approach additionally introduces speaking style modeling and conversion (i.e., pitch contour and rhythm) thus performing SSC.

Speaking Style Conversion.

Various existing methods aim to change the prosody of a given utterance. However, many convert it based on a single target utterance Qian et al. (2020); Chen and Duan (2022) which does not truly capture the general speaking style of the speaker, and often creates artifacts when the contents do not match. Other methods only linearly alter the speaking rate Kuhlmann et al. (2022) thus ignoring the change of rhythm for different content.

Even leading SSC methods are still limited. Some focus only on rhythm, paying less attention to pitch changes Qian et al. (2021); Lee et al. (2022c). Other approaches require text transcriptions for training Lee et al. (2022c); Liu et al. (2021). In contrast, our approach uses HuBERT units trained in a SSL fashion without using textual annotations, thus can support lower resource languages (see Fig. 6 in Appendix A.5). In addition, two concurrent methods for SSC Qu et al. (2022); Chen et al. (2022) were proposed recently. Although showing impressive results, they are based on spectrogram representations, hence require an additional vocoding step.

3 Our Approach - DISSC

We operate under the following setup: we perform any-to-many SSC which means we take any recording and convert it to any of the training set speakers. Our training set contains only unpaired utterances - e.g. no training utterance is said by more than one speaker. The learning process is completely textless with no form of text supervision.

Our approach uses DI screte units for S peaking S tyle C onversion, and is denoted DISSC. We decompose speech to few representations to synthesise speech in the target speaking style. We consider three components in the decomposition: phonetic-content, prosodic features (i.e., F0 and duration) and speaker identity, denoted by 𝐳 c,(𝐳 d⁢u⁢r,𝐳 F 0),𝐳 s⁢p⁢k subscript 𝐳 𝑐 subscript 𝐳 𝑑 𝑢 𝑟 subscript 𝐳 subscript 𝐹 0 subscript 𝐳 𝑠 𝑝 𝑘\mathbf{z}{c},(\mathbf{z}{dur},\mathbf{z}{F{0}}),\mathbf{z}{spk}bold_z start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT , ( bold_z start_POSTSUBSCRIPT italic_d italic_u italic_r end_POSTSUBSCRIPT , bold_z start_POSTSUBSCRIPT italic_F start_POSTSUBSCRIPT 0 end_POSTSUBSCRIPT end_POSTSUBSCRIPT ) , bold_z start_POSTSUBSCRIPT italic_s italic_p italic_k end_POSTSUBSCRIPT respectively. We propose a cascaded pipeline: (i) extract 𝐳 c subscript 𝐳 𝑐\mathbf{z}{c}bold_z start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT from the raw waveform using a SSL model; (ii) predict the prosodic features of the target speaker from 𝐳 c subscript 𝐳 𝑐\mathbf{z}{c}bold_z start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT and 𝐳 s⁢p⁢k subscript 𝐳 𝑠 𝑝 𝑘\mathbf{z}{spk}bold_z start_POSTSUBSCRIPT italic_s italic_p italic_k end_POSTSUBSCRIPT; (iii) synthesise the waveform speech from the content, predicted prosody and target speaker identity. See Figure2 for a visual description of DISSC. This modelling approach means moving from a continuous space to a discrete space, hence resulting in easier optimisation and better quality generations.

3.1 Speech Input Representation

Phonetic-content representation.

To represent speech phonetic-content we extract a discrete representation of the audio signal using a pre-trained SSL model, namely HuBERT Hsu et al. (2021). We use a SSL representation for phonetic-content, and not text or text based methods, to maintain non-textual cues like laughter and allow support for diverse languages. We discretise this representation to use as a rhythm proxy (by number of repetitions) and for ease of modelling. We chose HuBERT for the phonetic-content units as Polyak et al. (2021) showed it better disentangles between speech content and both speaker and prosody compared to other SSL-based models.

Denote the domain of audio samples by 𝒳⊂ℝ 𝒳 ℝ\mathcal{X}\subset\mathbb{R}caligraphic_X ⊂ blackboard_R. Audio waveforms are therefore represented by a sequence of samples 𝐱=(x 1,…,x T)𝐱 superscript 𝑥 1…superscript 𝑥 𝑇\mathbf{x}=(x^{1},\dots,x^{T})bold_x = ( italic_x start_POSTSUPERSCRIPT 1 end_POSTSUPERSCRIPT , … , italic_x start_POSTSUPERSCRIPT italic_T end_POSTSUPERSCRIPT ), where each x t∈𝒳 superscript 𝑥 𝑡 𝒳 x^{t}\in\mathcal{X}italic_x start_POSTSUPERSCRIPT italic_t end_POSTSUPERSCRIPT ∈ caligraphic_X for all 1≤t≤T 1 𝑡 𝑇 1\leq t\leq T 1 ≤ italic_t ≤ italic_T. The content encoder E c subscript 𝐸 𝑐 E_{c}italic_E start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT is a HuBERT model pre-trained on the LibriSpeech corpus Panayotov et al. (2015). The input to the content encoder E c subscript 𝐸 𝑐 E_{c}italic_E start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT is an audio waveform 𝐱 𝐱\mathbf{x}bold_x, and the output is a latent representation with lower temporal resolution 𝐳′=(z c 1′,…,z c L′)\mathbf{z}^{\prime}=(z_{c}^{{}^{\prime}1},\dots,z_{c}^{{}^{\prime}L})bold_z start_POSTSUPERSCRIPT ′ end_POSTSUPERSCRIPT = ( italic_z start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT start_POSTSUPERSCRIPT start_FLOATSUPERSCRIPT ′ end_FLOATSUPERSCRIPT 1 end_POSTSUPERSCRIPT , … , italic_z start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT start_POSTSUPERSCRIPT start_FLOATSUPERSCRIPT ′ end_FLOATSUPERSCRIPT italic_L end_POSTSUPERSCRIPT ) where L=T/n 𝐿 𝑇 𝑛 L=T/n italic_L = italic_T / italic_n. Since HuBERT outputs continuous representations, one needs an additional k-means step in order to quantise these representations into a discrete unit sequence. This sequence is denoted by 𝐳 c=(z c 1,…,z c L)subscript 𝐳 𝑐 superscript subscript 𝑧 𝑐 1…superscript subscript 𝑧 𝑐 𝐿\mathbf{z}{c}=(z{c}^{1},\dots,z_{c}^{L})bold_z start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT = ( italic_z start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT start_POSTSUPERSCRIPT 1 end_POSTSUPERSCRIPT , … , italic_z start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_L end_POSTSUPERSCRIPT ) where z c i∈{1,…,K}superscript subscript 𝑧 𝑐 𝑖 1…𝐾 z_{c}^{i}\in{1,\dots,K}italic_z start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_i end_POSTSUPERSCRIPT ∈ { 1 , … , italic_K } and K 𝐾 K italic_K is the vocabulary size. For the rest of the paper, we refer to these discrete representations as “units”. We extracted representations from the final layer of HuBERT and set K=100 𝐾 100 K=100 italic_K = 100. When we wish to predict the rhythm, we must decompose it from the sequence, therefore repeated units are omitted (e.g., 0,0,0,1,2,2→0,1,2 formulae-sequence→0 0 0 1 2 2 0 1 2 0,0,0,1,2,2\rightarrow 0,1,2 0 , 0 , 0 , 1 , 2 , 2 → 0 , 1 , 2) and the number of repetitions correlates to the rhythm. Such sequences are denoted as “deduped”.

Image 2: Refer to caption

Figure 2: An overview of DISSC. We use an SSL pretrained content encoder E c subscript 𝐸 𝑐 E_{c}italic_E start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT to extract discrete units from the waveform. We independently train a rhythm predictor E d⁢u⁢r subscript 𝐸 𝑑 𝑢 𝑟 E_{dur}italic_E start_POSTSUBSCRIPT italic_d italic_u italic_r end_POSTSUBSCRIPT and pitch predictor E F 0 subscript 𝐸 subscript 𝐹 0 E_{F_{0}}italic_E start_POSTSUBSCRIPT italic_F start_POSTSUBSCRIPT 0 end_POSTSUBSCRIPT end_POSTSUBSCRIPT to reconstruct the original unit repetitions and pitch contour, respectively. We input all representations into a pretrained vocoder G 𝐺 G italic_G.

Speaker representation.

Our goal is to convert the speaking style while keeping the content fixed. To that end, we construct a speaker representation 𝐳 s⁢p⁢k subscript 𝐳 𝑠 𝑝 𝑘\mathbf{z}{spk}bold_z start_POSTSUBSCRIPT italic_s italic_p italic_k end_POSTSUBSCRIPT, and include it as additional conditioning during the prosody prediction and waveform synthesis phases. To learn 𝐳 s⁢p⁢k subscript 𝐳 𝑠 𝑝 𝑘\mathbf{z}{spk}bold_z start_POSTSUBSCRIPT italic_s italic_p italic_k end_POSTSUBSCRIPT we optimise the parameters of a fixed size look-up-table. Although such modeling limits our ability to generate voices of new and unseen speakers, it produces higher quality generations Polyak et al. (2021).

Prosody representation.

As explained in the content representation section, the number of repetitions of units indicates the rhythm of the speaker in the context. More repetitions means a sound was produced for a longer time. Therefore when converting rhythm the sequence is “deduped”, and when only converting the pitch the full 𝐳 c subscript 𝐳 𝑐\mathbf{z}_{c}bold_z start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT is used.

Following Polyak et al. (2021) we encode the pitch contour of a sample using YAAPT Kasi and Zahorian (2002). We mark this pitch contour 𝒛 F 0=(z F 0 1,…,z F 0 L)subscript 𝒛 subscript 𝐹 0 superscript subscript 𝑧 subscript 𝐹 0 1…superscript subscript 𝑧 subscript 𝐹 0 𝐿\bm{z}{F{0}}=(z_{F_{0}}^{1},\dots,z_{F_{0}}^{L})bold_italic_z start_POSTSUBSCRIPT italic_F start_POSTSUBSCRIPT 0 end_POSTSUBSCRIPT end_POSTSUBSCRIPT = ( italic_z start_POSTSUBSCRIPT italic_F start_POSTSUBSCRIPT 0 end_POSTSUBSCRIPT end_POSTSUBSCRIPT start_POSTSUPERSCRIPT 1 end_POSTSUPERSCRIPT , … , italic_z start_POSTSUBSCRIPT italic_F start_POSTSUBSCRIPT 0 end_POSTSUBSCRIPT end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_L end_POSTSUPERSCRIPT ), where z F 0 t∈ℛ≥0 superscript subscript 𝑧 subscript 𝐹 0 𝑡 subscript ℛ absent 0 z_{F_{0}}^{t}\in\mathcal{R}{\geq 0}italic_z start_POSTSUBSCRIPT italic_F start_POSTSUBSCRIPT 0 end_POSTSUBSCRIPT end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_t end_POSTSUPERSCRIPT ∈ caligraphic_R start_POSTSUBSCRIPT ≥ 0 end_POSTSUBSCRIPT. We mark the per-speaker normalised version as 𝒛F 0 subscript𝒛 subscript 𝐹 0\widetilde{\bm{z}}{F_{0}}over~ start_ARG bold_italic_z end_ARG start_POSTSUBSCRIPT italic_F start_POSTSUBSCRIPT 0 end_POSTSUBSCRIPT end_POSTSUBSCRIPT.

3.2 Speaking Style Conversion

Using the above representations we propose to synthesise the speech signal in the target speaking style. We aim to predict the prosodic features (duration and F0) based on the phonetic-content units, while conditioned on the target speaker representation. We inflate the sequence according to the predicted durations, and use it, the predicted pitch contour, and the speaker vector as a conditioning for the waveform synthesis phase.

Image 3: Refer to caption

Figure 3: An overview of the E F 0 subscript 𝐸 subscript 𝐹 0 E_{F_{0}}italic_E start_POSTSUBSCRIPT italic_F start_POSTSUBSCRIPT 0 end_POSTSUBSCRIPT end_POSTSUBSCRIPT architecture. The positional encoding is a linear positional embedding relative to the start and the end of the sequence, added to the repeated speaker vector. The model jointly predicts the presence of vocalised units (binary classification) together with continuous F0 values for the voiced units.

Rhythm prediction.

Each HuBERT unit correlates with the phoneme in the given 20 20 20 20 ms. This means that longer repetitions of the same unit indicate a longer voicing of the sound - i.e. rhythm. In order to learn to predict, and convert the rhythm we build a model which learns to predict unit durations, denoted as E d⁢u⁢r subscript 𝐸 𝑑 𝑢 𝑟 E_{dur}italic_E start_POSTSUBSCRIPT italic_d italic_u italic_r end_POSTSUBSCRIPT. During training of E d⁢u⁢r subscript 𝐸 𝑑 𝑢 𝑟 E_{dur}italic_E start_POSTSUBSCRIPT italic_d italic_u italic_r end_POSTSUBSCRIPT, we input the deduped phonetic-content units 𝐳 c subscript 𝐳 𝑐\mathbf{z}{c}bold_z start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT and speaker representation conditioning 𝐳 s⁢p⁢k subscript 𝐳 𝑠 𝑝 𝑘\mathbf{z}{spk}bold_z start_POSTSUBSCRIPT italic_s italic_p italic_k end_POSTSUBSCRIPT, with the original unit durations (before dedup) as supervision.

This simple reconstruction task is formulated as a regression task with Mean Squared Error (MSE) loss. As the units do not perfectly correlate to phonemes, neighbouring units often represent the same phoneme even after "dedup". This makes it harder to predict each unit’s length separately. Therefore, we encourage the model’s nearby errors to cancel each other out by adding an MSE loss on the total error of 4 neighbouring units. This helped with errors all being on the negative end and creating a shortening bias in our preliminary test. In addition, we round the regression outputs to be used as integer repetitions of the unit, however this can also cause a bias in which predictions like 1.51,1.51,1.51 1.51 1.51,1.51,1.51 , 1.51 , …, 1.51 1.51 1.51 1.51 and 2.49,2.49,2.49 2.49 2.49,2.49,2.49 , 2.49 , …, 2.49 2.49 2.49 2.49 - would result in the same total length. To mitigate this we carryover the remainder after rounding and sum those differences.

During conversion we replace the speaker vector with that of the target and wish for the model to convert the rhythm. To this end our predictor model must strongly depend on Z s⁢p⁢k subscript 𝑍 𝑠 𝑝 𝑘 Z_{spk}italic_Z start_POSTSUBSCRIPT italic_s italic_p italic_k end_POSTSUBSCRIPT. In order to encourage this we introduce two minor adjustments. HuBERT units hold speaker information and are not fully disentangled, however, the shorter the sequence the less information exists Kharitonov et al. (2022a). We therefore use a CNN with a small receptive field. In addition, we introduce train time masking of units in the sequence, this means that E d⁢u⁢r subscript 𝐸 𝑑 𝑢 𝑟 E_{dur}italic_E start_POSTSUBSCRIPT italic_d italic_u italic_r end_POSTSUBSCRIPT has access to even less units at train time, hence they contain less speaker information. Exact details can be found in Appendix A.3.

Pitch prediction.

We train the pitch predictor to predict the per-speaker normalised pitch contour 𝒛F⁢0 subscript𝒛 𝐹 0\widetilde{\bm{z}}{F0}over~ start_ARG bold_italic_z end_ARG start_POSTSUBSCRIPT italic_F 0 end_POSTSUBSCRIPT, from the HuBERT units 𝐳 c subscript 𝐳 𝑐\mathbf{z}{c}bold_z start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT. The pitch values are normalised to zero mean and standard deviation of one per-speaker on the vocalised sections. We use these values as labels to a predictor conditioned on content units 𝐳 c subscript 𝐳 𝑐\mathbf{z}{c}bold_z start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT, positional encoding and a speaker vector 𝐳 s⁢p⁢k subscript 𝐳 𝑠 𝑝 𝑘\mathbf{z}{spk}bold_z start_POSTSUBSCRIPT italic_s italic_p italic_k end_POSTSUBSCRIPT. The positional encoding is added to 𝐳 s⁢p⁢k subscript 𝐳 𝑠 𝑝 𝑘\mathbf{z}_{spk}bold_z start_POSTSUBSCRIPT italic_s italic_p italic_k end_POSTSUBSCRIPT and denotes the absolute position of the unit compared to the start and finish of the sequence, though preliminary results showed that any standard positional encoding works. Intuitively, positional encoding will help learning localised speaker pitch patterns such as ending sentences in a pitch drop.

Our pitch predictor is denoted E F 0 subscript 𝐸 subscript 𝐹 0 E_{F_{0}}italic_E start_POSTSUBSCRIPT italic_F start_POSTSUBSCRIPT 0 end_POSTSUBSCRIPT end_POSTSUBSCRIPT. An overview of it is described in Figure3. It is a CNN which has two predicted outputs: binary speaking detection, and a regression head to predict the exact pitch. The binary speaking classification is trained with BCE loss on the binarised pitch labels - i.e. 𝒛 F⁢0>0 subscript 𝒛 𝐹 0 0\bm{z}_{F0}>0 bold_italic_z start_POSTSUBSCRIPT italic_F 0 end_POSTSUBSCRIPT > 0. The regression head is trained with MSE loss on the non-zero, vocalised pitch labels. Like the rhythm predictor this model is trained with a small unit receptive field and unit masking.

3.3 Speech Synthesis

We follow Polyak et al. (2021); Kreuk et al. (2021) using a variation of the HiFi-GAN vocoder Kong et al. (2020). Like previous work, we adapt the generator to take as input a sequence of phonetic-content units, pitch contour and a speaker-embedding. During training, these are only features from real samples as the vocoder is trained independently for reconstruction. At inference time we use the unit sequence inflated with predicted durations, the predicted pitch contour and the target speaker-embedding. The above features are concatenated along the temporal axis and fed into a sequence of convolutional layers that output a 1 1 1 1-dimensional audio signal. We match the sample rates of unit sequence and F0 by means of linear interpolation, while replicating the speaker embedding.

To sum up, after each of the components described in previous sections are trained, the full pipeline of DISSC is composed of: first extracting content representation, 𝐳 c subscript 𝐳 𝑐\mathbf{z}{c}bold_z start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT, then, predicting the prosodic features of the target speaker based on 𝐳 c subscript 𝐳 𝑐\mathbf{z}{c}bold_z start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT and 𝐳 s⁢p⁢k subscript 𝐳 𝑠 𝑝 𝑘\mathbf{z}_{spk}bold_z start_POSTSUBSCRIPT italic_s italic_p italic_k end_POSTSUBSCRIPT, and lastly, synthesising the waveform using these new representations.

4 SSC Evaluation Suite

In this section we describe our setup which uses several datasets and fine-grained offline metrics for each part of the conversion - timbre, pitch, rhythm. As no standardised framework for evaluating SSC currently exists, we hope the proposed metrics, will help further advance research in this field. A detailed description of training configurations and hyper-parameters can be found in the Appendix.

4.1 Datasets

Most previous work focus on a single dataset such as VCTK Veaux et al. (2017), since few large, multi-speaker datasets with parallel test sets exist. However, VCTK has fairly monotonous speech with less distinct speaking style.

To address this, we use several datasets. First, we use VCTK as it has many speakers. We leave all paired utterances (speakers say the same content) to the test set, to mimic a truly unpaired setup. In preliminary results, having paired data in training improved the results notably, but harms the wanted setup. We follow Qian et al. (2021) in selecting the two fastest (P 231 231 231 231, P 239 239 239 239) and two slowest (P 245 245 245 245, P 270 270 270 270) speakers by mean log duration.

In addition, we wished to add a more expressive dataset, yet found that most expressive datasets are from the emotional speech domain. In order to adapt them to our use case we take the Emotional Speech Dataset (ESD) Zhou et al. (2022), and select one emotion for each of the 10 10 10 10 speakers. This means that all recordings of a given speaker are in a certain emotion thus making it part of their speaking style. Here, we also take the two fastest and two slowest speakers for rhythm metric evaluation.

After preliminary experiments we found that the speakers in the above datasets do not always have clear pitch patterns. This means that the speech is quite monotonous and that variation across recordings was greater than across speakers. To effectively evaluate pitch conversion we created a synthetic dataset based on VCTK - denoted as Syn_VCTK. In this dataset we took 6 6 6 6 speakers from VCTK, encoded them based on the pipeline by Polyak et al. (2021). We then altered the pitch contour by introducing a linear trend of either up, down or flat for each of the speakers. We then generated the new synthetic recordings in the same approach. We encourage readers to listen to these samples as well as the other datasets in the accompanying page to grasp the different speaking styles.

Table 1: A comparison of the proposed method against VC and SSC baselines considering content (WER, CER), Rhythm and F0 (EMD, TLE, WLE, PLE), and speaker identification (EER). Only DISSC improves rhythm metrics compared to standard VC baseline Speech Resynthesis.

4.2 Metrics

4.2.1 Objective Metrics

We introduce new metrics which aim to capture the rhythm in a fine-grained manner, and objectively evaluate the pitch contour (even when the rhythm does not match). We hope these metrics will contribute to other advances in this field.

Timbre.

For regular VC we use the common Equal Error Rate (EER) of speaker verification using SpeechBrain’s Ravanelli et al. (2021) ECAPA-TDNN Desplanques et al. (2020) which achieves 0.8 0.8 0.8 0.8% EER on Voxceleb. For each sample, we randomly select n 𝑛 n italic_n positive and n 𝑛 n italic_n negative samples. Positive samples are random utterances of other speakers converted to the current speaker, and negative samples are random utterances of the current speaker converted to a random target speaker.

Content.

For phonetic content preservation we use the standard Word Error Rate (WER) and Character Error Rate (CER), using Whisper Radford et al. (2022) as our Automatic Speech Recognition (ASR) model.

Rhythm.

In order to evaluate the rhythm we wish to go beyond the length of the recording, which only indicates the average rhythm and not which sounds are uttered slowly and which quickly. We therefore use the text transcriptions of the samples, and the Montreal Forced Aligner (MFA) McAuliffe et al. (2017) to get start and end times for each word and phoneme in the recording. We then compare the lengths of the converted speech and the target speech, thus defining Phoneme Length Error (PLE), Word Length Error (WLE) and Total Length Error (TLE). We do not compare the length of silences as their occurrence can differ between recordings thus breaking the alignment.

Formally, for each input recording and matching text, MFA returns a list of tuples T w=((s w 1,e w 1,c w 1)⁢…⁢(s w m,e w m,c w m))subscript 𝑇 𝑤 subscript 𝑠 subscript 𝑤 1 subscript 𝑒 subscript 𝑤 1 subscript 𝑐 subscript 𝑤 1…subscript 𝑠 subscript 𝑤 𝑚 subscript 𝑒 subscript 𝑤 𝑚 subscript 𝑐 subscript 𝑤 𝑚 T_{w}=((s_{w_{1}},e_{w_{1}},c_{w_{1}})\dots(s_{w_{m}},e_{w_{m}},c_{w_{m}}))italic_T start_POSTSUBSCRIPT italic_w end_POSTSUBSCRIPT = ( ( italic_s start_POSTSUBSCRIPT italic_w start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT end_POSTSUBSCRIPT , italic_e start_POSTSUBSCRIPT italic_w start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT end_POSTSUBSCRIPT , italic_c start_POSTSUBSCRIPT italic_w start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT end_POSTSUBSCRIPT ) … ( italic_s start_POSTSUBSCRIPT italic_w start_POSTSUBSCRIPT italic_m end_POSTSUBSCRIPT end_POSTSUBSCRIPT , italic_e start_POSTSUBSCRIPT italic_w start_POSTSUBSCRIPT italic_m end_POSTSUBSCRIPT end_POSTSUBSCRIPT , italic_c start_POSTSUBSCRIPT italic_w start_POSTSUBSCRIPT italic_m end_POSTSUBSCRIPT end_POSTSUBSCRIPT ) ) for word level and another for phoneme level - T p=((s p 1,e p 1,c p 1)⁢…⁢(s p m,e p m,c p m))subscript 𝑇 𝑝 subscript 𝑠 subscript 𝑝 1 subscript 𝑒 subscript 𝑝 1 subscript 𝑐 subscript 𝑝 1…subscript 𝑠 subscript 𝑝 𝑚 subscript 𝑒 subscript 𝑝 𝑚 subscript 𝑐 subscript 𝑝 𝑚 T_{p}=((s_{p_{1}},e_{p_{1}},c_{p_{1}})\dots(s_{p_{m}},e_{p_{m}},c_{p_{m}}))italic_T start_POSTSUBSCRIPT italic_p end_POSTSUBSCRIPT = ( ( italic_s start_POSTSUBSCRIPT italic_p start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT end_POSTSUBSCRIPT , italic_e start_POSTSUBSCRIPT italic_p start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT end_POSTSUBSCRIPT , italic_c start_POSTSUBSCRIPT italic_p start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT end_POSTSUBSCRIPT ) … ( italic_s start_POSTSUBSCRIPT italic_p start_POSTSUBSCRIPT italic_m end_POSTSUBSCRIPT end_POSTSUBSCRIPT , italic_e start_POSTSUBSCRIPT italic_p start_POSTSUBSCRIPT italic_m end_POSTSUBSCRIPT end_POSTSUBSCRIPT , italic_c start_POSTSUBSCRIPT italic_p start_POSTSUBSCRIPT italic_m end_POSTSUBSCRIPT end_POSTSUBSCRIPT ) ). s{w,p}i subscript 𝑠 subscript 𝑤 𝑝 𝑖 s_{{w,p}{i}}italic_s start_POSTSUBSCRIPT { italic_w , italic_p } start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT end_POSTSUBSCRIPT indicates the start time of the sound at index i 𝑖 i italic_i in the recording (word or phoneme according to the level), and e{w,p}i subscript 𝑒 subscript 𝑤 𝑝 𝑖 e{{w,p}{i}}italic_e start_POSTSUBSCRIPT { italic_w , italic_p } start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT end_POSTSUBSCRIPT indicates the end time. c{w,p}i subscript 𝑐 subscript 𝑤 𝑝 𝑖 c{{w,p}{i}}italic_c start_POSTSUBSCRIPT { italic_w , italic_p } start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT end_POSTSUBSCRIPT represents the content of the time segment, i.e the word or phoneme. We filter only non-silence time segments leaving T~{w,p}=((s{w,p}i,e{w,p}i,c{w,p}i)∀i s.t.c i{w,p}≠s i l e n c e)\widetilde{T}{{w,p}}=((s_{{w,p}{i}},e{{w,p}{i}},c{{w,p}{i}})\ % \forall i\ s.t.\ c^{{w,p}}{i}\neq silence)over~ start_ARG italic_T end_ARG start_POSTSUBSCRIPT { italic_w , italic_p } end_POSTSUBSCRIPT = ( ( italic_s start_POSTSUBSCRIPT { italic_w , italic_p } start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT end_POSTSUBSCRIPT , italic_e start_POSTSUBSCRIPT { italic_w , italic_p } start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT end_POSTSUBSCRIPT , italic_c start_POSTSUBSCRIPT { italic_w , italic_p } start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT end_POSTSUBSCRIPT ) ∀ italic_i italic_s . italic_t . italic_c start_POSTSUPERSCRIPT { italic_w , italic_p } end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ≠ italic_s italic_i italic_l italic_e italic_n italic_c italic_e ). Thus formally, the errors per sample compare durations of a reference sample and a synthesised sample:

PLE(Tp s⁢y⁢n,Tp r⁢e⁢f)=∑j=1 l⁢e⁢n⁢(Tp)|(e p j s⁢y⁢n−s p j s⁢y⁢n)−(e p j r⁢e⁢f−s p j r⁢e⁢f)|,WLE(Tw s⁢y⁢n,Tw r⁢e⁢f)=∑j=1 l⁢e⁢n⁢(Tw)|(e w j s⁢y⁢n−s w j s⁢y⁢n)−(e w j r⁢e⁢f−s w j r⁢e⁢f)|.formulae-sequence PLE superscript subscript𝑇 𝑝 𝑠 𝑦 𝑛 superscript subscript𝑇 𝑝 𝑟 𝑒 𝑓 superscript subscript 𝑗 1 𝑙 𝑒 𝑛 subscript𝑇 𝑝 subscript superscript 𝑒 𝑠 𝑦 𝑛 subscript 𝑝 𝑗 subscript superscript 𝑠 𝑠 𝑦 𝑛 subscript 𝑝 𝑗 subscript superscript 𝑒 𝑟 𝑒 𝑓 subscript 𝑝 𝑗 subscript superscript 𝑠 𝑟 𝑒 𝑓 subscript 𝑝 𝑗 WLE superscript subscript𝑇 𝑤 𝑠 𝑦 𝑛 superscript subscript𝑇 𝑤 𝑟 𝑒 𝑓 superscript subscript 𝑗 1 𝑙 𝑒 𝑛 subscript𝑇 𝑤 subscript superscript 𝑒 𝑠 𝑦 𝑛 subscript 𝑤 𝑗 subscript superscript 𝑠 𝑠 𝑦 𝑛 subscript 𝑤 𝑗 subscript superscript 𝑒 𝑟 𝑒 𝑓 subscript 𝑤 𝑗 subscript superscript 𝑠 𝑟 𝑒 𝑓 subscript 𝑤 𝑗\begin{split}\text{PLE}(&\widetilde{T}{p}^{syn},\widetilde{T}{p}^{ref})=\ &\sum_{j=1}^{len(\widetilde{T}{p})}|(e^{syn}{p_{j}}-s^{syn}{p{j}})-(e^{ref% }{p{j}}-s^{ref}{p{j}})|,\ \text{WLE}(&\widetilde{T}{w}^{syn},\widetilde{T}{w}^{ref})=\ &\sum_{j=1}^{len(\widetilde{T}{w})}|(e^{syn}{w_{j}}-s^{syn}{w{j}})-(e^{ref% }{w{j}}-s^{ref}{w{j}})|.\end{split}start_ROW start_CELL PLE ( end_CELL start_CELL over~ start_ARG italic_T end_ARG start_POSTSUBSCRIPT italic_p end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_s italic_y italic_n end_POSTSUPERSCRIPT , over~ start_ARG italic_T end_ARG start_POSTSUBSCRIPT italic_p end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_r italic_e italic_f end_POSTSUPERSCRIPT ) = end_CELL end_ROW start_ROW start_CELL end_CELL start_CELL ∑ start_POSTSUBSCRIPT italic_j = 1 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l italic_e italic_n ( over~ start_ARG italic_T end_ARG start_POSTSUBSCRIPT italic_p end_POSTSUBSCRIPT ) end_POSTSUPERSCRIPT | ( italic_e start_POSTSUPERSCRIPT italic_s italic_y italic_n end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_p start_POSTSUBSCRIPT italic_j end_POSTSUBSCRIPT end_POSTSUBSCRIPT - italic_s start_POSTSUPERSCRIPT italic_s italic_y italic_n end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_p start_POSTSUBSCRIPT italic_j end_POSTSUBSCRIPT end_POSTSUBSCRIPT ) - ( italic_e start_POSTSUPERSCRIPT italic_r italic_e italic_f end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_p start_POSTSUBSCRIPT italic_j end_POSTSUBSCRIPT end_POSTSUBSCRIPT - italic_s start_POSTSUPERSCRIPT italic_r italic_e italic_f end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_p start_POSTSUBSCRIPT italic_j end_POSTSUBSCRIPT end_POSTSUBSCRIPT ) | , end_CELL end_ROW start_ROW start_CELL WLE ( end_CELL start_CELL over~ start_ARG italic_T end_ARG start_POSTSUBSCRIPT italic_w end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_s italic_y italic_n end_POSTSUPERSCRIPT , over~ start_ARG italic_T end_ARG start_POSTSUBSCRIPT italic_w end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_r italic_e italic_f end_POSTSUPERSCRIPT ) = end_CELL end_ROW start_ROW start_CELL end_CELL start_CELL ∑ start_POSTSUBSCRIPT italic_j = 1 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_l italic_e italic_n ( over~ start_ARG italic_T end_ARG start_POSTSUBSCRIPT italic_w end_POSTSUBSCRIPT ) end_POSTSUPERSCRIPT | ( italic_e start_POSTSUPERSCRIPT italic_s italic_y italic_n end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_w start_POSTSUBSCRIPT italic_j end_POSTSUBSCRIPT end_POSTSUBSCRIPT - italic_s start_POSTSUPERSCRIPT italic_s italic_y italic_n end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_w start_POSTSUBSCRIPT italic_j end_POSTSUBSCRIPT end_POSTSUBSCRIPT ) - ( italic_e start_POSTSUPERSCRIPT italic_r italic_e italic_f end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_w start_POSTSUBSCRIPT italic_j end_POSTSUBSCRIPT end_POSTSUBSCRIPT - italic_s start_POSTSUPERSCRIPT italic_r italic_e italic_f end_POSTSUPERSCRIPT start_POSTSUBSCRIPT italic_w start_POSTSUBSCRIPT italic_j end_POSTSUBSCRIPT end_POSTSUBSCRIPT ) | . end_CELL end_ROW(1)

The TLE is just the duration difference between the entire recordings. All rhythm metrics are in time units, in this case seconds. In addition, on rare occasions the number of phonemes for the same sound by different speakers is different (due to accents etc.), we ignore these samples for a more well-defined metric. In addition, when the phonetic content is harmed dramatically in the conversion, MFA might fail to align the text. We penalise such samples by assuming a linear and even split of the recording to phonemes and words.

Pitch.

Popular evaluation methods for F0 similarity include the Voicing Decision Error (VDE) and F0 frame error (FFE). Intuitively, VDE measures the portion of frames with voicing decision error, while FFE measures the percentage of frames that contain a deviation of more than 20% in pitch value or have a voicing decision error. Formally,

VDE⁢(𝒗,𝒗^)=∑t=1 T−1 𝟙⁢[𝒗 t≠𝒗^t]T,VDE 𝒗^𝒗 superscript subscript 𝑡 1 𝑇 1 1 delimited-[]subscript 𝒗 𝑡 subscript^𝒗 𝑡 𝑇\text{VDE}(\bm{v},\hat{\bm{v}})=\frac{\sum_{t=1}^{T-1}\mathbbm{1}[\bm{v}{t}% \neq\hat{\bm{v}}{t}]}{T},VDE ( bold_italic_v , over^ start_ARG bold_italic_v end_ARG ) = divide start_ARG ∑ start_POSTSUBSCRIPT italic_t = 1 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_T - 1 end_POSTSUPERSCRIPT blackboard_1 [ bold_italic_v start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT ≠ over^ start_ARG bold_italic_v end_ARG start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT ] end_ARG start_ARG italic_T end_ARG ,(2)

FFE(𝒑,𝒑^,𝒗,𝒗^)=VDE⁢(𝒗,𝒗^)+∑t=1 T−1 𝟙⁢[|𝒑 t−𝒑^t|>0.2⋅𝒑 t]⁢𝟙⁢[𝒗 t]⁢𝟙⁢[𝒗^t]T,FFE 𝒑^𝒑 𝒗^𝒗 VDE 𝒗^𝒗 superscript subscript 𝑡 1 𝑇 1 1 delimited-[]subscript 𝒑 𝑡 subscript^𝒑 𝑡⋅0.2 subscript 𝒑 𝑡 1 delimited-[]subscript 𝒗 𝑡 1 delimited-[]subscript^𝒗 𝑡 𝑇\begin{split}\text{FFE}&(\bm{p},\hat{\bm{p}},\bm{v},\hat{\bm{v}})=\text{VDE}(% \bm{v},\hat{\bm{v}})\ +&\frac{\sum_{t=1}^{T-1}\mathbbm{1}[|\bm{p}{t}-\hat{\bm{p}}{t}|>0.2\cdot\bm{% p}{t}]\mathbbm{1}[\bm{v}{t}]\mathbbm{1}[\hat{\bm{v}}_{t}]}{T},\end{split}start_ROW start_CELL FFE end_CELL start_CELL ( bold_italic_p , over^ start_ARG bold_italic_p end_ARG , bold_italic_v , over^ start_ARG bold_italic_v end_ARG ) = VDE ( bold_italic_v , over^ start_ARG bold_italic_v end_ARG ) end_CELL end_ROW start_ROW start_CELL + end_CELL start_CELL divide start_ARG ∑ start_POSTSUBSCRIPT italic_t = 1 end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_T - 1 end_POSTSUPERSCRIPT blackboard_1 [ | bold_italic_p start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT - over^ start_ARG bold_italic_p end_ARG start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT | > 0.2 ⋅ bold_italic_p start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT ] blackboard_1 [ bold_italic_v start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT ] blackboard_1 [ over^ start_ARG bold_italic_v end_ARG start_POSTSUBSCRIPT italic_t end_POSTSUBSCRIPT ] end_ARG start_ARG italic_T end_ARG , end_CELL end_ROW(3)

where 𝒑,𝒑^𝒑^𝒑\bm{p},\hat{\bm{p}}bold_italic_p , over^ start_ARG bold_italic_p end_ARG are the pitch frames from the target and generated signals, 𝒗,𝒗^𝒗^𝒗\bm{v},\hat{\bm{v}}bold_italic_v , over^ start_ARG bold_italic_v end_ARG are the respective voicing decisions, and 𝟙 1\mathbbm{1}blackboard_1 is the indicator function.

Although these metrics are widely used in speech synthesis Wang et al. (2021), these methods are not well adapted to misaligned samples, i.e when they are spoken at different rates. These metrics require the samples to have the same time length, therefore many implementations linearly interpolate the generated sample’s pitch contour to that of the target before computing them. However, rhythm differences are non-linear (i.e certain phonemes could be “stretched” more than others) thus not alleviating the misalignment problem. Instead, we suggest using segments obtained from the forced-aligner and calculate the FFE while interpolating the segments to match the segment sizes. This measures the match in shape and value of the pitch contour for each word and phoneme, even if they are misaligned. We denoted these as P_FFE for phoneme segments and W_FFE for words.

Rhythm & Pitch.

We compute Earth Movers Distance (EMD) Rubner et al. (1998) between the pitch contour of converted and target samples. Essentially we check how much “pitch mass” needs to be moved between them. This metric considers errors along the time domain (rhythm) and pitch contour errors (such as predicting the wrong trend). Our results show strong correlation to human perception of rhythm and pitch with the above metrics.

This newly proposed evaluation suite is generic and can transfer to other languages. However, certain metrics are based on language-specific pre-trained models (e.g., ASR-based metrics). In such cases, a multi-lingual ASR can be used. In our setup, we use the Whisper model Radford et al. (2022) which supports ∼100 similar-to absent 100\sim 100∼ 100 languages. Another option would be to use the MMS model Pratap et al. (2023) which supports more than 1000 1000 1000 1000 languages for both ASR and forced-alignment.

4.2.2 Human evaluation

To confirm that the proposed objective metrics correlate to human perception, and to evaluate to what extent the prosody impacts humans’ speaker recognition we conduct a human evaluation study. We use the common mean opinion score (MOS) to evaluate the quality and naturalness of generated and real samples. Specifically, each sample was scored from 1 1 1 1-5 5 5 5 on naturalness and quality. We took the same 40 40 40 40 samples for each method and the original. Each sample was annotated by at least three raters.

Next, to evaluate the effectiveness of the speaking style conversion, we present the users a sample from the target speaker and a sample converted by each method. Users rated each sample from 1 1 1 1-5 5 5 5 (“Very Different”-“Very Similar”) on how likely it is to be uttered by the same speaker and instructed to “pay attention to speaking style such as rhythm and pitch changes, and not only the voice texture". We evaluated 40 40 40 40 samples per-method, with at least three raters per-sample. We report the mean score and 95 95 95 95% confidence intervals for both tests. We modified the WebMUSHRA 1 1 1https://github.com/audiolabs/webMUSHRA tool for the tests. A detailed description of the subjective tests can be found in the Appendix.

Image 4: Refer to caption

Figure 4: We compare the human evaluated naturalness of real reference utterances, Speech Resynthesis (SR), AutoPST, two versions of DISSC which convert rhythm only or also F0. We also compare the similarity of how likely the converted sample is to come from the the same speaker as the target recording.

Table 2: VC and SSC results on the Syn_VCTK dataset. This ablation measures the impact of rhythm and pitch conversion over SR considering content (WER, CER), Rhythm and F0 (EMD, TLE, WLE, PLE, W_FFE, P_FFE).

5 Results

We compare DISSC to several any-to-many (or many-to-many) baselines. Notably, we evaluate Speech Resynthesis (SR), and AutoVC Qian et al. (2019) which are well known VC methods. In addition, we compare to AutoPST and BNE-MoL-seq2seqVC Liu et al. (2021) which are prominent prosodic-aware VC (i.e., SSC) methods. We use the official Github implementations and train the models on our datasets using the authors recommended configuration. For a fair comparison, all baselines were evaluated under the any-to-many setup, i.e., generating only seen speakers. Only seq2seqVC supports new speakers, because only that approach was published. However, we evaluate them on known speakers, with the entire training set as reference recordings. We show three variants of our method: DISSC _Rhythm, DISSC _Pitch and DISSC _Both which predict duration, pitch contour and both respectively.

Objective results for VCTK and ESD appear in Table 1. On both datasets DISSC improves the length errors across all scales: phonemes, words and utterances. The relative improvement compared to the next best baseline is 56.8 56.8 56.8 56.8% on ESD TLE. At word level we get 27.6 27.6 27.6 27.6% and 26.7 26.7 26.7 26.7% relative improvement on ESD and VCTK respectively. All other baselines show rhythm errors comparable or worse than SR which does not convert rhythm.

We can see a minor decrease in content quality (by WER and CER) of DISSC compared to SR which does not convert prosody, but still superior to AutoPST. In addition, seq2seq also has better WER, at the expense of harming the rhythm conversion. Potentially, the supervised phoneme recogniser is too fine, thus does not impose enough of a rhythm bottleneck. The harm in quality of DISSC may be due to additional information bottlenecks by de-duplicating the speech units, and predicting F0.

We additionally provide subjective evaluation considering both naturalness of the converted samples and their similarity to the target recording considering speaking style. These results are shown in Figure4. We found that the naturalness of both DISSC variants was comparable to SR, and not far from the reference recordings while far outperforming AutoPST. In addition, we see a noticeable increase in speaking style similarity when converting prosody with DISSC compared to SR. However, also converting the pitch did not noticeably improve the results. We believe that this has to do with the lack of distinct speaking style which people can notice, especially from a single recording.

Ablation study.

In order to properly evaluate the ability of DISSC to learn the pitch speaking style, when such exists, we use the synthetic VCTK data (as explained in Section 4.1) with known pitch trends. These results (Table 2) are an ablation study for each converter of DISSC. They highlight the control-ability of our approach - when predicting the rhythm, only length errors improve, likewise for pitch. We see that when a pitch pattern exists in the speaking style - DISSC learns it, providing 88 88 88 88-113 113 113 113% relative increase in the metrics. We also see that they can be converted jointly without noticeable drop in pitch and length performance.

Irregular rhythm.

Lastly, to demonstrate that DISSC predicts rhythm depending on content, and not only on average - we create a sample with abnormal rhythm patterns. We take an utterance and artificially stretch one of the words using speech resynthesis. We then attempt to “convert” the sample to the original speaker thus checking the ability to only change the rhythm of the abnormal part. We compare DISSC and AutoPST in Figure5, and see that DISSC manages to mostly correct the irregular rhythm, while hardly changing the other parts. By contrast, AutoPST only slightly changes the rhythm. If E c subscript 𝐸 𝑐 E_{c}italic_E start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT performs optimally, we are guaranteed perfect results, as z c subscript 𝑧 𝑐 z_{c}italic_z start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT would be identical for irregular and original after dedup. However, HuBERT may introduce some errors specifically when considering signal variations Gat et al. (2022).

Image 5: Refer to caption

Figure 5: DISSC corrects abnormal rhythm patterns in specific words, compared to AutoPST. Here we take the utterance “people look, but no one ever finds it” and stretch the duration of the word “finds” times three.

6 Conclusion & Future Work

In this work, we propose a simple and effective method for speaking style conversion based on discrete self supervised speech representations. We provide a formal definition to the speaking style conversion setup while introducing a set of evaluation functions to analyse and evaluate different speech characteristics separately. Results suggest, the proposed approach is superior to the evaluated baselines considering both objective and subjective metrics. We hope these advance the study in the field, and lead to improved performance.

The current model suffers from slight content loss when using the deduped units. In addition, repeated utterances get slightly different units (in the abnormal rhythm experiment). Therefore, in future work we aim to improve the robustness and disentanglement of the speech representation considering speaker and prosodic information.

Limitations

Our approach can only perform conversion into a closed set of speakers seen at train time. This is due to the LUT used to learn speaker representations and to the use of per-speaker pitch statistics. This might be hard to overcome as speaker style (rhythm and pitch changes) can be hard to judge from a small number of samples. Additionally, we see that speaking style conversion and especially rhythm conversion harm the WER and CER slightly. The additional bottleneck of de-duplication which allows us to convert rhythm comes with the cost of slight content loss. We hope that future improvement of the hidden representations will help improve or alleviate the tradeoff.

Ethical Statement

The broader impact of this study is, as in any generative model, the development of a high quality and natural speech synthesis model. This has special sensitivity and ethical concerns as such technology might be used to alter voices and speaking style in speech recordings, which can be considered as biometric data and biometric processing. To deal with that we limit the number of speakers that can be synthesized with the proposed approach using a pre-defined voices using a look-up-table. Another potential risk of the proposed method which is also one of its limitations is that the generated speech content is not always perfect, hence might lead to wrong pronunciations.

References

  • Borsos et al. (2022) Zalán Borsos, Raphaël Marinier, Damien Vincent, Eugene Kharitonov, Olivier Pietquin, Matt Sharifi, Olivier Teboul, David Grangier, Marco Tagliasacchi, and Neil Zeghidour. 2022. Audiolm: a language modeling approach to audio generation. arXiv preprint arXiv:2209.03143.
  • Chen et al. (2022) Li-Wei Chen, Shinji Watanabe, and Alexander Rudnicky. 2022. A unified one-shot prosody and speaker conversion system with self-supervised discrete speech units. arXiv preprint arXiv:2211.06535.
  • Chen and Duan (2022) Meiying Chen and Zhiyao Duan. 2022. Controlvc: Zero-shot voice conversion with time-varying controls on pitch and rhythm. arXiv preprint arXiv:2209.11866.
  • Chen et al. (2021) Yen-Hao Chen, Da-Yi Wu, Tsung-Han Wu, and Hung-yi Lee. 2021. Again-vc: A one-shot voice conversion using activation guidance and adaptive instance normalization. In ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 5954–5958. IEEE.
  • Chou et al. (2019) Ju-chieh Chou, Cheng-chieh Yeh, and Hung-yi Lee. 2019. One-shot voice conversion by separating speaker and content representations with instance normalization. arXiv preprint arXiv:1904.05742.
  • Desplanques et al. (2020) Brecht Desplanques, Jenthe Thienpondt, and Kris Demuynck. 2020. ECAPA-TDNN: emphasized channel attention, propagation and aggregation in TDNN based speaker verification. In Interspeech 2020, pages 3830–3834. ISCA.
  • Gat et al. (2022) Itai Gat, Felix Kreuk, Ann Lee, Jade Copet, Gabriel Synnaeve, Emmanuel Dupoux, and Yossi Adi. 2022. On the robustness of self-supervised representations for spoken language modeling. arXiv preprint arXiv:2209.15483.
  • Hsu et al. (2021) Wei-Ning Hsu, Benjamin Bolte, Yao-Hung Hubert Tsai, Kushal Lakhotia, Ruslan Salakhutdinov, and Abdelrahman Mohamed. 2021. Hubert: Self-supervised speech representation learning by masked prediction of hidden units. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 29:3451–3460.
  • Huang et al. (2021) Wen-Chin Huang, Yi-Chiao Wu, and Tomoki Hayashi. 2021. Any-to-one sequence-to-sequence voice conversion using self-supervised discrete speech representations. In ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 5944–5948. IEEE.
  • Huang et al. (2022) Wen-Chin Huang, Shu-Wen Yang, Tomoki Hayashi, Hung-Yi Lee, Shinji Watanabe, and Tomoki Toda. 2022. S3prl-vc: Open-source voice conversion framework with self-supervised speech representations. In ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 6552–6556. IEEE.
  • Kain and Macon (1998) Alexander Kain and Michael W Macon. 1998. Spectral voice conversion for text-to-speech synthesis. In Proceedings of the 1998 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP’98 (Cat. No. 98CH36181), volume 1, pages 285–288. IEEE.
  • Kameoka et al. (2018) Hirokazu Kameoka, Takuhiro Kaneko, Kou Tanaka, and Nobukatsu Hojo. 2018. Stargan-vc: Non-parallel many-to-many voice conversion using star generative adversarial networks. In 2018 IEEE Spoken Language Technology Workshop (SLT), pages 266–273. IEEE.
  • Kaneko and Kameoka (2017) Takuhiro Kaneko and Hirokazu Kameoka. 2017. Parallel-data-free voice conversion using cycle-consistent adversarial networks. arXiv preprint arXiv:1711.11293.
  • Kasi and Zahorian (2002) Kavita Kasi and Stephen A Zahorian. 2002. Yet another algorithm for pitch tracking. In 2002 ieee international conference on acoustics, speech, and signal processing, volume 1, pages I–361. IEEE.
  • Kharitonov et al. (2022a) Eugene Kharitonov, Jade Copet, Kushal Lakhotia, Tu Anh Nguyen, Paden Tomasello, Ann Lee, Ali Elkahky, Wei-Ning Hsu, Abdelrahman Mohamed, Emmanuel Dupoux, and Yossi Adi. 2022a. textless-lib: a library for textless spoken language processing. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies: System Demonstrations, pages 1–9, Hybrid: Seattle, Washington + Online. Association for Computational Linguistics.
  • Kharitonov et al. (2022b) Eugene Kharitonov, Ann Lee, Adam Polyak, Yossi Adi, Jade Copet, Kushal Lakhotia, Tu Anh Nguyen, Morgane Riviere, Abdelrahman Mohamed, Emmanuel Dupoux, and Wei-Ning Hsu. 2022b. Text-free prosody-aware generative spoken language modeling. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 8666–8681, Dublin, Ireland. Association for Computational Linguistics.
  • Kim et al. (2022) Kang-wook Kim, Seung-won Park, Junhyeok Lee, and Myun-chul Joe. 2022. Assem-vc: Realistic voice conversion by assembling modern speech synthesis techniques. In ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 6997–7001. IEEE.
  • Kong et al. (2020) Jungil Kong, Jaehyeon Kim, and Jaekyoung Bae. 2020. Hifi-gan: Generative adversarial networks for efficient and high fidelity speech synthesis. Advances in Neural Information Processing Systems, 33:17022–17033.
  • Kreuk et al. (2021) Felix Kreuk, Adam Polyak, Jade Copet, Eugene Kharitonov, Tu-Anh Nguyen, Morgane Rivière, Wei-Ning Hsu, Abdelrahman Mohamed, Emmanuel Dupoux, and Yossi Adi. 2021. Textless speech emotion conversion using decomposed and discrete representations. arXiv preprint arXiv:2111.07402.
  • Kreuk et al. (2022a) Felix Kreuk, Gabriel Synnaeve, Adam Polyak, Uriel Singer, Alexandre Défossez, Jade Copet, Devi Parikh, Yaniv Taigman, and Yossi Adi. 2022a. Audiogen: Textually guided audio generation. arXiv preprint arXiv:2209.15352.
  • Kreuk et al. (2022b) Felix Kreuk, Yaniv Taigman, Adam Polyak, Jade Copet, Gabriel Synnaeve, Alexandre Défossez, and Yossi Adi. 2022b. Audio language modeling using perceptually-guided discrete representations. arXiv preprint arXiv:2211.01223.
  • Kuhlmann et al. (2022) Michael Kuhlmann, Fritz Seebauer, Janek Ebbers, Petra Wagner, and Reinhold Haeb-Umbach. 2022. Investigation into target speaking rate adaptation for voice conversion. arXiv preprint arXiv:2209.01978.
  • Lakhotia et al. (2021) Kushal Lakhotia, Eugene Kharitonov, Wei-Ning Hsu, Yossi Adi, Adam Polyak, Benjamin Bolte, Tu-Anh Nguyen, Jade Copet, Alexei Baevski, Abdelrahman Mohamed, et al. 2021. On generative spoken language modeling from raw audio. Transactions of the Association for Computational Linguistics, 9:1336–1354.
  • Lee et al. (2022a) Ann Lee, Peng-Jen Chen, Changhan Wang, Jiatao Gu, Sravya Popuri, Xutai Ma, Adam Polyak, Yossi Adi, Qing He, Yun Tang, Juan Pino, and Wei-Ning Hsu. 2022a. Direct speech-to-speech translation with discrete units. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 3327–3339, Dublin, Ireland. Association for Computational Linguistics.
  • Lee et al. (2022b) Ann Lee, Hongyu Gong, Paul-Ambroise Duquenne, Holger Schwenk, Peng-Jen Chen, Changhan Wang, Sravya Popuri, Yossi Adi, Juan Pino, Jiatao Gu, and Wei-Ning Hsu. 2022b. Textless speech-to-speech translation on real data. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 860–872, Seattle, United States. Association for Computational Linguistics.
  • Lee et al. (2021) Sang-Hoon Lee, Ji-Hoon Kim, Hyunseung Chung, and Seong-Whan Lee. 2021. Voicemixer: Adversarial voice style mixup. In Advances in Neural Information Processing Systems, volume 34, pages 294–308. Curran Associates, Inc.
  • Lee et al. (2022c) Sang-Hoon Lee, Hyeong-Rae Noh, Woo-Jeoung Nam, and Seong-Whan Lee. 2022c. Duration controllable voice conversion via phoneme-based information bottleneck. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 30:1173–1183.
  • Lin et al. (2021a) Jheng-hao Lin, Yist Y Lin, Chung-Ming Chien, and Hung-yi Lee. 2021a. S2vc: a framework for any-to-any voice conversion with self-supervised pretrained representations. arXiv preprint arXiv:2104.02901.
  • Lin et al. (2021b) Yist Y Lin, Chung-Ming Chien, Jheng-Hao Lin, Hung-yi Lee, and Lin-shan Lee. 2021b. Fragmentvc: Any-to-any voice conversion by end-to-end extracting and fusing fine-grained voice fragments with attention. In ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 5939–5943. IEEE.
  • Liu et al. (2021) Songxiang Liu, Yuewen Cao, Disong Wang, Xixin Wu, Xunying Liu, and Helen Meng. 2021. Any-to-many voice conversion with location-relative sequence-to-sequence modeling. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 29:1717–1728.
  • McAuliffe et al. (2017) Michael McAuliffe, Michaela Socolof, Sarah Mihuc, Michael Wagner, and Morgan Sonderegger. 2017. Montreal Forced Aligner: Trainable Text-Speech Alignment Using Kaldi. In Proc. Interspeech 2017, pages 498–502.
  • Nakashika et al. (2013) Toru Nakashika, Ryoichi Takashima, Tetsuya Takiguchi, and Yasuo Ariki. 2013. Voice conversion in high-order eigen space using deep belief nets. In Interspeech, pages 369–372.
  • Nguyen et al. (2022) Tu Anh Nguyen, Eugene Kharitonov, Jade Copet, Yossi Adi, Wei-Ning Hsu, Ali Elkahky, Paden Tomasello, Robin Algayres, Benoit Sagot, Abdelrahman Mohamed, et al. 2022. Generative spoken dialogue language modeling. arXiv preprint arXiv:2203.16502.
  • Panayotov et al. (2015) Vassil Panayotov, Guoguo Chen, Daniel Povey, and Sanjeev Khudanpur. 2015. Librispeech: an asr corpus based on public domain audio books. In 2015 IEEE international conference on acoustics, speech and signal processing (ICASSP), pages 5206–5210. IEEE.
  • Polyak et al. (2021) Adam Polyak, Yossi Adi, Jade Copet, Eugene Kharitonov, Kushal Lakhotia, Wei-Ning Hsu, Abdelrahman Mohamed, and Emmanuel Dupoux. 2021. Speech resynthesis from discrete disentangled self-supervised representations. arXiv preprint arXiv:2104.00355.
  • Popuri et al. (2022) Sravya Popuri, Peng-Jen Chen, Changhan Wang, Juan Pino, Yossi Adi, Jiatao Gu, Wei-Ning Hsu, and Ann Lee. 2022. Enhanced direct speech-to-speech translation using self-supervised pre-training and data augmentation. arXiv preprint arXiv:2204.02967.
  • Pratap et al. (2023) Vineel Pratap, Andros Tjandra, Bowen Shi, Paden Tomasello, Arun Babu, Sayani Kundu, Ali Elkahky, Zhaoheng Ni, Apoorv Vyas, Maryam Fazel-Zarandi, et al. 2023. Scaling speech technology to 1,000+ languages. arXiv preprint arXiv:2305.13516.
  • Qian et al. (2020) Kaizhi Qian, Yang Zhang, Shiyu Chang, Mark Hasegawa-Johnson, and David Cox. 2020. Unsupervised speech decomposition via triple information bottleneck. In International Conference on Machine Learning, pages 7836–7846. PMLR.
  • Qian et al. (2021) Kaizhi Qian, Yang Zhang, Shiyu Chang, Jinjun Xiong, Chuang Gan, David Cox, and Mark Hasegawa-Johnson. 2021. Global prosody style transfer without text transcriptions. In International Conference on Machine Learning, pages 8650–8660. PMLR.
  • Qian et al. (2019) Kaizhi Qian, Yang Zhang, Shiyu Chang, Xuesong Yang, and Mark Hasegawa-Johnson. 2019. Autovc: Zero-shot voice style transfer with only autoencoder loss. In International Conference on Machine Learning, pages 5210–5219. PMLR.
  • Qian et al. (2022) Kaizhi Qian, Yang Zhang, Heting Gao, Junrui Ni, Cheng-I Lai, David Cox, Mark Hasegawa-Johnson, and Shiyu Chang. 2022. Contentvec: An improved self-supervised speech representation by disentangling speakers. In International Conference on Machine Learning, pages 18003–18017. PMLR.
  • Qu et al. (2022) Leyuan Qu, Taihao Li, Cornelius Weber, Theresa Pekarek-Rosin, Fuji Ren, and Stefan Wermter. 2022. Disentangling prosody representations with unsupervised speech reconstruction. arXiv preprint arXiv:2212.06972.
  • Radford et al. (2022) Alec Radford, Jong Wook Kim, Tao Xu, Greg Brockman, Christine McLeavey, and Ilya Sutskever. 2022. Robust speech recognition via large-scale weak supervision. arXiv preprint arXiv:2212.04356.
  • Ravanelli et al. (2021) Mirco Ravanelli, Titouan Parcollet, Peter Plantinga, Aku Rouhe, Samuele Cornell, Loren Lugosch, Cem Subakan, Nauman Dawalatabad, Abdelwahab Heba, Jianyuan Zhong, Ju-Chieh Chou, Sung-Lin Yeh, Szu-Wei Fu, Chien-Feng Liao, Elena Rastorgueva, François Grondin, William Aris, Hwidong Na, Yan Gao, Renato De Mori, and Yoshua Bengio. 2021. SpeechBrain: A general-purpose speech toolkit. ArXiv:2106.04624.
  • Rubner et al. (1998) Yossi Rubner, Carlo Tomasi, and Leonidas J Guibas. 1998. A metric for distributions with applications to image databases. In Sixth international conference on computer vision (IEEE Cat. No. 98CH36271), pages 59–66. IEEE.
  • Stylianou et al. (1998) Yannis Stylianou, Olivier Cappé, and Eric Moulines. 1998. Continuous probabilistic transform for voice conversion. IEEE Transactions on speech and audio processing, 6(2):131–142.
  • Veaux et al. (2017) Christophe Veaux, Junichi Yamagishi, Kirsten MacDonald, et al. 2017. Superseded-cstr vctk corpus: English multi-speaker corpus for cstr voice cloning toolkit.
  • Wang et al. (2021) Changhan Wang, Wei-Ning Hsu, Yossi Adi, Adam Polyak, Ann Lee, Peng-Jen Chen, Jiatao Gu, and Juan Pino. 2021. fairseq s^2: A scalable and integrable speech synthesis toolkit. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pages 143–152, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
  • Williams (1965) Carl Earl Williams. 1965. The effects of selected factors on the aural identification of speakers. University of Illinois at Urbana-Champaign.
  • Yuan et al. (2021) Siyang Yuan, Pengyu Cheng, Ruiyi Zhang, Weituo Hao, Zhe Gan, and Lawrence Carin. 2021. Improving zero-shot voice style transfer via disentangled representation learning. In International Conference on Learning Representations.
  • Zhou et al. (2022) Kun Zhou, Berrak Sisman, Rui Liu, and Haizhou Li. 2022. Emotional voice conversion: Theory, databases and esd. Speech Communication, 137:1–18.

Appendix A Appendix

Here we give fine-grained details of computational costs, hyper-parameters and other implementation details omitted from the main part for brevity and readability.

A.1 Dataset details

We use several datasets: (i) VCTK dataset Veaux et al. (2017); (ii) ESD Zhou et al. (2022); (iii) and the a synthetic VCTK dataset (Syn_VCTK), generated by us. For VCTK the training data contains 41k samples, and the evaluation set contains 2.5k. For ESD the training data contains 3k samples, the validation contains 200, and the test 300. For Syn_VCTK the training data contains 2.4k samples and the validation has 142. For ESD we use the official train, val, test split. For both VCTK versions we enforce unpaired data in the train sets, and use paired data only for evaluation.

A.2 Computational Price

We use a pretrained HuBERT_Base model for input encoding. This model has 90 million parameters and takes less than 15 minutes to inference over all of our datasets. For training times see the original paper. For the HiFi-GAN vocoder which is trained once, independently of the speaking style conversion, we use a model with 13.8 million parameters (for the generator). It takes 4 days on to train on VCTK using two RTX 6000 GPUs, other datasets take less than that.

The rhythm predictor E d⁢u⁢r subscript 𝐸 𝑑 𝑢 𝑟 E_{dur}italic_E start_POSTSUBSCRIPT italic_d italic_u italic_r end_POSTSUBSCRIPT has 330 thousand parameters, and takes about 30 minutes to train on a single RTX2080 GPU. Likewise, the pitch predictor E f⁢0 subscript 𝐸 𝑓 0 E_{f0}italic_E start_POSTSUBSCRIPT italic_f 0 end_POSTSUBSCRIPT has 527 thousand parameters, and also takes about 30 minutes to train on a single RTX2080 GPU.

A.3 Hyper-Parameters

We encode the HuBERT units and pitch contour with textlesslib. For the content endocer we use hubert-base-ls960 and use K-means with 100 clusters for the clustering step. For the pitch contour we use the default parameters from textless lib which match the content encoder we used.

For the vocoder we use the official implementation of speech resynthesis with the configuration suitable for VCTK, hubert100, using a fixed size lookup-table. We adapt the model so that it receives as input the normalised pitch without quantisation. We train it with a batch size of 64 over two GPUs for the minimum between 2000 epochs and 400,000 batches.

The rhythm predictor E d⁢u⁢r subscript 𝐸 𝑑 𝑢 𝑟 E_{dur}italic_E start_POSTSUBSCRIPT italic_d italic_u italic_r end_POSTSUBSCRIPT is built of 8 blocks of Conv1D followed by Batch Normalisation, with 128 hidden units. The speaker embedding and token embedding are learned with a lookup-table, both with size 32. The speaker vector is repeated along the time axis and concatenated with the unit embedding. Each CNN has a receptive field of 3 and padding of 1 so that the output shape is the same as the input. For activation we use Leaky ReLU. We normalise the length for prediction with the mean and standard deviation of the training data. Train time masking masks each unit in the input independently with 20% probability. We train the model for 30 epochs, with a batch size of 32, a learning rate of 3⁢e−4 3 𝑒 4 3e-4 3 italic_e - 4 and Adam optimiser. We take the model with lowest validation MSE. This architecture and parameters were fairly standard values which worked out of the box, no manual or automated hyper-tuning was done.

The pitch predictor E F⁢0 subscript 𝐸 𝐹 0 E_{F0}italic_E start_POSTSUBSCRIPT italic_F 0 end_POSTSUBSCRIPT encodes the speaker and units in the same way as E d⁢u⁢r subscript 𝐸 𝑑 𝑢 𝑟 E_{dur}italic_E start_POSTSUBSCRIPT italic_d italic_u italic_r end_POSTSUBSCRIPT with a fixed size lookup-table of size 32. However, here we add to the speaker embedding a linear positional encoding - this means that indices 0-15 of the speaker vector get added the unit index relative to the start and indices 16-31 get added the unit index relative to the end. In preliminary experiments more typical sinusoidal positional encoding worked similarly but we chose the linear version for simplicity. The model is comprised of 9 shared Conv1D layers with kernel size 3, and followed by 2 Conv1D with kernel sizes 3 and 1 for each head (speaking classification and F0 regression). We use Leaky ReLU activation. The labels for training are received from the YAAPT algorithm, normalised per-speaker with the mean and standard deviation of the vocalised parts of the training set. We do not normalise the non-vocalised parts. The BCE loss for binary classification is multiplied by 100 and added to the pitch MSE loss, to attempt to give both losses a similar scale. We train the model for 20 epochs with learning rate 3⁢e−4 3 𝑒 4 3e-4 3 italic_e - 4, batch size of 32 and Adam optimiser. We take the model with the lowest validation MSE. The results in Table 1, and Figure 4 are with a similar model, however it does not use positional encoding. The positional encoding was added afterwards as an improvement. This architecture and parameters were fairly standard values which worked out of the box, no manual or automated hyper-tuning was done.

A.4 Human Evaluation Details

We had over 20 different annotators in our subjective tests, who were all non-paid volunteers. They are all fluent, though not necessarily native, in English, and vary in academic background and connection to the field. For privacy reasons the questionnaires are anonymous and no details are collected about the participants.

The quality question is phrased as follows: “Evaluate the quality and naturalness of the following audio segment”, and the raters give a scale of 1-5 with the labels: “Poor”, “Decent”, “Good”, “Very Good”, “Excellent”.

The similarity question is phrased as follows: “The first recording in this section is a reference, mark all recordings in how similar the speaking style is to the original. speaking style refers to the speaking rhythm, intonation etc. Try to ignore the quality of the recording itself”, and the raters give a scale of 1-5 with the labels: “Very different”, “Different”, “Slightly similar”, “Similar”, “Very Similar”.

A.5 Low Resource Language Conversion

As mentioned, DISSC does not use text as any intermediate representation in conversion (or text supervision for training). This means that it can work in low-resource languages which do not have large annotated datasets.

We demonstrate the ability of DISSC to convert a Hebrew utterance to a target VCTK speaker, while not being trained on any Hebrew data (see Fig. 6). While converting the accent in cross-lingual (Hebrew utterance to English speaker) remains challenging, this shows good potential. Notably, this example is out of domain for all model training data. Training DISSC in a textless self-supervised way on Hebrew data will likely improve the performance dramatically.

Image 6: Refer to caption

Figure 6: An example using DISSC to convert a non-English sample. Hebrew is a lower resource language, hence there are little to no transcribed datasets. Thus, no expressive TTS methods exist, making an ASR-TTS VC pipeline impossible. Training DISSC on unlabelled Hebrew data, will likely improve results further. We encourage readers to listen to the accompanying samples.

A.6 WER detailed analysis

While DISSC is the only evaluated method to successfully convert prosody, it has a limitation of slightly harming content compared to some methods which do not convert prosody (e.g. speech resythesis). We see this in the WER and CER metrics and wish to further analyse the cause for this.

We first manually inspect the samples with highest word error rate. We have found that the content loss is mainly due to minor phonological errors which cause transcription change. For example, in cases with a high error rate, the original text is: “Please call Stella” and the transcription after conversion is: “Peace cool Stalla”. However, when looking at the WER of both DISSC and speech resynthesis, we observe that such phonological errors are observed in the same samples for both methods (even without conversion). Therefore, we believe that the root cause of such phenomenon is the quality of the SSL model and DISSC amplified this effect.

Table 3: Word error rate comparison on VCTK, split by source speaker when converting to all other speakers as targets. For instance the first row denotes the WER when converting a fixed set of samples from speaker p231 to all other speakers in the list. We compare DISSC _Rhythm and Speech resynthesis and see that the results vary greatly, and that the success of both methods are correlated.

Table 4: Word error rate comparison on VCTK, split by target speaker when converting from all other speakers as sources. For instance the first row denotes the WER when converting a fixed set of samples from all source speakers in the list (except p231) to speaker p231. We compare DISSC _Rhythm and Speech resynthesis and see that the results hardly depend on the target speaker.

To further, evalute this hypothesis we evaluate the WER per source speaker and per target speaker (results in Tables 3 and 4). This analysis clearly shows that WER per-speaker varies greatly based on the source speaker, but does not depend on the target speaker, this further supporting the fact that this is an issue with the SSL encoding method. An interesting line of future work would be in improving SSL for such cases, or even applying error-correcting codes on the converted speech.

Xet Storage Details

Size:
82.2 kB
·
Xet hash:
37e0b276ac5a4cd2682681241cbb650bf4d1cae7df7953259631148709cb7d9f

Xet efficiently stores files, intelligently splitting them into unique chunks and accelerating uploads and downloads. More info.