| { |
| "paper_id": "O13-1008", |
| "header": { |
| "generated_with": "S2ORC 1.0.0", |
| "date_generated": "2023-01-19T08:03:29.832340Z" |
| }, |
| "title": "\u5408\u6210\u55ae\u5143\u8207\u554f\u984c\u96c6\u4e4b\u5b9a\u7fa9\u65bc\u96b1\u85cf\u5f0f\u99ac\u53ef\u592b\u6a21\u578b\u4e2d\u6587\u6b4c\u8072\u5408\u6210\u7cfb\u7d71\u4e4b\u5efa\u7acb Synthesis Unit and Question Set Definition for Mandarin HMM-based Singing Voice Synthesis", |
| "authors": [ |
| { |
| "first": "Ju-Yun", |
| "middle": [], |
| "last": "Cheng", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "National Cheng Kung University", |
| "location": { |
| "settlement": "Tainan", |
| "country": "Taiwan" |
| } |
| }, |
| "email": "" |
| }, |
| { |
| "first": "Yi-Chin", |
| "middle": [], |
| "last": "Huang", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "National Cheng Kung University", |
| "location": { |
| "settlement": "Tainan", |
| "country": "Taiwan" |
| } |
| }, |
| "email": "ychin.huang@gmail.com" |
| }, |
| { |
| "first": "Chung-Hsien", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "", |
| "affiliation": { |
| "laboratory": "", |
| "institution": "National Cheng Kung University", |
| "location": { |
| "settlement": "Tainan", |
| "country": "Taiwan" |
| } |
| }, |
| "email": "chunghsienwu@gmail.com" |
| } |
| ], |
| "year": "", |
| "venue": null, |
| "identifiers": {}, |
| "abstract": "", |
| "pdf_parse": { |
| "paper_id": "O13-1008", |
| "_pdf_hash": "", |
| "abstract": [], |
| "body_text": [ |
| { |
| "text": "The fluency and continuity properties are very important in singing voice synthesis. In order to synthesize smooth and continuous singing voice, the Hidden Markov Model (HMM)-based synthesis approach is employed to build our Mandarin singing voice synthesis system. The system is designed to generate Mandarin songs with arbitrary lyrics and melodies in a certain pitch range. We also build a singing voice database for system training and synthesis, which is designed based on the phonetic converge of Mandarin speech. In addition, the acoustic feature extraction using STRAIGHT algorithm is employed to generate satisfactory vocoded singing voices.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "The purpose of this paper is to elaborate the construction of Mandarin singing voice synthesis system by defining the synthesis model and question set for HMM-based singing voice synthesis. In addition, we implemented two techniques, including pitch-shift pseudo data extension and vibrato post-processing, to make synthesized singing voice more natural.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "The proposed system framework consists of two main phases, the training phase and the synthesis phase. In the training phase, excitation, spectral and aperiodic factors are extracted from a singing voice database. The lyrics and notes of songs in the singing voice corpus are considered as contextual information for generating context-dependent label sequences. Then, the sequences are clustered with context-dependent question set and then the context-dependent HMMs are trained based on the clustered phone segments. In the synthesis phase, the input musical score and the lyric are converted into a context-dependent label sequence. The label sequence, consisting of excitation, spectrum and aperiodic factors, for the given song is constructed by concatenating the parameters generated from the context-dependent HMMs. Finally, the generated parameter sequences are synthesized using Mel Log Spectrum Approximation (MLSA) filter to generate the singing voice.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "The approaches used in this study are to improve the model accuracy by defining the question set, extending the singing voice database through generating pitch-shift pseudo data, and adding the vibrato singing skill using signal post-processing. The selection of question set is crucial to generate proper synthesis models. In the baseline system, the most frequently used questions of F0 and mel-cepstral clustering trees are sub-syllables types, position of note and phrase level. Since the recorded singing database is not large enough to contain each combination of contextual factors. Thus, only essential and suitable questions are defined compared to the traditional method. Besides, the extended pitch-shift pseudo data are helpful to cover the missing pitch information of sub-syllables and increase the size of the training data. Based on the analysis results of the defined pitch range (C4~B4) of the recorded singing corpus, shifting the frequency of a note too much would change the timbre. Thus, the missing pitch information of sub-syllables of the recorded corpus is compensated using the nearby notes from other songs, and shifting the frequency of signal to the corresponding Hertz by a pitch-to-frequency mapping table. The vocal vibrato is a natural oscillation of musical pitch and the singers generally employ vibrato as an expressive and musically useful aspect of the performance. So adding vibrato can make synthesized singing voice more natural and expressive. The frequency and the amplitude can be considered since the two fundamental parameters affect the singing voice with vibrato effect. The method to create vibrato is to vary the time delay periodically and use the principle of Doppler Effect. Our system implemented this phenomenon by a delay line and a low frequency oscillator (LFO) to vary the delay.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "For evaluation, the singing voice signals were sampled at a rate of 48 kHz and windowed by a 25ms Blackman window with a 5ms shift. Then mel-cepstral coefficients were obtained from STRAIGHTextracted spectra. The feature vectors consist of spectrum, excitation and aperiodic factor. The spectrum parameter vectors consist of 49th-order STRAIGHT mel-cepstral coefficients including the zero-th coefficient, their delta, and delta-delta coefficients. The excitation parameter vectors consist of log F0, its Proceedings of the Twenty-Fifth Conference on Computational Linguistics and Speech Processing (ROCLING 2013) delta, and delta-delta. A five-state, left-to-right Hidden Semi-Markov Models (HSMM) was employed in which the spectral part of the state was modeled by a single diagonal Gaussian output distribution. The excitation stream was modeled with multi-space probability distributions HSMM (MSD-HSMM), each of which consists of a Gaussian distribution for \"voiced\" frames and a discrete distribution for \"unvoiced\" frames. We evaluate the nature of synthesized singing voice with the long duration model, and the result show that the system with long duration model obtained 62% preference higher than 38% for the system without long duration model. It shows that long duration model can actually improve the nature of phones with longer duration. Besides, the experimental results show that suitable question set definition can improve the quality and intelligibility of synthesized singing voice, and pitch-shift pseudo data and vibrato post-processing can successfully improve the quality and naturalness of the synthesized singing voice.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "In conclusion, a corpus-based Mandarin singing voice synthesis system based on HMM framework was implemented in this paper. We defined the Mandarin synthesis models and the question set for model clustering and construction. In the context-dependent HMM, linguistic information and musical information are considered. Music information such as pitch, duration, is included to model the singing characteristics. Furthermore, we used three methods to refine our system, i.e. question set definition, pitch-shift pseudo data extension and vibrato post-processing. Experimental results show that our system can synthesize singing voice successfully and the refinements can actually improve the fluency and continuity of the proposed Mandarin singing voice synthesis system.", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| }, |
| { |
| "text": "Proceedings of the Twenty-Fifth Conference on Computational Linguistics and Speech Processing(ROCLING 2013)", |
| "cite_spans": [], |
| "ref_spans": [], |
| "eq_spans": [], |
| "section": "", |
| "sec_num": null |
| } |
| ], |
| "back_matter": [], |
| "bib_entries": { |
| "BIBREF0": { |
| "ref_id": "b0", |
| "title": "VOCALOID-Commercial singing synthesizer based on sample concatenation", |
| "authors": [ |
| { |
| "first": "H", |
| "middle": [], |
| "last": "Kenmochi", |
| "suffix": "" |
| }, |
| { |
| "first": "H", |
| "middle": [], |
| "last": "Ohshita", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "INTERSPEECH", |
| "volume": "", |
| "issue": "", |
| "pages": "4009--4010", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "H. Kenmochi, H. Ohshita, \"VOCALOID-Commercial singing synthesizer based on sample concatenation\", in INTERSPEECH, pp.4009-4010, 2007.", |
| "links": null |
| }, |
| "BIBREF1": { |
| "ref_id": "b1", |
| "title": "A Corpus-Based Concatenative Mandarin Singing voice Synthesis System", |
| "authors": [ |
| { |
| "first": "S.-S", |
| "middle": [], |
| "last": "Zhou", |
| "suffix": "" |
| }, |
| { |
| "first": "Q.-C", |
| "middle": [], |
| "last": "Chen", |
| "suffix": "" |
| }, |
| { |
| "first": "D.-D", |
| "middle": [], |
| "last": "Wang", |
| "suffix": "" |
| }, |
| { |
| "first": "X.-H", |
| "middle": [], |
| "last": "Yang", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "Machine Learning and Cybernetics", |
| "volume": "5", |
| "issue": "", |
| "pages": "2695--2699", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "S.-S. Zhou, Q.-C. Chen, D.-D. Wang, X.-H. Yang, \"A Corpus-Based Concatenative Mandarin Singing voice Synthesis System\", in Machine Learning and Cybernetics, 2008 International Conference on, vol.5, no., pp.2695-2699, 2008.", |
| "links": null |
| }, |
| "BIBREF2": { |
| "ref_id": "b2", |
| "title": "Simultaneous modeling of spectrum, pitch and duration in HMM-based speech synthesis", |
| "authors": [ |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Yoshimura", |
| "suffix": "" |
| }, |
| { |
| "first": "K", |
| "middle": [], |
| "last": "Tokuda", |
| "suffix": "" |
| }, |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Masuko", |
| "suffix": "" |
| }, |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Kobayashi", |
| "suffix": "" |
| }, |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Kitamura", |
| "suffix": "" |
| } |
| ], |
| "year": 1999, |
| "venue": "in EUROSPEECH", |
| "volume": "5", |
| "issue": "", |
| "pages": "2347--2350", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "T. Yoshimura, K. Tokuda, T. Masuko, T. Kobayashi and T. Kitamura, \"Simultaneous modeling of spectrum, pitch and duration in HMM-based speech synthesis\", in EUROSPEECH, vol.5, pp.2347- 2350, 1999.", |
| "links": null |
| }, |
| "BIBREF3": { |
| "ref_id": "b3", |
| "title": "Recent Development of the HMM-based Singing Voice Synthesis System-Sinsy", |
| "authors": [ |
| { |
| "first": "K", |
| "middle": [], |
| "last": "Oura", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Mase", |
| "suffix": "" |
| }, |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Yamada", |
| "suffix": "" |
| }, |
| { |
| "first": "S", |
| "middle": [], |
| "last": "Muto", |
| "suffix": "" |
| }, |
| { |
| "first": "Y", |
| "middle": [], |
| "last": "Nankaku", |
| "suffix": "" |
| }, |
| { |
| "first": "K", |
| "middle": [], |
| "last": "Tokuda", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "th ISCA Speech Synthesis Workshop", |
| "volume": "", |
| "issue": "", |
| "pages": "211--216", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "K. Oura, A. Mase, T. Yamada, S. Muto, Y. Nankaku, and K. Tokuda, \"Recent Development of the HMM-based Singing Voice Synthesis System-Sinsy\", in 7 th ISCA Speech Synthesis Workshop, pp.211-216, 2010.", |
| "links": null |
| }, |
| "BIBREF4": { |
| "ref_id": "b4", |
| "title": "Mandarin Singing Voice Synthesis Using an HNM Based Scheme", |
| "authors": [ |
| { |
| "first": "H.-Y", |
| "middle": [], |
| "last": "Gu", |
| "suffix": "" |
| }, |
| { |
| "first": "H.-L", |
| "middle": [], |
| "last": "Liau", |
| "suffix": "" |
| } |
| ], |
| "year": 2008, |
| "venue": "International Congress on Image and Signal Processing", |
| "volume": "5", |
| "issue": "", |
| "pages": "347--351", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "H.-Y. Gu, H.-L. Liau, \"Mandarin Singing Voice Synthesis Using an HNM Based Scheme,\" in International Congress on Image and Signal Processing (CISP), vol.5, no., pp.347-351, 2008.", |
| "links": null |
| }, |
| "BIBREF5": { |
| "ref_id": "b5", |
| "title": "Speech-to-Singing Synthesis: Converting Speaking Voices to Singing Voices by Controlling Acoustic Features Unique to Singing Voices", |
| "authors": [ |
| { |
| "first": "T", |
| "middle": [], |
| "last": "Saitou", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Goto", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Unoki", |
| "suffix": "" |
| }, |
| { |
| "first": "M", |
| "middle": [], |
| "last": "Akagi", |
| "suffix": "" |
| } |
| ], |
| "year": 2007, |
| "venue": "Applications of Signal Processing to Audio and Acoustics Workshop on", |
| "volume": "218", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "T. Saitou, M. Goto, M. Unoki, M. Akagi, \"Speech-to-Singing Synthesis: Converting Speaking Voices to Singing Voices by Controlling Acoustic Features Unique to Singing Voices\", in Applications of Signal Processing to Audio and Acoustics Workshop on, vol., no., pp.215,218, 2007", |
| "links": null |
| }, |
| "BIBREF6": { |
| "ref_id": "b6", |
| "title": "A Lyrics to Singing Voice Synthesis System with Variable Timbre", |
| "authors": [ |
| { |
| "first": "J", |
| "middle": [], |
| "last": "Li", |
| "suffix": "" |
| }, |
| { |
| "first": "H", |
| "middle": [], |
| "last": "Yang", |
| "suffix": "" |
| }, |
| { |
| "first": "W", |
| "middle": [], |
| "last": "Zhang", |
| "suffix": "" |
| }, |
| { |
| "first": "L", |
| "middle": [], |
| "last": "Cai", |
| "suffix": "" |
| } |
| ], |
| "year": 2011, |
| "venue": "Applied Informatics and Communication Communications in Computer and Information Science", |
| "volume": "225", |
| "issue": "", |
| "pages": "186--193", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "J. Li, H. Yang, W. Zhang, L. Cai, \"A Lyrics to Singing Voice Synthesis System with Variable Timbre\", in Applied Informatics and Communication Communications in Computer and Information Science, vol.225, pp.186-193, 2011.", |
| "links": null |
| }, |
| "BIBREF7": { |
| "ref_id": "b7", |
| "title": "Singing Voice Analysis/Synthesis", |
| "authors": [ |
| { |
| "first": "Y", |
| "middle": [ |
| "E" |
| ], |
| "last": "Kim", |
| "suffix": "" |
| } |
| ], |
| "year": 2003, |
| "venue": "", |
| "volume": "", |
| "issue": "", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Y. E. Kim, \"Singing Voice Analysis/Synthesis\", Massachusetts Institute of Technology, 2003.", |
| "links": null |
| }, |
| "BIBREF8": { |
| "ref_id": "b8", |
| "title": "Exploiting Prosody Hierarchy and Dynamic Features for Pitch Modeling and Generation in HMM-based Speech Synthesis", |
| "authors": [ |
| { |
| "first": "C.-C", |
| "middle": [], |
| "last": "Hsia", |
| "suffix": "" |
| }, |
| { |
| "first": "C.-H", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| }, |
| { |
| "first": "J.-Y", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| } |
| ], |
| "year": 2010, |
| "venue": "IEEE Trans. Audio, Speech, and Language Processing", |
| "volume": "18", |
| "issue": "8", |
| "pages": "", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "C.-C. Hsia, C.-H. Wu and J.-Y. Wu, \"Exploiting Prosody Hierarchy and Dynamic Features for Pitch Modeling and Generation in HMM-based Speech Synthesis,\" IEEE Trans. Audio, Speech, and Language Processing, Vol. 18, No. 8, November 2010, pp. 1994~2003.", |
| "links": null |
| }, |
| "BIBREF9": { |
| "ref_id": "b9", |
| "title": "Hierarchical prosodic pattern selection based on Fujisaki model for natural mandarin speech synthesis", |
| "authors": [ |
| { |
| "first": "Y.-C", |
| "middle": [], |
| "last": "Huang", |
| "suffix": "" |
| }, |
| { |
| "first": "C.-H", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| }, |
| { |
| "first": "S.-T", |
| "middle": [], |
| "last": "Weng", |
| "suffix": "" |
| } |
| ], |
| "year": 2012, |
| "venue": "Chinese Spoken Language Processing (ICASSP), 2012 8th International Symposium on", |
| "volume": "", |
| "issue": "", |
| "pages": "79--83", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Y.-C. Huang, C.-H. Wu, S.-T. Weng, \"Hierarchical prosodic pattern selection based on Fujisaki model for natural mandarin speech synthesis\", in Chinese Spoken Language Processing (ICASSP), 2012 8th International Symposium on , vol., no., pp.79-83, 2012", |
| "links": null |
| }, |
| "BIBREF10": { |
| "ref_id": "b10", |
| "title": "Personalized Spectral and Prosody Conversion using Frame-Based Codeword Distribution and Adaptive CRF", |
| "authors": [ |
| { |
| "first": "Y.-C", |
| "middle": [], |
| "last": "Huang", |
| "suffix": "" |
| }, |
| { |
| "first": "C.-H", |
| "middle": [], |
| "last": "Wu", |
| "suffix": "" |
| }, |
| { |
| "first": "Y.-T", |
| "middle": [], |
| "last": "Chao", |
| "suffix": "" |
| } |
| ], |
| "year": 2013, |
| "venue": "IEEE Trans. Audio, Speech, and Language Processing", |
| "volume": "21", |
| "issue": "1", |
| "pages": "51--62", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "Y.-C. Huang, C.-H. Wu, and Y.-T. Chao, \"Personalized Spectral and Prosody Conversion using Frame-Based Codeword Distribution and Adaptive CRF,\" IEEE Trans. Audio, Speech, and Language Processing, Vol. 21, No. 1, January 2013, pp. 51~62.", |
| "links": null |
| }, |
| "BIBREF11": { |
| "ref_id": "b11", |
| "title": "An HMM-based Singing Voice Synthesis System", |
| "authors": [ |
| { |
| "first": "K", |
| "middle": [], |
| "last": "Saino", |
| "suffix": "" |
| }, |
| { |
| "first": "H", |
| "middle": [], |
| "last": "Zen", |
| "suffix": "" |
| }, |
| { |
| "first": "Y", |
| "middle": [], |
| "last": "Nankaku", |
| "suffix": "" |
| }, |
| { |
| "first": "A", |
| "middle": [], |
| "last": "Lee", |
| "suffix": "" |
| }, |
| { |
| "first": "K", |
| "middle": [], |
| "last": "Tokuda", |
| "suffix": "" |
| } |
| ], |
| "year": 2006, |
| "venue": "International Conference on Spoken Language Processing (ICSLP)", |
| "volume": "", |
| "issue": "", |
| "pages": "1141--1144", |
| "other_ids": {}, |
| "num": null, |
| "urls": [], |
| "raw_text": "K. Saino, H. Zen, Y. Nankaku, A. Lee, and K. Tokuda \"An HMM-based Singing Voice Synthesis System\", in International Conference on Spoken Language Processing (ICSLP), pp. 1141-1144, 2006.", |
| "links": null |
| } |
| }, |
| "ref_entries": {} |
| } |
| } |