HMM-based speech synthesis with unsupervised labeling of accentual context based on F0 quantization and average voice model
暂无分享,去创建一个
[1] Tomoki Toda,et al. Voice Conversion Based on Maximum-Likelihood Estimation of Spectral Parameter Trajectory , 2007, IEEE Transactions on Audio, Speech, and Language Processing.
[2] Heiga Zen,et al. Hidden Semi-Markov Model Based Speech Synthesis System , 2006 .
[3] Heiga Zen,et al. Unsupervised adaptation for HMM-based speech synthesis , 2008, INTERSPEECH.
[4] Keiichi Tokuda,et al. Simultaneous modeling of spectrum, pitch and duration in HMM-based speech synthesis , 1999, EUROSPEECH.
[5] Takashi Nose,et al. A Speaker Adaptation Technique for MRHSMM-Based Style Control of Synthetic Speech , 2007, 2007 IEEE International Conference on Acoustics, Speech and Signal Processing - ICASSP '07.
[6] Keiichi Tokuda,et al. Text-to-speech synthesis with arbitrary speaker's voice from average voice , 2001, INTERSPEECH.
[7] Matthew Gibson. Two-pass decision tree construction for unsupervised adaptation of HMM-based synthesis models , 2009, INTERSPEECH.
[8] K. Tokuda,et al. A Training Method of Average Voice Model for HMM-Based Speech Synthesis , 2003, IEICE Trans. Fundam. Electron. Commun. Comput. Sci..
[9] Takao Kobayashi,et al. Acoustic Modeling of Speaking Styles and Emotional Expressions in HMM-Based Speech Synthesis , 2005, IEICE Trans. Inf. Syst..