Extraction of speaker's feeling using facial image and speech

While all sorts of systems have been computerized, the importance of the interactive communication between human and computer rises. In this research we propose a method that merges information of prosody, facial expression in macro and the movement of the head, and classify these into a category of feelings. By using both visual and speech information the accuracy of the recognition of human feeling up to 75% was achieved.

[1]  Roberto Brunelli,et al.  Face Recognition: Features Versus Templates , 1993, IEEE Trans. Pattern Anal. Mach. Intell..

[2]  Patrick Bouthemy,et al.  Multimodal Estimation of Discontinuous Optical Flow using Markov Random Fields , 1993, IEEE Trans. Pattern Anal. Mach. Intell..

[3]  Takaaki Akimoto,et al.  Automatic creation of 3D facial models , 1993, IEEE Computer Graphics and Applications.