Emotion space for analysis and synthesis of facial expression

This paper presents a new emotion model which gives a criteria to decide human's emotion condition from the face image. Our final goal is to realize very natural and user-friendly human-machine communication environment by giving a face to computer terminal or communication system which can also understand the user's emotion condition. So it is necessary for the emotion model to express emotional meanings of a parameterized face expression and its motion quantitatively. Our emotion model is based on 5-layered neural network which has generalization and nonlinear mapping performance. Both input and output layer has the same number of units. So identity mapping can be realized and emotion space can be constructed in the middle-layer (3rd layer). The mapping from input layer to middle layer means emotion recognition and that from middle layer to output layer corresponds to expression synthesis from the emotion value. Training is performed by typical 13 emotion patterns which are expressed by expression parameters. Subjective test of this emotion space proves the propriety of this model. The facial action coding system is selected as an efficient criteria to describe delicate face expression and motion.<<ETX>>

[1]  Shigeo Morishima,et al.  Speech-to-image media conversion based on VQ and neural network , 1991, [Proceedings] ICASSP 91: 1991 International Conference on Acoustics, Speech, and Signal Processing.

[2]  Hiroshi Harashima,et al.  A proposal of a knowledge based isolated word recognition , 1986, ICASSP '86. IEEE International Conference on Acoustics, Speech, and Signal Processing.

[3]  Shizuo Hiki,et al.  Characteristics of the mouth shape in the production of Japanese , 1982 .

[4]  H. Harashima,et al.  Analysis and synthesis of facial expressions in knowledge-based coding of facial image sequences , 1991, [Proceedings] ICASSP 91: 1991 International Conference on Acoustics, Speech, and Signal Processing.

[5]  Hiroshi Harashima,et al.  A Media Conversion from Speech to Facial Image for Intelligent Man-Machine Interface , 1991, IEEE J. Sel. Areas Commun..

[6]  K. Stevens,et al.  Emotions and speech: some acoustical correlates. , 1972, The Journal of the Acoustical Society of America.

[7]  Shigeo Morishima,et al.  A facial image synthesis system for human-machine interface , 1992, [1992] Proceedings IEEE International Workshop on Robot and Human Communication.

[8]  Hiroshi Harashima,et al.  Facial Animation Synthesis for Human-Machine Communication System , 1993, HCI.

[9]  Shigeo Morishima,et al.  Human machine interface using media conversion and model-based coding schemes , 1992 .

[10]  Michael Potmesil,et al.  The pixel machine: a parallel image computer , 1989, SIGGRAPH.