A composite framework for affective sensing

A system capable of interpreting affect from a speaking face must recognise and fuse signals from multiple cues. Building such a system requires the integration of software components to perform tasks such as image registration, video segmentation, speech recognition and classification. Such software components tend to be idiosyncratic, purpose-built, and driven by scripts and textual configuration files. Integrating components to achieve the necessary degree of flexibility to perform full multimodal affective recognition is challenging. We discuss the key requirements and describe a system to perform multimodal affect sensing which integrates such software components and meets these requirements.

[1]  E. Vesterinen,et al.  Affective Computing , 2009, Encyclopedia of Biometrics.

[2]  Philippe Salembier,et al.  MPEG-7 Systems: overview , 2001, IEEE Trans. Circuits Syst. Video Technol..

[3]  Timothy F. Cootes,et al.  Active Appearance Models , 2001, IEEE Trans. Pattern Anal. Mach. Intell..

[4]  Alan Ezust,et al.  An Introduction to Design Patterns in C++ with Qt 4 (Bruce Perens Open Source) , 2006 .

[5]  Ralph Johnson,et al.  design patterns elements of reusable object oriented software , 2019 .

[6]  Changbo Hu,et al.  AAM derived face representations for robust facial action recognition , 2006, 7th International Conference on Automatic Face and Gesture Recognition (FGR06).