A novel human–machine interface based on recognition of multi-channel facial bioelectric signals

This paper presents a novel human–machine interface for disabled people to interact with assistive systems for a better quality of life. It is based on multi-channel forehead bioelectric signals acquired by placing three pairs of electrodes (physical channels) on the Frontalis and Temporalis facial muscles. The acquired signals are passed through a parallel filter bank to explore three different sub-bands related to facial electromyogram, electrooculogram and electroencephalogram. The root mean square features of the bioelectric signals analyzed within non-overlapping 256 ms windows were extracted. The subtractive fuzzy c-means clustering method (SFCM) was applied to segment the feature space and generate initial fuzzy based Takagi–Sugeno rules. Then, an adaptive neuro-fuzzy inference system is exploited to tune up the premises and consequence parameters of the extracted SFCMs rules. The average classifier discriminating ratio for eight different facial gestures (smiling, frowning, pulling up left/right lips corner, eye movement to left/right/up/down) is between 93.04% and 96.99% according to different combinations and fusions of logical features. Experimental results show that the proposed interface has a high degree of accuracy and robustness for discrimination of 8 fundamental facial gestures. Some potential and further capabilities of our approach in human–machine interfaces are also discussed.

[1]  Veikko Surakka,et al.  Facial Activation Control Effect (FACE) , 2007, ACII.

[2]  Rui Wang,et al.  Toward affective hands-free human-machine interface approach in virtual environments-based equipment operation training , 2009 .

[3]  Soo-Young Lee,et al.  A Practical Biosignal-Based Human Interface Applicable to the Assistive Systems for People with Motor Impairment , 2006, IEICE Trans. Inf. Syst..

[4]  Cuntai Guan,et al.  Multiclass voluntary facial expression classification based on Filter Bank Common Spatial Pattern , 2008, 2008 30th Annual International Conference of the IEEE Engineering in Medicine and Biology Society.

[5]  S. Mahlke,et al.  Emotions and EMG measures of facial muscles in interactive contexts , 2006 .

[6]  R. A. Rahmat,et al.  GENERATION OF FUZZY RULES WITH SUBTRACTIVE CLUSTERING , 2005 .

[7]  Mehran Jahed,et al.  Real-time intelligent pattern recognition algorithm for surface EMG signals , 2007, Biomedical engineering online.

[8]  Huosheng Hu,et al.  A Human – Computer Interface based on Forehead Multi-Channel Bio-signals to Control a Virtual Wheelchair , 2002 .

[9]  Xiangyu Wang,et al.  Using affective human-machine interface to increase the operation performance in virtual construction crane training system: A novel approach , 2011 .

[10]  Huosheng Hu,et al.  Myoelectric control systems - A survey , 2007, Biomed. Signal Process. Control..

[11]  Martti Juhola,et al.  Detection of electromyographic signals from facial muscles with neural networks , 2006, TAP.

[12]  Robert Chen,et al.  Representation of facial muscles in human motor cortex , 2005, The Journal of physiology.

[13]  Rosalind W. Picard Affective Computing , 1997 .

[14]  M Adjouadi,et al.  A practical EMG-based human-computer interface for users with motor disabilities. , 2000, Journal of rehabilitation research and development.

[15]  Toshio Tsuji,et al.  A human-assisting manipulator teleoperated by EMG signals and arm motions , 2003, IEEE Trans. Robotics Autom..

[16]  S. Krishnan,et al.  Real-Time Classification of Forearm Electromyographic Signals Corresponding to User-Selected Intentional Movements for Multifunction Prosthesis Control , 2007, IEEE Transactions on Neural Systems and Rehabilitation Engineering.

[17]  Wanderley Cardoso Celeste,et al.  Human–machine interface based on muscular and brain signals applied to a robotic wheelchair , 2007 .

[18]  Cheng-Ning Huang,et al.  The Review of Applications and Measurements in Facial Electromyography , 2004 .

[19]  Othman Omran Khalifa,et al.  EMG signal classification for human computer interaction: a review , 2009 .

[20]  Robert C. Wolpert,et al.  A Review of the , 1985 .

[21]  Huosheng Hu,et al.  EMG-based hands-free wheelchair control with EOG attention shift detection , 2007, 2007 IEEE International Conference on Robotics and Biomimetics (ROBIO).

[22]  Kemal Polat,et al.  Comparison of Different Classifier Algorithms on the Automated Detection of Obstructive Sleep Apnea Syndrome , 2008, Journal of Medical Systems.

[23]  Veikko Surakka Contagion and modulation of human emotions , 1998 .

[24]  R.Fff. Weir,et al.  A heuristic fuzzy logic approach to EMG pattern recognition for multifunctional prosthesis control , 2005, IEEE Transactions on Neural Systems and Rehabilitation Engineering.

[25]  Yoshiaki Saitoh,et al.  Development of communication supporting device controlled by eye movements and voluntary eye blink , 2006, The 26th Annual International Conference of the IEEE Engineering in Medicine and Biology Society.

[26]  L.B.P. Ang,et al.  Facial expression recognition through pattern analysis of facial muscle movements utilizing electromyogram sensors , 2004, 2004 IEEE Region 10 Conference TENCON 2004..

[27]  E. Vesterinen,et al.  Affective Computing , 2009, Encyclopedia of Biometrics.