Identification of Multimodal Human-Robot Interaction Using Combined Kernels

In this paper we propose a methodology to build multiclass classifiers for the human-robot interaction problem. Our solution uses kernel-based classifiers and assumes that each data type is better represented by a different kernel. The kernels are then combined into one single kernel that uses all the dataset involved in the HRI process. The results on real data shows that our proposal is capable of obtaining lower generalization errors due to the use of specific kernels for each data type. Also, we show that our proposal is more robust when presented to noise in either or both data types.

[1]  Ethem Alpaydin,et al.  Multiple Kernel Learning Algorithms , 2011, J. Mach. Learn. Res..

[2]  Bernhard Schölkopf,et al.  Estimating the Support of a High-Dimensional Distribution , 2001, Neural Computation.

[3]  P. Dhanalakshmi,et al.  Classification of audio signals using SVM and RBFNN , 2009, Expert Syst. Appl..

[4]  Sven Behnke,et al.  RoboCup 2013: Robot World Cup XVII , 2014, Lecture Notes in Computer Science.

[5]  S. Palanivel,et al.  Audio-Video based Classification using SVM and AANN , 2012 .

[6]  Chih-Jen Lin,et al.  LIBSVM: A library for support vector machines , 2011, TIST.

[7]  Michael A. Goodrich,et al.  Human-Robot Interaction: A Survey , 2008, Found. Trends Hum. Comput. Interact..

[8]  S. Bhattacharya,et al.  Gesture classification with machine learning using Kinect sensor data , 2012, 2012 Third International Conference on Emerging Applications of Information Technology.

[9]  Malik Yousef,et al.  One-Class SVMs for Document Classification , 2002, J. Mach. Learn. Res..

[10]  C. Krishna Mohan,et al.  Content-Based Video Classification Using Support Vector Machines , 2004, ICONIP.