Gesture recognition algorithm based on multi-scale feature fusion in RGB-D images

With the rapid development of sensor technology and artificial intelligence, the video gesture recognition technology under the background of big data makes human–computer interaction more natural and flexible, bringing the richer interactive experience to teaching, on-board control, electronic games etc. To perform robust recognition under the conditions of illumination change, background clutter, rapid movement, and partial occlusion, an algorithm based on multi-level feature fusion of two-stream convolutional neural network is proposed, which includes three main steps. Firstly, the Kinect sensor obtains red–green–blue-depth (RGB-D) images to establish a gesture database. At the same time, data enhancement is performed on the training set and test set. Then, a model of multi-level feature fusion of a two-stream convolutional neural network is established and trained. Experiments show that the proposed network model can robustly track and recognise gestures under complex backgrounds (such as similar complexion, illumination changes, and occlusion), and compared with the single-channel model, the average detection accuracy is improved by 1.08%, and mean average precision is improved by 3.56%.

[1]  Ying Sun,et al.  Surface EMG hand gesture recognition system based on PCA and GRNN , 2019, Neural Computing and Applications.

[2]  Gongfa Li,et al.  Grasping force prediction based on sEMG signals , 2020 .

[3]  Honghai Liu,et al.  Hand gesture recognition based on convolution neural network , 2017, Cluster Computing.

[4]  Renqiang Xie,et al.  Accelerometer-Based Hand Gesture Recognition by Neural Network and Similarity Matching , 2016, IEEE Sensors Journal.

[5]  Yuan Li,et al.  Deep attention network for joint hand gesture localization and recognition using static RGB-D images , 2018, Inf. Sci..

[6]  Karl F. MacDorman,et al.  Review of constraints on vision-based gesture recognition for human-computer interaction , 2018, IET Comput. Vis..

[7]  Gongfa Li,et al.  CNN-Based Facial Expression Recognition from Annotated RGB-D Images for Human-Robot Interaction , 2019, Int. J. Humanoid Robotics.

[8]  Biing-Hwang Juang,et al.  Feature Processing and Modeling for 6D Motion Gesture Recognition , 2013, IEEE Transactions on Multimedia.

[9]  Honghai Liu,et al.  Gesture recognition based on modified adaptive orthogonal matching pursuit algorithm , 2017, Cluster Computing.

[10]  Gongfa Li,et al.  Human Lesion Detection Method Based on Image Information and Brain Signal , 2019, IEEE Access.

[11]  Bo Tao,et al.  Probability analysis for grasp planning facing the field of medical robotics , 2019, Measurement.

[12]  Dieter Fox,et al.  Unsupervised Feature Learning for RGB-D Based Object Recognition , 2012, ISER.

[13]  Ying Sun,et al.  Towards the sEMG hand: internet of things sensors and haptic feedback application , 2018, Multimedia Tools and Applications.

[14]  Hu He,et al.  Saliency and depth-based unsupervised object segmentation , 2016, IET Image Process..

[15]  Honghai Liu,et al.  Gesture Recognition Based on Kinect and sEMG Signal Fusion , 2018, Mobile Networks and Applications.

[16]  Honghai Liu,et al.  Simultaneous Calibration: A Joint Optimization Approach for Multiple Kinect and External Cameras , 2017, Sensors.

[17]  Bo Tao,et al.  Intelligent Human-Computer Interaction Based on Surface EMG Gesture Recognition , 2019, IEEE Access.

[18]  Hao Wu,et al.  Dynamic Gesture Recognition in the Internet of Things , 2019, IEEE Access.

[19]  Bo Tao,et al.  Gesture recognition based on skeletonization algorithm and CNN with ASL database , 2018, Multimedia Tools and Applications.

[20]  Hui Yu,et al.  Gesture recognition based on binocular vision , 2018, Cluster Computing.

[21]  Tanima Dutta,et al.  A Continuous Hand Gestures Recognition Technique for Human-Machine Interaction Using Accelerometer and Gyroscope Sensors , 2016, IEEE Sensors Journal.

[22]  Yi Yao,et al.  Continuous camera placement using multiple objective optimisation process , 2015, IET Comput. Vis..

[23]  D. Jude Hemanth,et al.  Hand posture and gesture recognition techniques for virtual reality applications: a survey , 2016, Virtual Reality.

[24]  Gongfa Li,et al.  A novel feature extraction method for machine learning based on surface electromyography from healthy brain , 2019, Neural Computing and Applications.

[25]  Honghai Liu,et al.  Gesture recognition based on an improved local sparse representation classification algorithm , 2017, Cluster Computing.

[26]  Honghai Liu,et al.  Jointly network: a network based on CNN and RBM for gesture recognition , 2018, Neural Computing and Applications.

[27]  Honghai Liu,et al.  Optimal grasp planning of multi-fingered robotic hands:a review , 2015 .

[28]  Honghai Liu,et al.  An Interactive Image Segmentation Method in Hand Gesture Recognition , 2017, Sensors.

[29]  Jun Li,et al.  Deeply Learned View-Invariant Features for Cross-View Action Recognition , 2017, IEEE Transactions on Image Processing.

[30]  Yizhou Yu,et al.  Automatic object extraction from images using deep neural networks and the level-set method , 2018, IET Image Process..