Visualization of Important Human Motion Feature Using Convolutional Neural Network

Human motion feature extraction is necessary for robot motion generation. In particular, feature extraction methods related to non-periodic motion should be proposed. Recently, the number of the human motion recognition studies utilizing Convolutional Neural Network (CNN) is increasing due to the brilliant ability of extracting and identifying features. CNN has the same or better discrimination ability than humans so that the features extracted by CNN are thought to be useful for human motion understanding. However, since the internal structure of CNN is like a black box, it is difficult to understand the extracted features. In order to understand the features, some visualization methods which are utilized in the image field are applied in this paper. Furthermore, the many conventional studies utilizing the gradient are not preferred for visualizing CNN which recognizes human motion because human motion is sensitive for the object which treated. That is, the gradient method does not work because the value varies greatly depending on the environment even in the same motion. Therefore, new visualizing method without using the gradient is proposed in this paper. The proposed method visualizes CNN focusing part by following the neuron in CNN. Since this method does not require the gradient, the stable and accurate visualization can be performed. The effectiveness is shown by experiments.

[1]  Avanti Shrikumar,et al.  Learning Important Features Through Propagating Activation Differences , 2017, ICML.

[2]  Danna Zhou,et al.  d. , 1934, Microbial pathogenesis.

[3]  Li Xue,et al.  Understanding and Improving Deep Neural Network for Activity Recognition , 2018 .

[4]  Alexander Binder,et al.  On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation , 2015, PloS one.

[5]  Rob Fergus,et al.  Visualizing and Understanding Convolutional Networks , 2013, ECCV.

[6]  Senem Velipasalar,et al.  A Survey on Activity Detection and Classification Using Wearable Sensors , 2017, IEEE Sensors Journal.

[7]  T Togawa,et al.  Classification of acceleration waveforms during walking by wavelet transform. , 1997, Methods of information in medicine.

[8]  Dumitru Erhan,et al.  Going deeper with convolutions , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[9]  Tsuyoshi Murata,et al.  {m , 1934, ACML.

[10]  Mi Zhang,et al.  A feature selection-based framework for human activity recognition using wearable multimodal sensors , 2011, BODYNETS.

[11]  Bo Yu,et al.  Convolutional Neural Networks for human activity recognition using mobile sensors , 2014, 6th International Conference on Mobile Computing, Applications and Services.

[12]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[13]  Zhenyu He,et al.  Activity recognition from acceleration data based on discrete consine transform and SVM , 2009, 2009 IEEE International Conference on Systems, Man and Cybernetics.

[14]  Thomas Brox,et al.  Striving for Simplicity: The All Convolutional Net , 2014, ICLR.

[15]  Yang Zhang,et al.  A Theoretical Explanation for Perplexing Behaviors of Backpropagation-based Visualizations , 2018, ICML.

[16]  Andreas Krause,et al.  Unsupervised, dynamic identification of physiological and activity context in wearable computing , 2003, Seventh IEEE International Symposium on Wearable Computers, 2003. Proceedings..

[17]  Miguel A. Labrador,et al.  A Survey on Human Activity Recognition using Wearable Sensors , 2013, IEEE Communications Surveys & Tutorials.

[18]  M. Tahar Kechadi,et al.  Human Activity Recognition with Convolutional Neural Networks , 2018, ECML/PKDD.

[19]  松永 卓也 Development of haptic end-effector for medicine and manufacturing(本文) , 2017 .