Accurate and Fast Classification of Foot Gestures for Virtual Locomotion

This work explores the use of foot gestures for locomotion in virtual environments. Foot gestures are represented as the distribution of plantar pressure and detected by three sparsely-located sensors on each insole. The Long Short-Term Memory model is chosen as the classifier to recognize the performer's foot gesture based on the captured signals of pressure information. The trained classifier directly takes the noisy and sparse input of sensor data, and handles seven categories of foot gestures (stand, walk forward/backward, run, jump, slide left and right) without manual definition of signal features for classifying these gestures. This classifier is capable of recognizing the foot gestures, even with the existence of large sensor-specific, inter-person and intra-person variations. Results show that an accuracy of ~80% can be achieved across different users with different shoe sizes and ~85% for users with the same shoe size. A novel method, Dual-Check Till Consensus, is proposed to reduce the latency of gesture recognition from 2 seconds to 0.5 seconds and increase the accuracy to over 97%. This method offers a promising solution to achieve lower latency and higher accuracy at a minor cost of computation workload. The characteristics of high accuracy and fast classification of our method could lead to wider applications of using foot patterns for human-computer interaction.

[1]  Kosuke Sato,et al.  LazyNav: 3D ground navigation with non-critical body parts , 2015, 2015 IEEE Symposium on 3D User Interfaces (3DUI).

[2]  Gerhard Tröster,et al.  Robust Recognition of Reading Activity in Transit Using Wearable Electrooculography , 2009, Pervasive.

[3]  Antonio Krüger,et al.  FootStriker: A Wearable EMS-based Foot Strike Assistant for Running , 2017, ISS.

[4]  Dieter Kranzlmüller,et al.  ShoeSoleSense: proof of concept for a wearable foot interface for virtual and real environments , 2013, VRST '13.

[5]  Xiaoli Li,et al.  Deep Convolutional Neural Networks on Multichannel Time Series for Human Activity Recognition , 2015, IJCAI.

[6]  Mel Slater,et al.  Taking steps: the influence of a walking technique on presence in virtual reality , 1995, TCHI.

[7]  James Deaton,et al.  Development of a smart insole tracking system for physical therapy and athletics , 2014, PETRA.

[8]  Luca Turchet,et al.  A multimodal architecture for simulating natural interactive walking in virtual environments , 2011, PsychNology J..

[9]  Stefania Serafin,et al.  The Perceived Naturalness of Virtual Locomotion Methods Devoid of Explicit Leg Movements , 2013, MIG.

[10]  Vincent Hayward,et al.  Audio-tactile Display of Ground Properties Using Interactive Shoes , 2010, HAID.

[11]  Joaquim A. Jorge,et al.  A New Approach to Walking in Place , 2013, INTERACT.

[12]  Paul Lukowicz,et al.  All for one or one for all? Combining heterogeneous features for activity spotting , 2010, 2010 8th IEEE International Conference on Pervasive Computing and Communications Workshops (PERCOM Workshops).

[13]  Takeo Igarashi,et al.  Exploring Subtle Foot Plantar-based Gestures with Sock-placed Pressure Sensors , 2015, CHI.

[14]  Stefania Serafin,et al.  Walking in Place Through Virtual Worlds , 2016, HCI.

[15]  Hyuntae Cho Design and Implementation of a Lightweight Smart Insole for Gait Analysis , 2017, 2017 IEEE Trustcom/BigDataSE/ICESS.

[16]  Mary C. Whitton,et al.  Walking > walking-in-place > flying, in virtual environments , 1999, SIGGRAPH.

[17]  E. Langbehn,et al.  Evaluation of an Omnidirectional Walking-in-Place User Interface with Virtual Locomotion Speed Scaled by Forward Leaning Angle , 2015 .

[18]  Aaron F. Bobick,et al.  Parametric Hidden Markov Models for Gesture Recognition , 1999, IEEE Trans. Pattern Anal. Mach. Intell..

[19]  Arjan Kuijper,et al.  CapSoles: who is walking on what kind of floor? , 2017, MobileHCI.

[20]  Li Wei,et al.  Fast time series classification using numerosity reduction , 2006, ICML.

[21]  Hacer Yalim Keles,et al.  Action recognition using MHI based Hu moments with HMMs , 2017, IEEE EUROCON 2017 -17th International Conference on Smart Technologies.

[22]  Stefania Serafin,et al.  Natural Walking in Virtual Reality , 2018, Comput. Entertain..

[23]  Zhihan Lv,et al.  Multimodal Hand and Foot Gesture Interaction for Handheld Devices , 2013, TOMM.

[24]  Vincent Hayward,et al.  Preliminary Experiment Combining Virtual Reality Haptic Shoes and Audio Synthesis , 2010, EuroHaptics.

[25]  Maud Marchal,et al.  Joyman: A human-scale joystick for navigating in virtual worlds , 2011, 2011 IEEE Symposium on 3D User Interfaces (3DUI).

[26]  Zhichao Dong,et al.  Smooth assembled mappings for large-scale real walking , 2017, ACM Trans. Graph..

[27]  Jason Alexander,et al.  The Feet in Human--Computer Interaction , 2015, ACM Comput. Surv..

[28]  Michael J. Black,et al.  Deep inertial poser , 2018, ACM Trans. Graph..

[29]  Robert S. Allison,et al.  New simple virtual walking method – walking on the spot , 2004 .

[30]  KwangYun Wohn,et al.  The control of avatar motion using hand gesture , 1998, VRST '98.

[31]  Frits H. Post,et al.  Using the Wii Balance Board#8482; as a low-cost VR interaction device , 2008, VRST '08.

[32]  Sang-Youn Kim,et al.  RealWalk: Feeling Ground Surfaces While Walking in Virtual Reality , 2018, CHI Extended Abstracts.

[33]  Rama Chellappa,et al.  A hidden Markov model based framework for recognition of humans from gait sequences , 2003, Proceedings 2003 International Conference on Image Processing (Cat. No.03CH37429).

[34]  Zhihan Lv,et al.  Foot motion sensing: augmented game interface based on foot interaction for smartphone , 2014, CHI Extended Abstracts.

[35]  Christos Mousas,et al.  Full-Body Locomotion Reconstruction of Virtual Characters Using a Single Inertial Measurement Unit , 2017, Sensors.

[36]  Majid Sarrafzadeh,et al.  Smart insole: a wearable system for gait analysis , 2012, PETRA '12.

[37]  Luca Turchet,et al.  Sound Synthesis and Evaluation of Interactive Footsteps and Environmental Sounds Rendering for Virtual Reality Applications , 2011, IEEE Transactions on Visualization and Computer Graphics.

[38]  Mary C. Whitton,et al.  GUD WIP: Gait-Understanding-Driven Walking-In-Place , 2010, 2010 IEEE Virtual Reality Conference (VR).

[39]  Maud Marchal,et al.  Shake-your-head: revisiting walking-in-place for desktop virtual reality , 2010, VRST '10.

[40]  Tao Yu,et al.  HybridFusion: Real-Time Performance Capture Using a Single Depth Sensor and Sparse IMUs , 2018, ECCV.

[41]  Luca Turchet,et al.  Enhancing realism in virtual environments by simulating the audio-haptic sensation of walking on ground surfaces , 2012, 2012 IEEE Virtual Reality Workshops (VRW).

[42]  Eelke Folmer,et al.  Legomotion: scalable walking-based virtual locomotion , 2017, VRST.

[43]  Theophano Mitsa,et al.  Temporal Data Mining , 2010 .

[44]  Dinesh K. Pai,et al.  FootSee: an interactive animation system , 2003, SCA '03.

[45]  Patricia S. Denbrook,et al.  Virtual Locomotion: Walking in Place through Virtual Environments , 1999, Presence.

[46]  Min-Hyung Choi,et al.  VRInsole: An unobtrusive and immersive mobility training system for stroke rehabilitation , 2018, 2018 IEEE 15th International Conference on Wearable and Implantable Body Sensor Networks (BSN).

[47]  Robert W. Lindeman,et al.  Leaning-based travel interfaces revisited: frontal versus sidewise stances for flying in 3D virtual spaces , 2012, VRST '12.

[48]  Wenyao Xu,et al.  Smart Insole: A Wearable Sensor Device for Unobtrusive Gait Monitoring in Daily Life , 2016, IEEE Transactions on Industrial Informatics.

[49]  Gerhard Rigoll,et al.  Hidden Markov model based continuous online gesture recognition , 1998, Proceedings. Fourteenth International Conference on Pattern Recognition (Cat. No.98EX170).

[50]  Tuncay Cakmak,et al.  Cyberith virtualizer: a locomotion device for virtual reality , 2014, SIGGRAPH '14.

[51]  Hans-Werner Gellersen,et al.  Feet movement in desktop 3D interaction , 2014, 2014 IEEE Symposium on 3D User Interfaces (3DUI).

[52]  Miguel A. Labrador,et al.  A Survey on Human Activity Recognition using Wearable Sensors , 2013, IEEE Communications Surveys & Tutorials.

[53]  Robert E. Morley,et al.  In-shoe multisensory data acquisition system , 2001, IEEE Transactions on Biomedical Engineering.

[54]  Li-Yi Wei,et al.  Mapping virtual and physical reality , 2016, ACM Trans. Graph..

[55]  Mary C. Whitton,et al.  LLCM-WIP: Low-Latency, Continuous-Motion Walking-in-Place , 2008, 2008 IEEE Symposium on 3D User Interfaces.

[56]  Kosuke Sato,et al.  Extended LazyNav: Virtual 3D Ground Navigation for Large Displays and Head-Mounted Displays , 2017, IEEE Transactions on Visualization and Computer Graphics.

[57]  Stefania Serafin,et al.  A Threefold Approach for Precise and Efficient Locomotion in Virtual Environments with Varying Accessibility , 2018, 2018 IEEE Conference on Virtual Reality and 3D User Interfaces (VR).

[58]  Graham Arnold,et al.  Foot pressure differences in men and women. , 2010, Foot and ankle surgery : official journal of the European Society of Foot and Ankle Surgeons.

[59]  Joel Lanir,et al.  Better Understanding of Foot Gestures: An Elicitation Study , 2018, CHI.

[60]  K. Tumkur,et al.  Modeling Human Walking for Step Detection and Stride Determination by 3-Axis Accelerometer Readings in Pedometer , 2012, 2012 Fourth International Conference on Computational Intelligence, Modelling and Simulation.

[61]  Chung-Lin Huang,et al.  Hand gesture recognition using a real-time tracking method and hidden Markov models , 2003, Image Vis. Comput..

[62]  Michael Mock,et al.  A step counter service for Java-enabled devices using a built-in accelerometer , 2009, CAMS '09.

[63]  Mike Bailey,et al.  Virtual Reality for the Masses , 2014, IEEE Computer Graphics and Applications.

[64]  Ming-Chun Huang,et al.  An Energy-Efficient Adaptive Sensing Framework for Gait Monitoring Using Smart Insole , 2015, IEEE Sensors Journal.

[65]  Christos Mousas,et al.  Performance-Driven Hybrid Full-Body Character Control for Navigation and Interaction in Virtual Environments , 2017 .

[66]  Stefania Serafin,et al.  Tapping-In-Place: Increasing the naturalness of immersive walking-in-place locomotion through novel gestural input , 2013, 2013 IEEE Symposium on 3D User Interfaces (3DUI).

[67]  Ana Tajadura-Jiménez,et al.  As Light as your Footsteps: Altering Walking Sounds to Change Perceived Body Weight, Emotional State and Gait , 2015, CHI.

[68]  Hans-Werner Gellersen,et al.  Interactions Under the Desk: A Characterisation of Foot Movements for Input in a Seated Position , 2015, INTERACT.

[69]  Bodo Rosenhahn,et al.  Sparse Inertial Poser: Automatic 3D Human Pose Estimation from Sparse IMUs , 2017, Comput. Graph. Forum.

[70]  Bernhard E. Riecke,et al.  Comparing leaning-based motion cueing interfaces for virtual reality locomotion , 2017, 2017 IEEE Symposium on 3D User Interfaces (3DUI).