Hip-directed walking-in-place using a single depth camera

Abstract Walking-in-place (WIP) is a locomotion technique that allows users to travel in virtual environments (VEs) without significantly changing their physical position on the floor. Hip-directed steering(HDS) is a novel physical technique for controlling direction changes in virtual travel using hip movements. We present an WIP-based navigation approach for controlling locomotion in VEs that combines the speed and direction in a scenario similar to a domestic setup in which people interact with a flat screen. Their physical motion data are captured by one depth camera properly aligned with the screen and oriented toward the user. We approach the characteristically noisy data generated by depth cameras via a user study to determine both the range of values and their robustness from the motion data associated with the joints relevant to WIP speed(knee, ankle and foot) and HDS (spine, hip and shoulder) to derive a reliable technique. Our WIP speed method is supported in a simple vocabulary of five different footstep types. Experimental results show that both the knee and hip provide the most robust data. We evaluated our techniques via usability tests exercising common locomotion tasks. The results show that users liked both the speed control and comfort afforded by our speed method. Regarding HDS, users reported that the angular-based method allowed them to travel faster and was both more controllable and easier to learn than the time-based method. Our work shows that a single depth camera can be used to combine locomotion and direction control in a simple and affordable setup.

[1]  Dietmar Bauer,et al.  Hands-Free Navigation in Immersive Environments for the Evaluation of the Effectiveness of Indoor Navigation Systems , 2012, J. Virtual Real. Broadcast..

[2]  Jason Alexander,et al.  The Feet in Human--Computer Interaction , 2015, ACM Comput. Surv..

[3]  Ryan P. McMahan,et al.  Shadow walking: An unencumbered locomotion technique for systems with under-floor projection , 2011, 2011 IEEE Virtual Reality Conference.

[4]  Makoto Sato,et al.  Virtual Locomotion Interface with Ground Surface Simulation , 2003, ICAT.

[5]  Stefania Serafin,et al.  The influence of step frequency on the range of perceptually natural visual walking speeds during walking-in-place and treadmill locomotion , 2014, VRST '14.

[6]  Victoria Interrante,et al.  Seven League Boots: A New Metaphor for Augmented Locomotion through Moderately Large Scale Immersive Virtual Environments , 2007, 2007 IEEE Symposium on 3D User Interfaces.

[7]  Patricia S. Denbrook,et al.  Virtual Locomotion: Walking in Place through Virtual Environments , 1999, Presence.

[8]  Béat Hirsbrunner,et al.  Active Walking Interface for Human-Scale Virtual Environment , 2005 .

[9]  Sharif Razzaque,et al.  Redirected Walking in Place , 2002, EGVE.

[10]  Maud Marchal,et al.  Shake-your-head: revisiting walking-in-place for desktop virtual reality , 2010, VRST '10.

[11]  Alberto Barbosa Raposo,et al.  A Study of Navigation and Selection Techniques in Virtual Environments Using Microsoft Kinect® , 2013, HCI.

[12]  Stefania Serafin,et al.  Establishing the Range of Perceptually Natural Visual Walking Speeds for Virtual Walking-In-Place Locomotion , 2014, IEEE Transactions on Visualization and Computer Graphics.

[13]  Mary C. Whitton,et al.  LLCM-WIP: Low-Latency, Continuous-Motion Walking-in-Place , 2008, 2008 IEEE Symposium on 3D User Interfaces.

[14]  Nathan Clark,et al.  Implementation of interactive arm playback behaviors of social robot Zeno for autism spectrum disorder therapy , 2012, PETRA '12.

[15]  Timothy P. McNamara,et al.  Updating orientation in large virtual environments using scaled translational gain , 2006, APGV '06.

[16]  Michèle Courant,et al.  Walking-pad: a step-in-place locomotion interface for virtual environments , 2004, ICMI '04.

[17]  Vassilis-Javed Khan,et al.  Efficient navigation in virtual environments: A comparative study of two interaction techniques: The Magic Wand vs. the Human Joystick , 2012, 2012 4th International Conference on Intelligent Human Computer Interaction (IHCI).

[18]  James N. Templeman,et al.  Immersive Simulation to Train Urban Infantry Combat , 2006 .

[19]  J. Gaber,et al.  Collision Avatar (CA): Adding collision objects for human body in augmented reality using Kinect , 2012, 2012 6th International Conference on Application of Information and Communication Technologies (AICT).

[20]  Ye Zheng,et al.  Torso versus gaze direction to navigate a VE by walking in place , 2013, SAP.

[21]  Mary C. Whitton,et al.  GUD WIP: Gait-Understanding-Driven Walking-In-Place , 2010, 2010 IEEE Virtual Reality Conference (VR).

[22]  Mel Slater,et al.  Taking steps: the influence of a walking technique on presence in virtual reality , 1995, TCHI.

[23]  Ye Zheng,et al.  Exploring a virtual environment by walking in place using the Microsoft Kinect , 2012, SAP.

[24]  Rita Francese,et al.  Wiimote and Kinect: gestural user interfaces add a natural third dimension to HCI , 2012, AVI.

[25]  Betsy Williams Sanders,et al.  Evaluation of walking in place on a Wii balance board to explore a virtual environment , 2011, TAP.

[26]  Joaquim A. Jorge,et al.  A New Approach to Walking in Place , 2013, INTERACT.

[27]  Samiul Monir,et al.  Rotation and scale invariant posture recognition using Microsoft Kinect skeletal tracking feature , 2012, 2012 12th International Conference on Intelligent Systems Design and Applications (ISDA).

[28]  Jessica K. Hodgins,et al.  Accelerometer-based user interfaces for the control of a physically simulated character , 2008, SIGGRAPH 2008.