The effects of spatial auditory and visual cues on mixed reality remote collaboration

Collaborative Mixed Reality (MR) technologies enable remote people to work together by sharing communication cues intrinsic to face-to-face conversations, such as eye gaze and hand gestures. While the role of visual cues has been investigated in many collaborative MR systems, the use of spatial auditory cues remains underexplored. In this paper, we present an MR remote collaboration system that shares both spatial auditory and visual cues between collaborators to help them complete a search task. Through two user studies in a large office, we found that compared to non-spatialized audio, the spatialized remote expert’s voice and auditory beacons enabled local workers to find small occluded objects with significantly stronger spatial perception. We also found that while the spatial auditory cues could indicate the spatial layout and a general direction to search for the target object, visual head frustum and hand gestures intuitively demonstrated the remote expert’s movements and the position of the target. Integrating visual cues (especially the head frustum) with the spatial auditory cues significantly improved the local worker’s task performance, social presence, and spatial perception of the environment.

[1]  Stephen Brewster,et al.  Non-visual interfaces for wearable computers , 2000 .

[2]  Lei Gao,et al.  Real-time Visual Representations for Mixed Reality Remote Collaboration , 2017, ICAT-EGVE.

[3]  Chris Schmandt,et al.  Nomadic radio: speech and audio interaction for contextual messaging in nomadic environments , 2000, TCHI.

[4]  Tobias Höllerer,et al.  World-stabilized annotations and virtual scene navigation for remote collaboration , 2014, UIST.

[5]  Titus Jia Jie Tang An assistive EyeWear prototype that interactively converts 3D object locations into spatial audio , 2014, ISWC '14 Adjunct.

[6]  Tapio Lokki,et al.  Augmented reality audio for mobile and wearable appliances , 2004 .

[7]  Matt Adcock,et al.  Hand gestures and visual annotation in live 360 panorama-based mixed reality remote collaboration , 2018, OZCHI.

[8]  William S. Helton,et al.  Attention Redirection Using Binaurally Spatialised Cues Delivered Over a Bone Conduction Headset , 2016 .

[9]  Roberta L. Klatzky,et al.  Navigation System for the Blind: Auditory Display Modes and Guidance , 1998, Presence.

[10]  Lei Gao,et al.  Real-time visual representations for mobile mixed reality remote collaboration , 2018, SIGGRAPH 2018.

[11]  Juan Liu,et al.  Downsizing: The Effect of Mixed-Reality Person Representations on Stress and Presence in Telecommunication , 2018, 2018 IEEE International Conference on Artificial Intelligence and Virtual Reality (AIVR).

[12]  Matti Karjalainen,et al.  AN AUGMENTED REALITY AUDIO HEADSET , 2008 .

[13]  Keita Higuchi,et al.  Can Eye Help You?: Effects of Visualizing Eye Fixations on Remote Collaboration Scenarios for Physical Tasks , 2016, CHI.

[14]  Mark Billinghurst,et al.  Improving co-presence with augmented visual communication cues for sharing experience through video conference , 2014, 2014 IEEE International Symposium on Mixed and Augmented Reality (ISMAR).

[15]  John C. Tang,et al.  3D Collaboration Method over HoloLens™ and Skype™ End Points , 2015, ImmersiveME@ACM Multimedia.

[16]  David McGookin,et al.  Diary in the Sky: A Spatial Audio Display for a Mobile Calendar , 2001, BCS HCI/IHM.

[17]  Maciej Tomczak,et al.  The need to report effect size estimates revisited. An overview of some recommended measures of effect size , 2014 .

[18]  Huiyang Li,et al.  Effects of HoloLens in Collaboration: A Case in Navigation Tasks , 2017 .

[19]  J. B. Brooke,et al.  SUS: A 'Quick and Dirty' Usability Scale , 1996 .

[20]  Leila Alem,et al.  A Study of Gestures in a Video-Mediated Collaborative Assembly Task , 2011, Adv. Hum. Comput. Interact..

[21]  F. Biocca,et al.  Internal Consistency and Reliability of the Networked MindsMeasure of Social Presence , 2004 .

[22]  Mark Billinghurst,et al.  Do You See What I See? The Effect of Gaze Tracking on Task Space Remote Collaboration , 2016, IEEE Transactions on Visualization and Computer Graphics.

[23]  Hirokazu Kato,et al.  Collaborative augmented reality , 2002, CACM.

[24]  Julian Villegas,et al.  “GABRIEL”: Geo-Aware BRoadcasting for In-Vehicle Entertainment and Localizability , 2010 .

[25]  Robert E. Kraut,et al.  Coordination of communication: effects of shared visual context on collaborative work , 2000, CSCW '00.

[26]  Weidong Huang,et al.  Sharing hand gesture and sketch cues in remote collaboration , 2019, J. Vis. Commun. Image Represent..

[27]  Andreas Bulling,et al.  On the interplay between spontaneous spoken instructions and human visual behaviour in an indoor guidance task , 2015, CogSci.

[28]  Michael Neff,et al.  Communication Behavior in Embodied Virtual Reality , 2018, CHI.

[29]  Daniel Mendes,et al.  Warping Deixis: Distorting Gestures to Enhance Collaboration , 2019, CHI.

[30]  Mark Billinghurst,et al.  A wearable spatial conferencing space , 1998, Digest of Papers. Second International Symposium on Wearable Computers (Cat. No.98EX215).

[31]  Weidong Huang,et al.  3D helping hands: a gesture based MR system for remote collaboration , 2012, VRCAI '12.

[32]  Luc Van Gool,et al.  Blue-c: a spatially immersive display and 3D video portal for telepresence , 2003, IPT/EGVE.

[33]  Stephen A. Brewster,et al.  Spatial audio in small screen device displays , 2000, Personal Technologies.

[34]  Jing Yang,et al.  Spatial Audio for Human-Object Interactions in Small AR Workspaces , 2018, MobiSys.

[35]  Stuart Anderson,et al.  RemoteFusion: real time depth camera fusion for remote collaboration on physical tasks , 2013, VRCAI '13.

[36]  J. Helmert,et al.  Limitations of gaze transfer: without visual context, eye movements do not to help to coordinate joint action, whereas mouse movements do. , 2014, Acta psychologica.

[37]  Charles T. Loop,et al.  Holoportation: Virtual 3D Teleportation in Real-time , 2016, UIST.

[38]  Hiroshi Ishii,et al.  Kinected conference: augmenting video imaging with calibrated depth and audio , 2011, CSCW.

[39]  Bill Buxton,et al.  EuroPARC's integrated interactive intermedia facility (IIIF): early experiences , 1990 .

[40]  Mark Billinghurst,et al.  A User Study on MR Remote Collaboration Using Live 360 Video , 2018, 2018 IEEE International Symposium on Mixed and Augmented Reality (ISMAR).

[41]  Robert W. Lindeman,et al.  Mini-Me: An Adaptive Avatar for Mixed Reality Remote Collaboration , 2018, CHI.

[42]  Bernd Fröhlich,et al.  Immersive Group-to-Group Telepresence , 2013, IEEE Transactions on Visualization and Computer Graphics.

[43]  Mark Billinghurst,et al.  Mixed reality collaboration through sharing a live panorama , 2017, SIGGRAPH ASIA Mobile Graphics and Interactive Applications.

[44]  J. B. Brooke,et al.  SUS: a retrospective , 2013 .

[45]  Mark Billinghurst,et al.  Spatial sound localization in an augmented reality environment , 2006, OZCHI.

[46]  Jing Yang,et al.  A User Study on Mixed Reality Remote Collaboration with Eye Gaze and Hand Gesture Sharing , 2020, CHI.

[47]  Antonio Criminisi,et al.  Improving Indoor Mobility of the Visually Impaired with Depth-Based Spatial Sound , 2015, 2015 IEEE International Conference on Computer Vision Workshop (ICCVW).

[48]  Mark Billinghurst,et al.  Study of augmented gesture communication cues and view sharing in remote collaboration , 2013, 2013 IEEE International Symposium on Mixed and Augmented Reality (ISMAR).

[49]  Jing Yang,et al.  Hearing Is Believing: Synthesizing Spatial Audio from Everyday Objects to Users , 2019, AH.

[50]  Matt Adcock,et al.  Mixed Reality Remote Collaboration Combining 360 Video and 3D Reconstruction , 2019, CHI.