Multi-user augmented reality with communication efficient and spatially consistent virtual objects

Multi-user augmented reality (AR), where multiple co-located users view a common set of virtual objects, is becoming increasingly popular. For example, Google Just a Line allows multiple users to draw virtual graffiti in the same physical space. Multi-user AR requires network communications in order to coordinate the positions of the virtual objects on each user's display, yet there is currently little understanding of how such apps communicate. In this work, we address this key gap in knowledge by showing that the communicated data directly impacts the latency and positioning of the virtual objects rendered on the users' displays. We develop solutions to these problems that we find along three facets: (1) efficient communication strategies that trade off communication latency for spatial consistency of the virtual objects; (2) a new metric that enables mobile AR devices to update their virtual objects as they move around and observe more of the scene; and (3) a tool to automatically quantify how much the virtual objects' positions inadvertently change in time and space. Our evaluation is performed on Android smartphones running open-source AR. The results show that our system, SPAR, can decrease the latency by up to 55%, while decreasing the spatial inconsistency by up to 60%, compared to baseline methods.

[1]  Renaud Dubé,et al.  An online multi-robot SLAM system for 3D LiDARs , 2017, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[2]  Danping Zou,et al.  CoSLAM: Collaborative Visual SLAM in Dynamic Environments , 2013, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[3]  Kittipat Apicharttrisorn,et al.  Frugal following: power thrifty object detection and tracking for mobile augmented reality , 2019, SenSys.

[4]  Francisco José Madrid-Cuevas,et al.  Automatic generation and detection of highly reliable fiducial markers under occlusion , 2014, Pattern Recognit..

[5]  Qiang Liu,et al.  DARE: Dynamic Adaptive Mobile Augmented Reality with Edge Computing , 2018, 2018 IEEE 26th International Conference on Network Protocols (ICNP).

[6]  Karthik Dantu,et al.  Edge-SLAM: edge-assisted visual simultaneous localization and mapping , 2020, MobiSys.

[7]  Fakhri Karray,et al.  Multi-robot SLAM: An Overview and Quantitative Evaluation of MRGS ROS Framework for MR-SLAM , 2017, RiTA.

[8]  Feng Qian,et al.  CARS: Collaborative Augmented Reality for Socialization , 2018, HotMobile.

[9]  Mahadev Satyanarayanan,et al.  Towards wearable cognitive assistance , 2014, MobiSys.

[10]  Alex Zelinsky,et al.  Learning OpenCV---Computer Vision with the OpenCV Library (Bradski, G.R. et al.; 2008)[On the Shelf] , 2009, IEEE Robotics & Automation Magazine.

[11]  Carlee Joe-Wong,et al.  CollabAR: Edge-assisted Collaborative Image Recognition for Mobile Augmented Reality , 2020, 2020 19th ACM/IEEE International Conference on Information Processing in Sensor Networks (IPSN).

[12]  Justin Manweiler,et al.  OverLay: Practical Mobile Augmented Reality , 2015, MobiSys.

[13]  Rafael Muñoz-Salinas,et al.  Speeded up detection of squared fiducial markers , 2018, Image Vis. Comput..

[14]  Marco Gruteser,et al.  Edge Assisted Real-time Object Detection for Mobile Augmented Reality , 2019, MobiCom.

[15]  Zhenming Liu,et al.  DeepDecision: A Mobile Deep Learning Framework for Edge Video Analytics , 2018, IEEE INFOCOM 2018 - IEEE Conference on Computer Communications.

[16]  Michael Gassner,et al.  SVO: Semidirect Visual Odometry for Monocular and Multicamera Systems , 2017, IEEE Transactions on Robotics.

[17]  Kyoung Shin Park,et al.  Effects of network characteristics on human performance in a collaborative virtual environment , 1999, Proceedings IEEE Virtual Reality (Cat. No. 99CB36316).

[18]  Yubin Kuang,et al.  Revisiting the PnP Problem: A Fast, General and Optimal Solution , 2013, 2013 IEEE International Conference on Computer Vision.

[19]  Gary R. Bradski,et al.  Learning OpenCV 3: Computer Vision in C++ with the OpenCV Library , 2016 .

[20]  Richard L. Holloway,et al.  Registration errors in augmented reality systems , 1996 .

[21]  Guohui Wang,et al.  Practical Urban Localization for Mobile AR , 2020, HotMobile.

[22]  Randy H. Katz,et al.  MARVEL: Enabling Mobile Augmented Reality with Low Energy and Low Latency , 2018, SenSys.

[23]  Rittwik Jana,et al.  Characterization of Multi-User Augmented Reality over Cellular Networks , 2020, 2020 17th Annual IEEE International Conference on Sensing, Communication, and Networking (SECON).

[24]  Wolfram Burgard,et al.  A benchmark for the evaluation of RGB-D SLAM systems , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[25]  Bo Han,et al.  Jaguar: Low Latency Mobile Augmented Reality with Flexible Tracking , 2018, ACM Multimedia.

[26]  Feng Zheng,et al.  Spatio-temporal registration in augmented reality , 2015 .

[27]  Juan D. Tardós,et al.  ORB-SLAM2: An Open-Source SLAM System for Monocular, Stereo, and RGB-D Cameras , 2016, IEEE Transactions on Robotics.

[28]  Andrew Howard,et al.  Multi-robot Simultaneous Localization and Mapping using Particle Filters , 2005, Int. J. Robotics Res..

[29]  Bo Han,et al.  Low Latency Mobile Augmented Reality with Flexible Tracking , 2018, MobiCom.

[30]  Hujun Bao,et al.  Survey and evaluation of monocular visual-inertial SLAM algorithms for augmented reality , 2019, Virtual Real. Intell. Hardw..

[31]  Patrik Schmuck,et al.  Multi-UAV collaborative monocular SLAM , 2017, 2017 IEEE International Conference on Robotics and Automation (ICRA).

[32]  Shaojie Shen,et al.  VINS-Mono: A Robust and Versatile Monocular Visual-Inertial State Estimator , 2017, IEEE Transactions on Robotics.

[33]  Ramesh Govindan,et al.  CarMap: Fast 3D Feature Map Updates for Automobiles , 2020, NSDI.

[34]  Syed Riaz un Nabi Jafri,et al.  A distributed multi robot SLAM system for environment learning , 2013, 2013 IEEE Workshop on Robotic Intelligence in Informationally Structured Space (RiiSS).

[35]  Justin Manweiler,et al.  Low Bandwidth Offload for Mobile AR , 2016, CoNEXT.

[36]  Mahadev Satyanarayanan,et al.  An empirical study of latency in an emerging class of edge computing applications for wearable cognitive assistance , 2017, SEC.

[37]  Hugh F. Durrant-Whyte,et al.  Simultaneous localization and mapping: part I , 2006, IEEE Robotics & Automation Magazine.

[38]  Paramvir Bahl,et al.  Glimpse: Continuous, Real-Time Object Recognition on Mobile Devices , 2015, SenSys.

[39]  Shaojie Shen,et al.  Monocular Visual-Inertial State Estimation for Mobile Augmented Reality , 2017, 2017 IEEE International Symposium on Mixed and Augmented Reality (ISMAR).

[40]  Robert LiKamWa,et al.  GLEAM: An Illumination Estimation Framework for Real-time Photorealistic Augmented Reality on Mobile Devices , 2019, MobiSys.

[41]  Ralf Dörner,et al.  Accuracy in optical tracking with fiducial markers: an accuracy function for ARToolKit , 2004, Third IEEE and ACM International Symposium on Mixed and Augmented Reality.

[42]  Dorian Gálvez-López,et al.  Bags of Binary Words for Fast Place Recognition in Image Sequences , 2012, IEEE Transactions on Robotics.

[43]  Maria Gorlatova,et al.  ShareAR: Communication-Efficient Multi-User Mobile Augmented Reality , 2019, HotNets.