Towards Augmented Reality-driven Human-City Interaction: Current Research and Future Challenges

Interaction design for Augmented Reality (AR) is gaining increasing attention from both academia and industry. This survey discusses 205 articles (75% of articles published between 2015 - 2019) to review the field of human interaction in connected cities with emphasis on augmented reality-driven interaction. We provide an overview of Human-City Interaction and related technological approaches, followed by a review of the latest trends of information visualization, constrained interfaces, and embodied interaction for AR headsets. We highlight under-explored issues in interface design and input techniques that warrant further research, and conjecture that AR with complementary Conversational User Interfaces (CUIs) is a key enabler for ubiquitous interaction with immersive systems in smart cities. Our work helps researchers understand the current potential and future needs of AR in Human-City Interaction.

[1]  Doug A. Bowman,et al.  Walking with adaptive augmented reality workspaces: design and usage patterns , 2019, IUI.

[2]  Xiang 'Anthony' Chen,et al.  WrisText: One-handed Text Entry on Smartwatch using Wrist Gestures , 2018, CHI.

[3]  Alessandro Bozzon,et al.  Chatterbox: Conversational Interfaces for Microtask Crowdsourcing , 2019, UMAP.

[4]  Jae Hoon Shin,et al.  On-wheel finger gesture control for in-vehicle systems on central consoles , 2015, AutomotiveUI.

[5]  Tsutomu Terada,et al.  An information layout method for an optical see-through head mounted display focusing on the viewability , 2008, 2008 7th IEEE/ACM International Symposium on Mixed and Augmented Reality.

[6]  Yu-Te Liao,et al.  A single-pixel wireless contact lens display , 2011, Journal of Micromechanics and Microengineering.

[7]  Ernst Kruijff,et al.  Non-Visual Cues for View Management in Narrow Field of View Augmented Reality Displays , 2019, 2019 IEEE International Symposium on Mixed and Augmented Reality (ISMAR).

[8]  Shamama Anwar,et al.  Hand Gesture Recognition: A Survey , 2018, Nanoelectronics, Circuits and Communication Systems.

[9]  Geehyuk Lee,et al.  Typing on a Smartwatch for Smart Glasses , 2017, ISS.

[10]  Bing-Yu Chen,et al.  InDexMo: exploring finger-worn RFID motion tracking for activity recognition on tagged objects , 2019, UbiComp.

[11]  Yang Zhang,et al.  ActiTouch: Robust Touch Detection for On-Skin AR/VR Interfaces , 2019, UIST.

[12]  Christian Holz,et al.  DuoSkin: rapidly prototyping on-skin user interfaces using skin-friendly materials , 2016, SEMWEB.

[13]  Alessandro Bogliolo,et al.  The Rise of Bots: A Survey of Conversational Interfaces, Patterns, and Paradigms , 2017, Conference on Designing Interactive Systems.

[14]  Marko Jurmu,et al.  Evaluation Probes , 2015, CHI.

[15]  Jussi Kangasharju,et al.  Enhanced Augmented Reality Applications in Vehicle-to-Edge Networks , 2019, 2019 22nd Conference on Innovation in Clouds, Internet and Networks and Workshops (ICIN).

[16]  Mark Billinghurst,et al.  Pinpointing: Precise Head- and Eye-Based Target Selection for Augmented Reality , 2018, CHI.

[17]  Joseph R. Keebler,et al.  Toward an Understanding of Trust Repair in Human-Robot Interaction , 2018, ACM Trans. Interact. Intell. Syst..

[18]  Jürgen Steimle,et al.  DeformWear , 2017, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[19]  Ivan Poupyrev,et al.  Soli , 2016, ACM Trans. Graph..

[20]  Isaac Wang,et al.  Exploring Virtual Agents for Augmented Reality , 2019, CHI.

[21]  Carlos Toxtli,et al.  Understanding Chatbot-mediated Task Management , 2018, CHI.

[22]  Alaric Hamacher,et al.  Augmented Reality User Interface Evaluation - Performance Measurement of Hololens, Moverio and Mouse Input , 2019, Int. J. Interact. Mob. Technol..

[23]  Hongan Wang,et al.  Pactolus: A Method for Mid-Air Gesture Segmentation within EMG , 2016, CHI Extended Abstracts.

[24]  Andreas A. Veglis,et al.  Embedding a chatbot in a news article: design and implementation , 2019, PCI.

[25]  Yong Gu Ji,et al.  Complexity Overloaded in Smart Car: How to Measure Complexity of In-vehicle Displays and Controls? , 2016, AutomotiveUI.

[26]  Mohamed Soliman,et al.  FingerInput: Capturing Expressive Single-Hand Thumb-to-Finger Microgestures , 2018, ISS.

[27]  Peter Kazanzides,et al.  Restoring the Awareness in the Occluded Visual Field for Optical See-Through Head-Mounted Displays , 2018, IEEE Transactions on Visualization and Computer Graphics.

[28]  Kentaro Go,et al.  Computational Alleviation of Homonymous Visual Field Defect with OST-HMD: The Effect of Size and Position of Overlaid Overview Window , 2019, 2019 International Conference on Cyberworlds (CW).

[29]  Byungjoo Lee,et al.  Geometrically Compensating Effect of End-to-End Latency in Moving-Target Selection Games , 2019, CHI.

[30]  Aaron J. Quigley,et al.  Factors influencing visual attention switch in multi-display user interfaces: a survey , 2012, PerDis.

[31]  Xincheng Li,et al.  One-Dimensional Handwriting: Inputting Letters and Words on Smart Glasses , 2016, CHI.

[32]  Antti Jylhä,et al.  Designing a Willing-to-Use-in-Public Hand Gestural Interaction Technique for Smart Glasses , 2016, CHI.

[33]  David Coyle,et al.  Empirical Evidence for a Diminished Sense of Agency in Speech Interfaces , 2015, CHI.

[34]  Geehyuk Lee,et al.  Gaze-Assisted Typing for Smart Glasses , 2019, UIST.

[35]  Per Ola Kristensson,et al.  An Evaluation of Discrete and Continuous Mid-Air Loop and Marking Menu Selection in Optical See-Through HMDs , 2019, MobileHCI.

[36]  Wendy A. Kellogg,et al.  "I'd be overwhelmed, but it's just one more thing to do": availability and interruption in research management , 2002, CHI.

[37]  Pan Hui,et al.  Peripheral vision: a new killer app for smart glasses , 2019, IUI.

[38]  Mario Di Francesco,et al.  Foraging-based optimization of pervasive displays , 2019, Pervasive Mob. Comput..

[39]  Jussi Kangasharju,et al.  ARVE: Augmented Reality Applications in Vehicle to Edge Networks , 2018, MECOMM@SIGCOMM.

[40]  John Paulin Hansen,et al.  Gaze typing in virtual reality: impact of keyboard design, selection method, and motion , 2018, ETRA.

[41]  Yang Zhang,et al.  LumiWatch: On-Arm Projected Graphics and Touch Input , 2018, CHI.

[42]  Cristina V. Lopes,et al.  Optical illusion in augmented reality , 2012, VRST '12.

[43]  Xiaojun Bi,et al.  COMPASS: Rotational Keyboard on Non-Touch Smartwatches , 2017, CHI.

[44]  Xin Liu,et al.  Body-Borne Computers as Extensions of Self , 2017, Comput..

[45]  Yvonne Rogers,et al.  Introduction to Human-Building Interaction (HBI) , 2019, ACM Trans. Comput. Hum. Interact..

[46]  Desney S. Tan,et al.  Humantenna: using the body as an antenna for real-time whole-body interaction , 2012, CHI.

[47]  Niels Henze,et al.  Notification in VR: The Effect of Notification Placement, Task and Environment , 2019, CHI PLAY.

[48]  Barbara S. Chaparro,et al.  Text Input Performance with a Mixed Reality Head-Mounted Display (HMD) , 2019 .

[49]  Mahbub Hassan,et al.  A Survey of Wearable Devices and Challenges , 2017, IEEE Communications Surveys & Tutorials.

[50]  Timothy L. Brown,et al.  Examining the interaction between timing and modality in forward collision warnings , 2015, AutomotiveUI.

[51]  Florian Alt,et al.  Should I Interrupt or Not?: Understanding Interruptions in Head-Mounted Display Settings , 2019, Conference on Designing Interactive Systems.

[52]  James A. Landay,et al.  Comparing Speech and Keyboard Text Entry for Short Messages in Two Languages on Touchscreen Phones , 2016, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[53]  Kiyoshi Kiyokawa,et al.  Dynamic text management for see-through wearable and heads-up display systems , 2013, IUI '13.

[54]  Orit Shaer,et al.  Reality-based interaction: a framework for post-WIMP interfaces , 2008, CHI.

[55]  Da-Yuan Huang,et al.  DigitSpace: Designing Thumb-to-Fingers Touch Interfaces for One-Handed and Eyes-Free Interactions , 2016, CHI.

[56]  Sean Gustafson,et al.  PinchWatch: A Wearable Device for One-Handed Microinteractions , 2010 .

[57]  Jürgen Steimle,et al.  Tacttoo: A Thin and Feel-Through Tattoo for On-Skin Tactile Output , 2018, UIST.

[58]  Andreea Danielescu,et al.  A Bot is Not a Polyglot: Designing Personalities for Multi-Lingual Conversational Agents , 2018, CHI Extended Abstracts.

[59]  Per Ola Kristensson,et al.  I did that! Measuring users' experience of agency in their own actions , 2012, CHI 2012.

[60]  Pan Hui,et al.  Interaction Methods for Smart Glasses: A Survey , 2017, IEEE Access.

[61]  Shuai Liu,et al.  SKIN+: Fabricating Soft Fluidic User Interfaces for Enhancing On-Skin Experiences and Interactions , 2018, CHI Extended Abstracts.

[62]  Ann Morrison,et al.  Like bees around the hive: a comparative study of a mobile augmented reality map , 2009, CHI.

[63]  Thomas R. Gross,et al.  Designing Groundless Body Channel Communication Systems: Performance and Implications , 2018, UIST.

[64]  Kening Zhu,et al.  FingerT9: Leveraging Thumb-to-finger Interaction for Same-side-hand Text Entry on Smartwatches , 2018, CHI.

[65]  Tsutomu Terada,et al.  Readability and legibility of fonts considering shakiness of head mounted displays , 2019, UbiComp.

[66]  Catherine Pelachaud,et al.  Skin-On Interfaces: A Bio-Driven Approach for Artificial Skin Design to Cover Interactive Devices , 2019, UIST.

[67]  Neda Mohammadi,et al.  Smart city digital twins , 2017, 2017 IEEE Symposium Series on Computational Intelligence (SSCI).

[68]  Mubbasir Kapadia,et al.  Towards a Conversational Interface for Authoring Intelligent Virtual Characters , 2019, IVA.

[69]  Andreas Schreiber,et al.  Adopting Conversational Interfaces for Exploring OSGi-Based Software Architectures in Augmented Reality , 2019, 2019 IEEE/ACM 1st International Workshop on Bots in Software Engineering (BotSE).

[70]  Matti Siekkinen,et al.  Multi-carrier Measurement Study of Mobile Network Latency: The Tale of Hong Kong and Helsinki , 2019, 2019 15th International Conference on Mobile Ad-Hoc and Sensor Networks (MSN).

[71]  Sung-Uk Jung,et al.  HandPoseMenu: Hand Posture-Based Virtual Menus for Changing Interaction Mode in 3D Space , 2019, ISS.

[72]  Dongyang Zhao,et al.  BiTipText: Bimanual Eyes-Free Text Entry on a Fingertip Keyboard , 2020, CHI.

[73]  Branislav Kveton,et al.  Minimal Interaction Content Discovery in Recommender Systems , 2016, TIIS.

[74]  Brian P. Bailey,et al.  Investigating the effectiveness of mental workload as a predictor of opportune moments for interruption , 2005, CHI Extended Abstracts.

[75]  Kenji Suzuki,et al.  EnhancedTouchX: Smart Bracelets for Augmenting Interpersonal Touch Interactions , 2019, CHI.

[76]  Hui-Shyong Yeo,et al.  Exploring Tangible Interactions with Radar Sensing , 2018, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[77]  Pan Hui,et al.  M2A: A Framework for Visualizing Information from Mobile Web to Mobile Augmented Reality , 2019, 2019 IEEE International Conference on Pervasive Computing and Communications (PerCom.

[78]  David Lindlbauer,et al.  Context-Aware Online Adaptation of Mixed Reality Interfaces , 2019, UIST.

[79]  Benjamin R. Cowan,et al.  What Makes a Good Conversation?: Challenges in Designing Truly Conversational Agents , 2019, CHI.

[80]  Sangki Yun,et al.  When IoT met Augmented Reality: Visualizing the Source of the Wireless Signal in AR View , 2019, MobiSys.

[81]  Shumin Zhai,et al.  Typing on an Invisible Keyboard , 2018, CHI.

[82]  Tao Mei,et al.  Video ChatBot: Triggering Live Social Interactions by Automatic Video Commenting , 2016, ACM Multimedia.

[83]  Oliver Brdiczka,et al.  Learning to detect user activity and availability from a variety of sensor data , 2004, Second IEEE Annual Conference on Pervasive Computing and Communications, 2004. Proceedings of the.

[84]  Pascale Fung,et al.  Exploring Perceived Emotional Intelligence of Personality-Driven Virtual Agents in Handling User Challenges , 2019, WWW.

[85]  Sanjay E. Sarma,et al.  MagicHand: Interact with IoT Devices in Augmented Reality Environment , 2019, 2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR).

[86]  Elena Mugellini,et al.  Tangible Meets Gestural: Comparing and Blending Post-WIMP Interaction Paradigms , 2015, Tangible and Embedded Interaction.

[87]  Ben Shneiderman,et al.  The limits of speech recognition , 2000, CACM.

[88]  Anant Bhaskar Garg,et al.  Embodied Cognition, Human Computer Interaction, and Application Areas , 2012 .

[89]  Joseph L. Gabbard,et al.  Augmented reality head-up displays effect on drivers’ spatial knowledge acquisition , 2019 .

[90]  Eric Paulos,et al.  Skintillates: Designing and Creating Epidermal Interactions , 2016, Conference on Designing Interactive Systems.

[91]  Ahmed Sabbir Arif,et al.  Context-sensitive app prediction on the suggestion bar of a mobile keyboard , 2019, MUM.

[92]  Norbert Reithinger,et al.  Conversation is multimodal: thus conversational user interfaces should be as well , 2019, CUI.

[93]  Elizabeth Gerber,et al.  Conversational Agents: Acting on the Wave of Research and Development , 2019, CHI Extended Abstracts.

[94]  Kyu-Han Kim,et al.  TypingRing: A Wearable Ring Platform for Text Input , 2015, MobiSys.

[95]  Robert Xiao,et al.  MRTouch: Adding Touch Input to Head-Mounted Mixed Reality , 2018, IEEE Transactions on Visualization and Computer Graphics.

[96]  Shwetak N. Patel,et al.  AuraRing , 2019, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[97]  Tracy Anne Hammond,et al.  Developing a Hand Gesture Recognition System for Mapping Symbolic Hand Gestures to Analogous Emojis in Computer-Mediated Communication , 2019, ACM Trans. Interact. Intell. Syst..

[98]  Pan Hui,et al.  HIBEY: Hide the Keyboard in Augmented Reality , 2019, 2019 IEEE International Conference on Pervasive Computing and Communications (PerCom.

[99]  Paul Strohmeier,et al.  Invisiboard: maximizing display and input space with a full screen text entry method for smartwatches , 2016, MobileHCI.

[100]  Pan Hui,et al.  Ubii: Towards Seamless Interaction between Digital and Physical Worlds , 2015, ACM Multimedia.

[101]  Ravin Balakrishnan,et al.  DualKey: Miniature Screen Text Entry via Finger Identification , 2016, CHI.

[102]  KiyokawaKiyoshi,et al.  Managing mobile text in head mounted displays , 2014 .

[103]  Yang Zhang,et al.  Pyro: Thumb-Tip Gesture Recognition Using Pyroelectric Infrared Sensing , 2017, UIST.

[104]  Hai-Ning Liang,et al.  Pointing and Selection Methods for Text Entry in Augmented Reality Head Mounted Displays , 2019, 2019 IEEE International Symposium on Mixed and Augmented Reality (ISMAR).

[105]  Karl J. Friston,et al.  Action and behavior: a free-energy formulation , 2010, Biological Cybernetics.

[106]  Benoît Martin,et al.  TEXTile: Eyes-Free Text Input on Smart Glasses Using Touch Enabled Textile on the Forearm , 2019, INTERACT.

[107]  Daniel Vogel,et al.  RotoSwype: Word-Gesture Typing using a Ring , 2019, CHI.

[108]  Darren Newtson Attribution and the unit of perception of ongoing behavior. , 1973 .

[109]  Kenji Suzuki,et al.  EnhancedTouch: A Smart Bracelet for Enhancing Human-Human Physical Touch , 2016, CHI.

[110]  Ana Paiva,et al.  Empathy in Virtual Agents and Robots , 2017, ACM Trans. Interact. Intell. Syst..

[111]  Xiaojun Bi,et al.  TipText: Eyes-Free Text Entry on a Fingertip Keyboard , 2019, UIST.

[112]  Mike Y. Chen,et al.  PalmType: Using Palms as Keyboards for Smart Glasses , 2015, MobileHCI.

[113]  Schahram Dustdar,et al.  Web AR: A Promising Future for Mobile Augmented Reality—State of the Art, Challenges, and Insights , 2019, Proceedings of the IEEE.

[114]  Victor Kaptelinin,et al.  Affordances in HCI: toward a mediated action perspective , 2012, CHI.

[115]  Yung-Ju Chang,et al.  When There is No Progress with a Task-Oriented Chatbot: A Conversation Analysis , 2019, MobileHCI.

[116]  Christopher G. Atkeson,et al.  Predicting human interruptibility with sensors: a Wizard of Oz feasibility study , 2003, CHI '03.

[117]  Tsutomu Terada,et al.  An information presentation method for head mounted display considering surrounding environments , 2014, AH.

[118]  Timothy W. Bickmore,et al.  Creating New Technologies for Companionable Agents to Support Isolated Older Adults , 2018, ACM Trans. Interact. Intell. Syst..

[119]  Pan Hui,et al.  When Augmented Reality meets Big Data , 2017, 2017 IEEE 37th International Conference on Distributed Computing Systems Workshops (ICDCSW).

[120]  Pourang Irani,et al.  ThumbText: Text Entry for Wearable Devices Using a Miniature Ring , 2018, Graphics Interface.

[121]  Robert J. Moore,et al.  Conversational UX Design: A Practitioner's Guide to the Natural Conversation Framework , 2019, Conversational UX Design.

[122]  Yuanchun Shi,et al.  Investigating Gesture Typing for Indirect Touch , 2019, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[123]  William S. Helton,et al.  Climbing With a Head-Mounted Display , 2016, Hum. Factors.

[124]  Kiyoshi Kiyokawa,et al.  Managing mobile text in head mounted displays: studies on visual preference and text placement , 2014, MOCO.

[125]  Heinrich Hußmann,et al.  Dynamic Subtitles in Cinematic Virtual Reality , 2018, TVX.

[126]  Koen van Boerdonk,et al.  An analysis of input-output relations in interaction with smart tangible objects , 2013, TIIS.

[127]  George Caridakis,et al.  Visualizing the Internet of Things: Naturalizing Human-Computer Interaction by Incorporating AR Features , 2018, IEEE Consumer Electronics Magazine.

[128]  Kangsoo Kim,et al.  Walking Your Virtual Dog: Analysis of Awareness and Proxemics with Simulated Support Animals in Augmented Reality , 2019, 2019 IEEE International Symposium on Mixed and Augmented Reality (ISMAR).

[129]  Takahiro Tanaka,et al.  Study of user interruptibility estimation based on focused application switching , 2011, CSCW '11.

[130]  Joseph A. Paradiso,et al.  NailO: Fingernails as an Input Surface , 2015, CHI.

[131]  Bipin Indurkhya,et al.  Painting an Apple with an Apple , 2018, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[132]  Teruhisa Misu,et al.  Effects of "Real-World" Visual Fidelity on AR Interface Assessment: A Case Study Using AR Head-up Display Graphics in Driving , 2019, 2019 IEEE International Symposium on Mixed and Augmented Reality (ISMAR).

[133]  Barbara S. Chaparro,et al.  The Intuitiveness of Gesture Control with a Mixed Reality Device , 2019 .

[134]  Niels Henze,et al.  Reading on Smart Glasses: The Effect of Text Position, Presentation Type and Walking , 2018, CHI.

[135]  Hyun Seung Yang,et al.  Annotation vs. Virtual Tutor: Comparative Analysis on the Effectiveness of Visual Instructions in Immersive Virtual Reality , 2019, 2019 IEEE International Symposium on Mixed and Augmented Reality (ISMAR).

[136]  M. McDaniel,et al.  (www.interscience.wiley.com) DOI: 10.1002/acp.1002 Delaying Execution of Intentions: Overcoming the Costs of Interruptions , 2022 .

[137]  Jürgen Steimle,et al.  Like A Second Skin: Understanding How Epidermal Devices Affect Human Tactile Perception , 2019, CHI.

[138]  Rainer Groh,et al.  HCI meets Material Science: A Literature Review of Morphing Materials for the Design of Shape-Changing Interfaces , 2018, CHI.

[139]  Tom Igoe,et al.  Physical computing: sensing and controlling the physical world with computers , 2004 .

[140]  Shengdong Zhao,et al.  Positioning Glass: Investigating Display Positions of Monocular Optical See-Through Head-Mounted Display , 2016, ChineseCHI2016.

[141]  Jodi Forlizzi,et al.  A study of fonts designed for screen display , 1998, CHI.

[142]  Saiganesh Swaminathan,et al.  Input, Output and Construction Methods for Custom Fabrication of Room-Scale Deployable Pneumatic Structures , 2019, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[143]  Xiaojun Bi,et al.  Optimal-T9: An Optimized T9-like Keyboard for Small Touchscreen Devices , 2018, ISS.

[144]  Tovi Grossman,et al.  BISHARE: Exploring Bidirectional Interactions Between Smartphones and Head-Mounted Augmented Reality , 2020, CHI.

[145]  I. Scott MacKenzie,et al.  1 thumb, 4 buttons, 20 words per minute: design and evaluation of H4-writer , 2011, UIST.

[146]  Antti Oulasvirta,et al.  GRIDS: Interactive Layout Design with Integer Programming , 2020, CHI.

[147]  Mark Packer,et al.  Wearable technology with head-mounted displays and visual function. , 2014, JAMA.

[148]  Bo Han,et al.  On the Networking Challenges of Mobile Augmented Reality , 2017, VR/AR Network@SIGCOMM.

[149]  Gregory D. Abowd,et al.  TapSkin: Recognizing On-Skin Input for Smartwatches , 2016, ISS.

[150]  David Holman,et al.  Bridging the gap between material science and human-computer interaction , 2019, Interactions.

[151]  Jichen Zhu,et al.  Patterns for How Users Overcome Obstacles in Voice User Interfaces , 2018, CHI.

[152]  Shuchang Xu,et al.  Accurate and Low-Latency Sensing of Touch Contact on Any Surface with Finger-Worn IMU Sensor , 2019, UIST.

[153]  Huamin Qu,et al.  Neighborhood Perception in Bar Charts , 2019, CHI.

[154]  Diogo Cabral,et al.  Augmented object selection through smart glasses , 2019, MUM.

[155]  Jürgen Steimle,et al.  Grasping Microgestures: Eliciting Single-hand Microgestures for Handheld Objects , 2019, CHI.

[156]  Joseph L. Gabbard,et al.  Legibility in Industrial AR: Text Style, Color Coding, and Illuminance , 2015, IEEE Computer Graphics and Applications.

[157]  Mark Ollila,et al.  UMAR: Ubiquitous Mobile Augmented Reality , 2004, MUM '04.

[158]  Gerard Jounghyun Kim,et al.  AR Enabled IoT for a Smart and Interactive Environment: A Survey and Future Directions , 2019, Sensors.

[159]  Tsutomu Terada,et al.  An Information Layout Method for an Optical See-through HMD Considering the Background , 2007, 2007 11th IEEE International Symposium on Wearable Computers.

[160]  Juan Pablo Wachs,et al.  The Effect of Embodied Interaction in Visual-Spatial Navigation , 2016, ACM Trans. Interact. Intell. Syst..

[161]  Sebastian van Delden,et al.  A System Architecture for Hands-Free UAV Drone Control Using Intuitive Voice Commands , 2017, HRI.

[162]  Tong Li,et al.  Quadmetric Optimized Thumb-to-Finger Interaction for Force Assisted One-Handed Text Entry on Mobile Headsets , 2019, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[163]  David R. Large,et al.  "It's small talk, jim, but not as we know it.": engendering trust through human-agent conversation in an autonomous, self-driving car , 2019, CUI.

[164]  Alexis Souchet,et al.  Investigating Cyclical Stereoscopy Effects Over Visual Discomfort and Fatigue in Virtual Reality While Learning , 2019, 2019 IEEE International Symposium on Mixed and Augmented Reality (ISMAR).

[165]  Ludger Schmidt,et al.  Text Presentation for Augmented Reality Applications in Dual-Task Situations , 2019, 2019 IEEE Conference on Virtual Reality and 3D User Interfaces (VR).

[166]  Yulyani Arifin,et al.  User Experience Metric for Augmented Reality Application: A Review , 2018 .

[167]  Per Ola Kristensson,et al.  How do People Type on Mobile Devices?: Observations from a Study with 37,000 Volunteers , 2019, MobileHCI.

[168]  Pan Hui,et al.  Future Networking Challenges: The Case of Mobile Augmented Reality , 2017, 2017 IEEE 37th International Conference on Distributed Computing Systems (ICDCS).

[169]  Yasaman Khazaeni,et al.  All Work and No Play? Conversations with a Question-and-Answer Chatbot in the Wild , 2018, CHI 2018.

[170]  Pan Hui,et al.  TiPoint: detecting fingertip for mid-air interaction on computational resource constrained smartglasses , 2019, UbiComp.

[171]  Niklas Elmqvist,et al.  When David Meets Goliath: Combining Smartwatches with a Large Vertical Display for Visual Data Exploration , 2018, CHI.

[172]  Daniel Vogel,et al.  Tip-Tap: Battery-free Discrete 2D Fingertip Input , 2019, UIST.

[173]  Theresa A. Pardo,et al.  Conceptualizing smart city with dimensions of technology, people, and institutions , 2011, dg.o '11.

[174]  Ewa Luger,et al.  Considering the privacy design issues arising from conversation as platform , 2017 .

[175]  Kristina Höök,et al.  Designing with the Body: Somaesthetic Interaction Design , 2018, CHIRA.

[176]  Suranga Nanayakkara,et al.  M-Hair: Creating Novel Tactile Feedback by Augmenting the Body Hair to Respond to Magnetic Field , 2019, UIST.

[177]  K. Chang,et al.  Embodiment in conversational interfaces: Rea , 1999, CHI '99.

[178]  Qian Wang,et al.  ForceBoard: Subtle Text Entry Leveraging Pressure , 2018, CHI.

[179]  Céline Coutrix,et al.  Morphees+: Studying Everyday Reconfigurable Objects for the Design and Taxonomy of Reconfigurable UIs , 2018, CHI.

[180]  Tong Lu,et al.  iSkin: Flexible, Stretchable and Visually Customizable On-Body Touch Sensors for Mobile Computing , 2015, CHI.

[181]  Shumin Zhai,et al.  i'sFree: Eyes-Free Gesture Typing via a Touch-Enabled Remote Control , 2019, CHI.

[182]  Abigail Sellen,et al.  "Like Having a Really Bad PA": The Gulf between User Expectation and Experience of Conversational Agents , 2016, CHI.

[183]  Manolya Kavakli,et al.  Towards optimization of mid-air gestures for in-vehicle interactions , 2017, OZCHI.

[184]  Gierad Laput,et al.  Enhancing Mobile Voice Assistants with WorldGaze , 2020, CHI.

[185]  Pan Hui,et al.  One-thumb Text Acquisition on Force-assisted Miniature Interfaces for Mobile Headsets , 2020, 2020 IEEE International Conference on Pervasive Computing and Communications (PerCom).

[186]  R. F. Haines,et al.  Cognitive issues in head-up displays , 1980 .