Maps, vibration or gaze? Comparison of novel navigation assistance in indoor and outdoor environments
暂无分享,去创建一个
Martin Raubal | Ioannis Giannopoulos | Charalampos Gkonos | M. Raubal | I. Giannopoulos | Charalampos Gkonos
[1] Andrew T. Duchowski,et al. Eye Tracking Methodology: Theory and Practice , 2003, Springer London.
[2] Oliver Lemon,et al. Summary of SpaceBook project results , 2014 .
[3] Linda R. Elliott,et al. Field-Based Validation of a Tactile Navigation Device , 2010, IEEE Transactions on Haptics.
[4] Martin Raubal,et al. Mobile Outdoor Gaze-Based GeoHCI , 2013 .
[5] Keith S. Karn,et al. Commentary on Section 4. Eye tracking in human-computer interaction and usability research: Ready to deliver the promises. , 2003 .
[6] Karl Rehrl,et al. Pedestrian navigation with augmented reality, voice and digital map: final results from an in situ field study assessing performance and user experience , 2014, J. Locat. Based Serv..
[7] Ergonomic requirements for office work with visual display terminals ( VDTs ) — Part 11 : Guidance on usability , 1998 .
[8] Howell O. Istance,et al. Gaze gestures or dwell-based interaction? , 2012, ETRA '12.
[9] D. Mark,et al. Cognitive and Linguistic Aspects of Geographic Space: An Introduction , 1991 .
[10] Georg Gartner,et al. Applications of location–based services: a selected review , 2007, J. Locat. Based Serv..
[11] Martin Raubal,et al. GazeNav: Gaze-Based Pedestrian Navigation , 2015, MobileHCI.
[12] Yueqing Li,et al. Assessing the Effectiveness of Vibrotactile Feedback on a 2D Navigation Task , 2013, HCI.
[13] D. R. Montello. The Cambridge Handbook of Visuospatial Thinking: Navigation , 2005 .
[14] Matthew Chalmers,et al. EyeSpy: supporting navigation through play , 2009, CHI.
[15] Raimund Dachselt,et al. Investigating gaze-supported multimodal pan and zoom , 2012, ETRA '12.
[16] Albrecht Schmidt,et al. Implicit human computer interaction through context , 2000, Personal Technologies.
[17] S. Hart,et al. Development of NASA-TLX (Task Load Index): Results of Empirical and Theoretical Research , 1988 .
[18] Martin Raubal,et al. Many to Many Mobile Maps , 2013 .
[19] Jason Wither,et al. Moving beyond the map: automated landmark based pedestrian guidance using street level panoramas , 2013, MobileHCI '13.
[20] Thomas Kieninger,et al. Gaze guided object recognition using a head-mounted eye tracker , 2012, ETRA '12.
[21] Martin Pielot,et al. PocketNavigator: studying tactile navigation systems in-situ , 2012, CHI.
[22] A H Rupert. An instrumentation solution for reducing spatial disorientation mishaps. , 2000, IEEE engineering in medicine and biology magazine : the quarterly magazine of the Engineering in Medicine & Biology Society.
[23] Nina Runge,et al. Using split screens to combine maps and images for pedestrian navigation , 2014, J. Locat. Based Serv..
[24] Hendrik A. H. C. van Veen,et al. Waypoint navigation with a vibrotactile waist belt , 2005, TAP.
[25] C J Scheirer,et al. The analysis of ranked data derived from completely randomized factorial designs. , 1976, Biometrics.
[26] David R. Morse,et al. AudioGPS: Spatial Audio Navigation with a Minimal Attention Interface , 2002, Personal and Ubiquitous Computing.
[27] Roel Vertegaal,et al. Attentive User Interfaces , 2003 .
[28] Oscar Mauricio Serrano Jaimes,et al. EVALUACION DE LA USABILIDAD EN SITIOS WEB, BASADA EN EL ESTANDAR ISO 9241-11 (International Standard (1998) Ergonomic requirements For office work with visual display terminals (VDTs)-Parts II: Guidance on usability , 2012 .
[29] Tommy Strandvall,et al. Eye Tracking in Human-Computer Interaction and Usability Research , 2009, INTERACT.
[30] J. F. Kelley,et al. An empirical methodology for writing user-friendly natural language computer applications , 1983, CHI '83.
[31] Martin Schrepp,et al. Construction and Evaluation of a User Experience Questionnaire , 2008, USAB.
[32] Martin Raubal,et al. GeoGazemarks: providing gaze history for the orientation on small display maps , 2012, ICMI '12.
[33] Martin Schrepp,et al. User Experience Questionnaire (UEQ) Benchmark. Praxiserfahrungen zur Auswertung und Anwendung von UEQ-Erhebungen im Business-Umfeld , 2013, UP.
[34] Bodo Urban,et al. Exploring vibrotactile feedback on the body and foot for the purpose of pedestrian navigation , 2015, iWOAR.
[35] Robert Hardy,et al. Design, implementation and evaluation of a novel public display for pedestrian navigation: the rotating compass , 2009, CHI.
[36] Robert J. K. Jacob,et al. Interacting with eye movements in virtual environments , 2000, CHI.
[37] Robert J. K. Jacob,et al. What you look at is what you get: eye movement-based interaction techniques , 1990, CHI '90.
[38] Maximilian Schirmer,et al. Shoe me the Way: A Shoe-Based Tactile Interface for Eyes-Free Urban Navigation , 2015, MobileHCI.
[39] Anthony E. Richardson,et al. Development of a self-report measure of environmental spatial ability. , 2002 .
[40] Robert Hardy,et al. NaviRadar: a novel tactile information display for pedestrian navigation , 2011, UIST.
[41] Dieter Schmalstieg,et al. Handheld augmented reality indoor navigation with activity-based instructions , 2011, Mobile HCI.
[42] Jennifer J. Richler,et al. Effect size estimates: current use, calculations, and interpretation. , 2012, Journal of experimental psychology. General.
[43] Martin Pielot,et al. Tactile Wayfinder: Comparison of Tactile Waypoint Navigation with Commercial Pedestrian Navigation Systems , 2010, Pervasive.