Human-Computer Interaction

This chapter focuses on looking at users to build more intuitive and friendly interfaces. We will cover two general types of input modalities: tracking of the head and eyes, and tracking of the hand, especially of the fingers. Head pose and eye–gaze estimation are useful in allowing the computer to understand the direction of the user’s focus. We will provide an overview of the many techniques that researchers have applied to this task. We will also consider the automated analysis of hand and finger gestures. These are more active modalities, designed to communicate and issue commands to the computer. We will provide a taxonomy of gestures in the context of human–computer interaction and survey the field of techniques. Finally, we will discuss possible applications of these input modalities. In general, we conclude that existing systems are not yet mature, but that there is great potential for future research.

[1]  Seong-Whan Lee,et al.  Recognizing hand gestures using dynamic Bayesian network , 2008, 2008 8th IEEE International Conference on Automatic Face & Gesture Recognition.

[2]  John M. Carroll,et al.  Softening Up Hard Science: Reply to Newell and Card , 1986, Hum. Comput. Interact..

[3]  R. Näsänen,et al.  Eye movements in the visual search of word lists , 2002, Vision Research.

[4]  Trevor Darrell,et al.  Head gesture recognition in intelligent interfaces: the role of context in improving recognition , 2006, IUI '06.

[5]  Jian-Gang Wang,et al.  Estimating the eye gaze from one eye , 2005, Comput. Vis. Image Underst..

[6]  Thomas S. Huang,et al.  Static Hand Gesture Recognition based on Local Orientation Histogram Feature Distribution Model , 2004, 2004 Conference on Computer Vision and Pattern Recognition Workshop.

[7]  David Beymer,et al.  Eye gaze tracking using an active stereo head , 2003, 2003 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2003. Proceedings..

[8]  H. Rex Hartson,et al.  Human-computer interaction: Interdisciplinary roots and trends , 1998, J. Syst. Softw..

[9]  Zhiwei Zhu,et al.  Eye and gaze tracking for interactive graphic display , 2002, SMARTGRAPH '02.

[10]  David Windridge,et al.  A Linguistic Feature Vector for the Visual Interpretation of Sign Language , 2004, ECCV.

[11]  Carlos Hitoshi Morimoto,et al.  Pupil detection and tracking using multiple light sources , 2000, Image Vis. Comput..

[12]  J. C. R. Licklider,et al.  Man-Computer Symbiosis , 1960 .

[13]  Mark Ashdown,et al.  Combining head tracking and mouse input for a GUI on multiple monitors , 2005, CHI Extended Abstracts.

[14]  David C. Hogg,et al.  Towards 3D hand tracking using a deformable model , 1996, Proceedings of the Second International Conference on Automatic Face and Gesture Recognition.

[15]  Zhiwei Zhu,et al.  Nonlinear Eye Gaze Mapping Function Estimation via Support Vector Regression , 2006, 18th International Conference on Pattern Recognition (ICPR'06).

[16]  Carlos Hitoshi Morimoto,et al.  Eye gaze tracking techniques for interactive applications , 2005, Comput. Vis. Image Underst..

[17]  Moshe Eizenman,et al.  General theory of remote gaze estimation using the pupil center and corneal reflections , 2006, IEEE Transactions on Biomedical Engineering.

[18]  Naoki Mukawa,et al.  FreeGaze: a gaze tracking system for everyday gaze interaction , 2002, ETRA.

[19]  Zhiwei Zhu,et al.  Eye and gaze tracking for interactive graphic display , 2002, SMARTGRAPH '02.

[20]  David J. Fleet,et al.  Model-based hand tracking with texture, shading and self-occlusions , 2008, 2008 IEEE Conference on Computer Vision and Pattern Recognition.

[21]  Alexander Zelinsky,et al.  Behavior recognition based on head pose and gaze direction measurement , 2000, Proceedings. 2000 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2000) (Cat. No.00CH37113).

[22]  K. Preston White,et al.  Spatially dynamic calibration of an eye-tracking system , 1993, IEEE Trans. Syst. Man Cybern..

[23]  Marcelo Knörich Zuffo,et al.  On the usability of gesture interfaces in virtual reality environments , 2005, CLIHC '05.

[24]  Michael Isard,et al.  CONDENSATION—Conditional Density Propagation for Visual Tracking , 1998, International Journal of Computer Vision.

[25]  Vladimir Pavlovic,et al.  Visual Interpretation of Hand Gestures for Human-Computer Interaction: A Review , 1997, IEEE Trans. Pattern Anal. Mach. Intell..

[26]  Harry Wechsler,et al.  Face pose discrimination using support vector machines (SVM) , 1998, Proceedings. Fourteenth International Conference on Pattern Recognition (Cat. No.98EX170).

[27]  Spiro Kiousis,et al.  Interactivity: a concept explication , 2002, New Media Soc..

[28]  Jakob Nielsen,et al.  Chapter 4 – The Usability Engineering Lifecycle , 1993 .

[29]  Takeo Kanade,et al.  Real-time combined 2D+3D active appearance models , 2004, CVPR 2004.

[30]  Mads Nielsen,et al.  Eye typing using Markov and active appearance models , 2002, Sixth IEEE Workshop on Applications of Computer Vision, 2002. (WACV 2002). Proceedings..

[31]  Yu-Te Wu,et al.  A calibration-free gaze tracking technique , 2000, Proceedings 15th International Conference on Pattern Recognition. ICPR-2000.

[32]  Shaogang Gong,et al.  Support vector machine based multi-view face detection and recognition , 2004, Image Vis. Comput..

[33]  Yvonne Kammerer,et al.  Looking my way through the menu: the impact of menu design and multimodal input on gaze-based menu selection , 2008, ETRA.

[34]  Brian Shackel,et al.  ERGONOMICS IN THE DESIGN OF A LARGE DIGITAL COMPUTER CONSOLE , 1962 .

[35]  Thomas S. Huang,et al.  Face as mouse through visual face tracking , 2007, Comput. Vis. Image Underst..

[36]  Roberto Cipolla,et al.  Determining the gaze of faces in images , 1994, Image Vis. Comput..

[37]  M.M. Trivedi,et al.  HyHOPE: Hybrid Head Orientation and Position Estimation for vision-based driver head tracking , 2008, 2008 IEEE Intelligent Vehicles Symposium.

[38]  Yael Edan,et al.  A Method for Selection of Optimal Hand Gesture Vocabularies , 2009, Gesture Workshop.

[39]  Allen Newell,et al.  The Prospects for Psychological Science in Human-Computer Interaction , 1985, Hum. Comput. Interact..

[40]  John Paulin Hansen,et al.  Noise tolerant selection by gaze-controlled pan and zoom in 3D , 2008, ETRA.

[41]  Sang Ryong Kim,et al.  Are physically embodied social agents better than disembodied social agents?: The effects of physical embodiment, tactile interaction, and people's loneliness in human-robot interaction , 2006, Int. J. Hum. Comput. Stud..

[42]  Mohan M. Trivedi,et al.  Head Pose Estimation in Computer Vision: A Survey , 2009, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[43]  Mircea Nicolescu,et al.  Vision-based hand pose estimation: A review , 2007, Comput. Vis. Image Underst..

[44]  Björn Stenger,et al.  Model-based hand tracking using a hierarchical Bayesian filter , 2006, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[45]  William T. Freeman,et al.  Example-based head tracking , 1996, Proceedings of the Second International Conference on Automatic Face and Gesture Recognition.

[46]  Rainer Stiefelhagen,et al.  Pointing gesture recognition based on 3D-tracking of face, hands and head orientation , 2003, ICMI '03.

[47]  Narendra Ahuja,et al.  Recognizing hand gesture using motion trajectories , 1999, Proceedings. 1999 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Cat. No PR00149).

[48]  Takehiko Ohno,et al.  One-point calibration gaze tracking method , 2006, ETRA.

[49]  Donald A. Norman,et al.  Affordance, conventions, and design , 1999, INTR.

[50]  Dan Witzner Hansen,et al.  Eye tracking in the wild , 2005, Comput. Vis. Image Underst..

[51]  Narendra Ahuja,et al.  Appearance-based eye gaze estimation , 2002, Sixth IEEE Workshop on Applications of Computer Vision, 2002. (WACV 2002). Proceedings..

[52]  Vassilis Athitsos,et al.  Nearest neighbor search methods for handshape recognition , 2008, PETRA '08.

[53]  Shumin Zhai,et al.  Manual and gaze input cascaded (MAGIC) pointing , 1999, CHI '99.

[54]  Stan Sclaroff,et al.  Estimating 3D hand pose from a cluttered image , 2003, 2003 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2003. Proceedings..

[55]  James M. Rehg Visual analysis of high DOF articulated objects with application to hand tracking , 1995 .

[56]  Mark Weiser The computer for the 21st century , 1991 .

[57]  Stan Sclaroff,et al.  A Unified Framework for Gesture Recognition and Spatiotemporal Gesture Segmentation , 2009, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[58]  Jakob Nielsen,et al.  Ten Usability Heuristics , 2006 .

[59]  Luc Van Gool,et al.  Fast stochastic optimization for articulated structure tracking , 2007, Image Vis. Comput..

[60]  Carlo Tomasi,et al.  3D tracking = classification + interpolation , 2003, Proceedings Ninth IEEE International Conference on Computer Vision.