Eye Controlled Media: Present and Future State

Today, the human eye-gaze can be recorded by relatively unobtrusive techniques. This thesis argues that it is possible to use the eye-gaze of a computer user in the interface to aid the control of the application. Care must be taken, though, that eye-gaze tracking data is used in a sensible way, since the nature of human eye-movements is a combination of several voluntary and involuntary cognitive processes. The main reason for eye-gaze based user interfaces being attractive is that the direction of the eye-gaze can express the interests of the user it is a potential porthole into the current cognitive processes and communication through the direction of the eyes is faster than any other mode of human communication. It is argued that eye-gaze tracking data is best used in multimodal interfaces where the user interacts with the data instead of the interface, in so-called noncommand user interfaces. Furthermore, ve usability criteria for eye-gaze media are given. This thesis also suggests research into a new, interactive lm medium: interest and emotion sensitive media (IES). IES exploits the viewer's eye-gaze and other a ection measures to determine how script paths of an IES lm are traversed. IES will be very di cult to implement today, as research is needed to investigate temporal problems in script construction, how multiple persons can use IES, how activation areas are constructed and how the producer of an IES lm is best assisted in writing the script. This thesis also contains reviews of current technical possibilities, psychological aspects of eye-gaze tracking and current eye-gaze based systems. Preface This thesis is submitted in partial ful lment of the requirements for a Danish Bachelor's Degree in information psychology for both authors. It contains work done from February to May 1995. Our supervisor on the project has been John Paulin Hansen, the Risø National Laboratory. The thesis has been made solely by the authors; most of the text, however, is based on the research of others, and we have done our best to provide references to these sources. In December 1994, our teacher in cognitive psychology at that time, John Paulin Hansen, introduced us to one of his latest developments: an application intended for public use and operated solely by the eyes of the user, called the EyeCatcher. We were both impressed and intrigued by this new means of controlling a computer, so when John presented a list of projects to Arne, including the present one, the choice of the EyeCatcher project was obvious. In the spring semester of 1995 we were to write up our minor's thesis on information psychology. Initially, we worked on two di erent projects, Theo working together with another group on a 3-D operating systems project and Arne working alone on the EyeCatcher project. Soon, however, it was evident that we could make a better project by joining forces, and we believe that the present thesis con rms this. Writing this thesis has been hard but in the process of writing we feel we have learned a lot and our initial conceptions of eye-gaze controlling have certainly changed! We have dealt with a lot of subjects, in an attempt to give this thesis a broad perspective on eye-gaze control, thus combining many aspects of cognitive psychology and human-computer interaction. The Aim of this Thesis Since this thesis is written as the nal thesis of the bachelor's degree in information psychology, the text is primarily aimed at teachers of information psychology at the University of Copenhagen and pupils attending the courses there, but we hope it will be of interest for the general HCI and cognitive psychology communities, too. iv Acknowledgements We would like to thank our supervisor, John Paulin Hansen, for lots of great inspiration, ideas, comments and an endless stream of articles. Especially our meetings at Risø National Laboratory have given us much inspiration, due to John's, to us, remarkable ability to cheer us up when we were in desperate need of it because everything seemed impossible and fruitless. Thanks also to Anja Philip and EyeCatcher -pilot Jesper Brix, both from the Experimentarium (a science museum in Hellerup in Denmark), for letting us stay there for two days, evaluating the EyeCatcher. Finally, lots of thanks are due to Claus Dantzer-Sørensen for expertly proofreading this thesis and especially Jill Glenstrup who had the laborious job of turning our attempts at the English language into readable prose. Hopefully she has removed all the serious errors, but any that might be left are wholly our responsibility. Arne John Glenstrup & Theo Engell-Nielsen v

[1]  Alex Waibel,et al.  Face locating and tracking for human-computer interaction , 1994, Proceedings of 1994 28th Asilomar Conference on Signals, Systems and Computers.

[2]  C. Eriksen,et al.  Allocation of attention in the visual field. , 1985, Journal of experimental psychology. Human perception and performance.

[3]  W. Graf,et al.  Ergonomic evaluation of user-interfaces by means of eye-movement data , 1989 .

[4]  M A Just,et al.  A theory of reading: from eye fixations to comprehension. , 1980, Psychological review.

[5]  R. C. Emerson,et al.  Paralysis of the awake human: Visual perceptions , 1976, Vision Research.

[6]  D. LaBerge Spatial extent of attention to letters and words. , 1983, Journal of experimental psychology. Human perception and performance.

[7]  M. Land Visual Perception: Physiology, Psychology and Ecology, Vicki Bruce, Patrick Green. Lawrence Erlbaum, London (1985), xiii, +369. Price £8.95 (paperback) , 1986 .

[8]  Richard A. Bolt,et al.  The human interface: Where people and computers meet , 1984 .

[9]  Jakob Nielsen,et al.  Noncommand user interfaces , 1993, CACM.

[10]  A. Treisman,et al.  Illusory conjunctions in the perception of objects , 1982, Cognitive Psychology.

[11]  C. Eriksen,et al.  Effects of noise letters upon the identification of a target letter in a nonsearch task , 1974 .

[12]  Alexander H. Waibel,et al.  Toward movement-invariant automatic lip-reading and speech recognition , 1995, 1995 International Conference on Acoustics, Speech, and Signal Processing.

[13]  Alexander H. Waibel,et al.  Knowing who to listen to in speech recognition: visually guided beamforming , 1995, 1995 International Conference on Acoustics, Speech, and Signal Processing.

[14]  L. Kaufman,et al.  Spontaneous fixation tendencies for visual forms , 1969 .

[15]  Worthy N. Martin,et al.  Human-computer interaction using eye-gaze input , 1989, IEEE Trans. Syst. Man Cybern..

[16]  A. L. Yarbus Eye Movements During Perception of Complex Objects , 1967 .

[17]  John Paulin Hansen,et al.  Eye-gaze control of multimedia systems , 1995 .

[18]  Shumeet Baluja,et al.  Non-Intrusive Gaze Tracking Using Artificial Neural Networks , 1993, NIPS.

[19]  Richard A. Bolt,et al.  A gaze-responsive self-disclosing display , 1990, CHI '90.

[20]  C. Ware,et al.  An evaluation of an eye tracker as a device for computer input2 , 1987, CHI '87.

[21]  J. D. Gould Pattern recognition and eye-movement parameters , 1967 .

[22]  K. Preston White,et al.  Eye-gaze word processing , 1990, IEEE Trans. Syst. Man Cybern..

[23]  H. Egeth,et al.  Further evidence for a time-independent shift of the focus of attention , 1991, Perception & psychophysics.

[24]  N. Mackworth,et al.  The gaze selects informative details within pictures , 1967 .

[25]  J K JacobRobert,et al.  The use of eye movements in human-computer interaction techniques , 1991 .

[26]  Allen Newell,et al.  The keystroke-level model for user performance time with interactive systems , 1980, CACM.

[27]  A. Baddeley The concept of working memory: A view of its current state and probable future development , 1981, Cognition.

[28]  J M Findlay,et al.  A probability vector and transition matrix analysis of eye movements during visual search. , 1995, Acta psychologica.

[29]  O. Mimura [Eye movements]. , 1992, Nippon Ganka Gakkai zasshi.

[30]  Robert J. K. Jacob,et al.  Eye tracking in advanced interface design , 1995 .

[31]  Géry d'Ydewalle,et al.  A comparison of a new limbus tracker, corneal reflection technique, Purkinje eye tracking and electro-oculography. , 1993 .

[32]  J Theeuwes,et al.  Visual selective attention: a theoretical analysis. , 1993, Acta psychologica.

[33]  M. Posner,et al.  Attention and the detection of signals. , 1980, Journal of experimental psychology.

[34]  John J. Leggett,et al.  Interaction styles and input/output devices , 1993, Behav. Inf. Technol..

[35]  Jens Rasmussen,et al.  Skills, rules, and knowledge; signals, signs, and symbols, and other distinctions in human performance models , 1983, IEEE Transactions on Systems, Man, and Cybernetics.

[36]  Wa James Tam,et al.  Static and dynamic spatial resolution in image coding: an investigation of eye movements , 1991, Electronic Imaging.

[37]  A. A. Skavenski,et al.  Miniature eye movement. , 1973, Science.

[38]  James D. Murray,et al.  Encyclopedia of graphics file formats , 1994 .

[39]  Robert J. K. Jacob,et al.  New Human-Computer Interaction Techniques , 1994 .

[40]  Mark T. Keane,et al.  Cognitive Psychology: A Student's Handbook , 1990 .

[41]  V. Bruce,et al.  Visual Cognition: Computational, Experimental, and Neuropsychological Perspectives , 1989 .

[42]  M. Wierda,et al.  Interpreting eye movements of traffic participants , 1993 .

[43]  James Gips,et al.  Direct Control of the Computer Through Electrodes Placed Around the Eyes , 1993, HCI.

[44]  G. McConkie,et al.  What guides a reader's eye movements? , 1976, Vision Research.