An application of passive human-robot interaction: human tracking based on attention distraction

In this research, a taxonomy is introduced to cover important considerations for human-robot interactions. As an application of passive human-robot interaction, two modalities for localizing humans based on sound source localization and infrared motion detection were developed and integrated with the face-tracker system of a humanoid ISAC (intelligent soft arm control), in order to direct ISACs attention and to prevent it from being quickly distracted. The sound source localization and passive infrared motion detection systems are used to provide the face-tracker system with candidate regions for finding a face. In order to avoid the situation where the robot appears to be "hyperactive" and cannot give sufficient attention to a newly discovered face, these sensing modules should not directly gain control of the tracking if the system has recently acquired a new face. Our goal is to allow a human to redirect the attention of the system but give the system a method to ignore the distraction if recently engaged.

[1]  H. Yamasaki,et al.  Audio-visual sensor fusion system for intelligent sound sensing , 1994, Proceedings of 1994 IEEE International Conference on MFI '94. Multisensor Fusion and Integration for Intelligent Systems.

[2]  D. Mitchell Wilkes,et al.  ISAC: Foundations in Human-Humanoid Interaction , 2000, IEEE Intell. Syst..

[3]  Steven George Goodridge Multimedia sensor fusion for intelligent camera control and human-computer interaction , 1997 .

[4]  Hong Wang,et al.  Voice source localization for automatic camera pointing system in videoconferencing , 1997, 1997 IEEE International Conference on Acoustics, Speech, and Signal Processing.

[5]  Benoît Champagne,et al.  A new cepstral prefiltering technique for estimating time delay under reverberant conditions , 1997, Signal Process..

[6]  James L. Flanagan,et al.  Estimation of wavefront arrival delay for acoustical signals using the cross‐power spectrum phase technique , 1996 .

[7]  Keiichiro Hoashi,et al.  Humanoid robot-development of an information assistant robot Hadaly , 1997, Proceedings 6th IEEE International Workshop on Robot and Human Communication. RO-MAN'97 SENDAI.

[8]  Kazuhiko Kawamura,et al.  Multi-agent system for a human-friendly robot , 1999, IEEE SMC'99 Conference Proceedings. 1999 IEEE International Conference on Systems, Man, and Cybernetics (Cat. No.99CH37028).

[9]  Michael S. Brandstein,et al.  A hybrid real-time face tracking system , 1998, Proceedings of the 1998 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP '98 (Cat. No.98CH36181).

[10]  Michele Rucci,et al.  Theoretical study , 2022 .

[11]  Robert Todd Pack,et al.  Ima: the intelligent machine architecture , 1998 .

[12]  P. C. Ching,et al.  Non-stationary time delay estimation with a multipath , 1989, International Conference on Acoustics, Speech, and Signal Processing,.

[13]  Julius O. Smith,et al.  Adaptive multipath delay estimation , 1984, IEEE Trans. Acoust. Speech Signal Process..

[14]  Maurizio Omologo,et al.  Acoustic event localization using a crosspower-spectrum phase based technique , 1994, Proceedings of ICASSP '94. IEEE International Conference on Acoustics, Speech and Signal Processing.

[15]  Kamen Y. Guentchev,et al.  Learning-Based Three Dimensional Sound Localization Using a Compact Non-Coplanar Array of Microphones , 1998 .

[16]  Cynthia Breazeal,et al.  A Motivational System for Regulating Human-Robot Interaction , 1998, AAAI/IAAI.

[17]  Maurizio Omologo,et al.  Acoustic source location in noisy and reverberant environment using CSP analysis , 1996, 1996 IEEE International Conference on Acoustics, Speech, and Signal Processing Conference Proceedings.

[18]  Harvey F. Silverman,et al.  A two-stage algorithm for determining talker location from linear microphone array data , 1992 .

[19]  Douglas E. Sturim,et al.  Tracking multiple talkers using microphone-array measurements , 1997, 1997 IEEE International Conference on Acoustics, Speech, and Signal Processing.

[20]  Alexander H. Waibel,et al.  Knowing who to listen to in speech recognition: visually guided beamforming , 1995, 1995 International Conference on Acoustics, Speech, and Signal Processing.

[21]  Yiu-Tong Chan,et al.  Constrained adaptation for time delay estimation with multipath propagation , 1991 .