Expectation-driven autonomous learning and interaction system

We introduce our latest autonomous learning and interaction system instance ALIS 2. It comprises different sensing modalities for visual (depth blobs, planar surfaces, motion) and auditory (speech, localization) signals and self-collision free behavior generation on the robot ASIMO. The system design emphasizes the split into a completely autonomous reactive layer and an expectation generation layer. Different feature channels can be classified and named with arbitrary speech labels in on-line learning sessions. The feasibility of the proposed approach is shown by interaction experiments.

[1]  Z. Pylyshyn Visual indexes, preconceptual objects, and situated vision , 2001, Cognition.

[2]  Marc Toussaint,et al.  Probabilistic inference for solving discrete and continuous state Markov Decision Processes , 2006, ICML.

[3]  Inna Mikhailova,et al.  Organizing multimodal perception for autonomous learning and interactive systems , 2008, Humanoids 2008 - 8th IEEE-RAS International Conference on Humanoid Robots.

[4]  Christian Goerick,et al.  Fast detection of arbitrary planar surfaces from unreliable 3D data , 2009, 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[5]  C. Balkenius Natural intelligence in artificial creatures , 1995 .

[6]  Werner von Seelen,et al.  Complex behavior by means of dynamical systems for an anthropomorphic robot , 1999, Neural Networks.

[7]  Chen Yu,et al.  A multimodal learning interface for word acquisition , 2003, 2003 IEEE International Conference on Acoustics, Speech, and Signal Processing, 2003. Proceedings. (ICASSP '03)..

[8]  Ronald A. Rensink Seeing, sensing, and scrutinizing , 2000, Vision Research.

[9]  Mark Dunn,et al.  Visually Guided Whole Body Interaction , 2007, Proceedings 2007 IEEE International Conference on Robotics and Automation.

[10]  Ronald C. Arkin,et al.  Biasing behavioral activation with intent for an entertainment robot , 2008, Intell. Serv. Robotics.

[11]  Heiko Wersing,et al.  Towards incremental hierarchical behavior generation for humanoids , 2007, 2007 7th IEEE-RAS International Conference on Humanoid Robots.

[12]  Ronald C. Arkin,et al.  Biasing Behavioral Activation with Intent , 2006 .

[13]  Michael I. Jordan,et al.  Hierarchical Mixtures of Experts and the EM Algorithm , 1994, Neural Computation.

[14]  Giulio Sandini,et al.  From sensorimotor development to object perception , 2005, 5th IEEE-RAS International Conference on Humanoid Robots, 2005..

[15]  Stephen Hart,et al.  A Framework for Learning Declarative Structure , 2006 .

[16]  Martin Heckmann,et al.  Real-time Sound Localization With a Binaural Head-system Using a Biologically-inspired Cue-triple Mapping , 2006, 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[17]  Giulio Sandini,et al.  Object-based Visual Attention: a Model for a Behaving Robot , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05) - Workshops.

[18]  R. Polikar,et al.  Multiple classifier systems for multisensor data fusion , 2006, Proceedings of the 2006 IEEE Sensors Applications Symposium, 2006..

[19]  Austen Clark,et al.  Feature-placing and proto-objects , 2004 .

[20]  Volker Willert,et al.  Estimating Object Proper Motion Using Optical Flow, Kinematics, and Depth Information , 2008, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[21]  Aaron Sloman,et al.  Towards an Integrated Robot with Multiple Cognitive Functions , 2007, AAAI.

[22]  Gökhan Ince,et al.  Using binaural and spectral cues for azimuth and elevation localization , 2008, 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[23]  Tomoaki Nakamura,et al.  Multimodal object categorization by a robot , 2007, 2007 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[24]  Robert A. Jacobs,et al.  Hierarchical Mixtures of Experts and the EM Algorithm , 1993, Neural Computation.