Help Me to Help You: How to Learn Intentions, Actions and Plans

The collaboration between a human and a robot is here understood as a learning process mediated by the instructor prompt behaviours and the apprentice collecting information from them to learn a plan. The instructor wears the Gaze Machine, a wearable device gathering and conveying visual and audio input from the instructor while executing a task. The robot, on the other hand, is eager to learn both the best sequence of actions, their timing and how they interlace. The cross relation among actions is specified both in terms of time intervals for their execution, and in terms of location in space to cope with the instruction interaction with people and objects in the scene. We outline this process: how to transform the rich information delivered by the Gaze Machine into a plan. Specifically, how to obtain a map of the instructor positions and his gaze position, via visual slam and gaze fixations; further, how to obtain an action map from the running commentaries and the topological maps and, finally, how to obtain a temporal net of the relevant actions that have been extracted. The learned structure is then managed by the flexible time paradigm of flexible planning in the Situation Calculus for execution monitoring and plan generation.

[1]  Andrew Calway,et al.  Real-Time Visual SLAM with Resilience to Erratic Motion , 2006, 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06).

[2]  Bernhard P. Wrobel,et al.  Multiple View Geometry in Computer Vision , 2001 .

[3]  G. Rizzolatti,et al.  Understanding motor events: a neurophysiological study , 2004, Experimental Brain Research.

[4]  Olivier Stasse,et al.  MonoSLAM: Real-Time Single Camera SLAM , 2007, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[5]  D. Simons,et al.  Failure to detect changes to people during a real-world interaction , 1998 .

[6]  三嶋 博之 The theory of affordances , 2008 .

[7]  Reinhard Koch,et al.  Visual Modeling with a Hand-Held Camera , 2004, International Journal of Computer Vision.

[8]  Analía Amandi,et al.  Plan recognition for interface agents , 2007, Artificial Intelligence Review.

[9]  Christopher W. Geib,et al.  Delaying Commitment in Plan Recognition Using Combinatory Categorial Grammars , 2009, IJCAI.

[10]  Fiora Pirri,et al.  The well-designed logical robot: Learning and experience from observations to the Situation Calculus , 2011, Artif. Intell..

[11]  Stefan Schaal,et al.  http://www.jstor.org/about/terms.html. JSTOR's Terms and Conditions of Use provides, in part, that unless you have obtained , 2007 .

[12]  Christopher W. Geib,et al.  The meaning of action: a review on action recognition and mapping , 2007, Adv. Robotics.

[13]  N. S. Sridharan,et al.  The Plan Recognition Problem: An Intersection of Psychology and Artificial Intelligence , 1978, Artif. Intell..

[14]  Alberto Finzi,et al.  Representing Flexible Temporal Behaviors in the Situation Calculus , 2005, IJCAI.

[15]  R. Shaw,et al.  Perceiving, Acting and Knowing : Toward an Ecological Psychology , 1978 .

[16]  Jake K. Aggarwal,et al.  Human Motion Analysis: A Review , 1999, Comput. Vis. Image Underst..

[17]  Andrew W. Fitzgibbon,et al.  Maintaining multiple motion model hypotheses over many views to recover matching and structure , 1998, Sixth International Conference on Computer Vision (IEEE Cat. No.98CH36271).

[18]  Fiora Pirri,et al.  Eyes and Cameras Calibration for 3D World Gaze Detection , 2008, ICVS.

[19]  G. Rizzolatti,et al.  Action recognition in the premotor cortex. , 1996, Brain : a journal of neurology.

[20]  Michael Werman,et al.  The viewing graph , 2003, 2003 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2003. Proceedings..

[21]  Adrian Hilton,et al.  A survey of advances in vision-based human motion capture and analysis , 2006, Comput. Vis. Image Underst..

[22]  Alessandro Rudi,et al.  Linear Solvability in the Viewing Graph , 2010, ACCV.

[23]  Brett Browning,et al.  A survey of robot learning from demonstration , 2009, Robotics Auton. Syst..

[24]  Robert P. Goldman,et al.  A Bayesian Model of Plan Recognition , 1993, Artif. Intell..

[25]  Ronald Poppe,et al.  A survey on vision-based human action recognition , 2010, Image Vis. Comput..

[26]  Henry A. Kautz A formal theory of plan recognition , 1987 .

[27]  Fiora Pirri,et al.  Bottom-Up Gaze Shifts and Fixations Learning by Imitation , 2007, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[28]  J. Gibson,et al.  Perceptual learning; differentiation or enrichment? , 1955, Psychological review.

[29]  Henry A. Kautz,et al.  Generalized Plan Recognition , 1986, AAAI.