Computationally efficient task execution is very important for autonomous mobile robots endowed with limited on-board computational capabilities. Most robot control approaches assume fixed state and action representations, and use a single algorithm to map states to actions. However, not all instances of a given task require equally complex algorithms and equally detailed representations. The main motivation for this work is a desire to reduce the computational footprint of performing a task by allowing the robot to run simpler algorithms whenever possible, and resort to more complex algorithms only when needed. We contribute the Multi-Resolution Task Execution (MRTE) algorithm that utilizes human feedback to learn a mapping from a given state to an appropriate detail resolution consisting of a state and action representation, and an algorithm. We then present Model Plus Correction (M+C), an algorithm that complements an existing robot controller with corrective human feedback to further improve the task execution performance. Finally, we introduce Multi-Resolution Model Plus Correction (MRM+C) as a combination of MRTE and M+C. We provide formal definitions of MRTE, M+C, and MRM+C, showing how they relate to general robot control problem and Learning from Demonstration (LfD) methods. We present detailed experimental results demonstrating the effectiveness of proposed methods on a simulated goal-directed humanoid obstacle avoidance task.
[1]
Pieter Abbeel,et al.
Hierarchical Apprenticeship Learning with Application to Quadruped Locomotion
,
2007,
NIPS.
[2]
Manuela M. Veloso,et al.
Interactive Policy Learning through Confidence-Based Autonomy
,
2014,
J. Artif. Intell. Res..
[3]
Çetin Meriçli,et al.
Task Refinement for Autonomous Robots Using Complementary Corrective Human Feedback
,
2011
.
[4]
Brett Browning,et al.
A survey of robot learning from demonstration
,
2009,
Robotics Auton. Syst..
[5]
Manuela M. Veloso,et al.
Visual sonar: fast obstacle avoidance using monocular vision
,
2003,
Proceedings 2003 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2003) (Cat. No.03CH37453).
[6]
Andrea Lockerd Thomaz,et al.
Reinforcement Learning with Human Teachers: Evidence of Feedback and Guidance with Implications for Learning Performance
,
2006,
AAAI.
[7]
Daniel H. Grollman,et al.
Incremental learning of subtasks from unsegmented demonstration
,
2010,
2010 IEEE/RSJ International Conference on Intelligent Robots and Systems.