A learning-based shared control architecture for interactive task execution

Shared control is a key technology for various robotic applications in which a robotic system and a human operator are meant to collaborate efficiently. In order to achieve efficient task execution in shared control, it is essential to predict the desired behavior for a given situation or context in order to simplify the control task for the human operator. This prediction is obtained by exploiting Learning from Demonstration (LfD), which is a popular approach for transferring human skills to robots. We encode the demonstrated behavior as trajectory distributions and generalize the learned distributions to new situations. The goal of this paper is to present a shared control framework that uses learned expert distributions to gain more autonomy. Our approach controls the balance between the controller's autonomy and the human preference based on the distributions of the demonstrated trajectories. Moreover, the learned distributions are autonomously refined from collaborative task executions, resulting in a master-slave system with increasing autonomy that requires less user input with an increasing number of task executions. We experimentally validated that our shared control approach enables efficient task executions. Moreover, the conducted experiments demonstrated that the developed system improves its performances through interactive task executions with our shared control.

[1]  Paolo Robuffo Giordano,et al.  A visual-based shared control architecture for remote telemanipulation , 2016, 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[2]  Antonio Franchi,et al.  Semi-autonomous trajectory generation for mobile robots with integral haptic shared control , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).

[3]  Louis B. Rosenberg,et al.  Virtual fixtures: Perceptual tools for telerobotic manipulation , 1993, Proceedings of IEEE Virtual Reality Annual International Symposium.

[4]  Allison M. Okamura,et al.  Pseudo-admittance Bilateral Telemanipulation with Guidance Virtual Fixtures , 2007 .

[5]  Andrew W. Moore,et al.  Locally Weighted Learning , 1997, Artificial Intelligence Review.

[6]  Mamoru Mitsuishi,et al.  Online Trajectory Planning in Dynamic Environments for Surgical Task Automation , 2014, Robotics: Science and Systems.

[7]  Geoffrey J. Gordon,et al.  A Reduction of Imitation Learning and Structured Prediction to No-Regret Online Learning , 2010, AISTATS.

[8]  Jan Peters,et al.  Probabilistic Movement Primitives , 2013, NIPS.

[9]  Robert D. Howe,et al.  Cooperative Human and Machine Perception in Teleoperated Assembly , 2000, ISER.

[10]  Bruno Castro da Silva,et al.  Learning parameterized motor skills on a humanoid robot , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).

[11]  Antonio Franchi,et al.  Interactive planning of persistent trajectories for human-assisted navigation of mobile robots , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[12]  Sylvain Calinon,et al.  A tutorial on task-parameterized movement learning and retrieval , 2016, Intell. Serv. Robotics.

[13]  David A. Cohn,et al.  Active Learning with Statistical Models , 1996, NIPS.

[14]  Andrew W. Moore,et al.  Locally Weighted Learning for Control , 1997, Artificial Intelligence Review.

[15]  Siddhartha S. Srinivasa,et al.  Formalizing Assistive Teleoperation , 2012, Robotics: Science and Systems.

[16]  Blake Hannaford,et al.  Force-reflection and shared compliant control in operating telemanipulators with time delay , 1992, IEEE Trans. Robotics Autom..

[17]  Oussama Khatib,et al.  Springer Handbook of Robotics , 2007, Springer Handbooks.

[18]  Wayne Book,et al.  Blended Shared Control of Zermelo's navigation problem , 2010, Proceedings of the 2010 American Control Conference.

[19]  Allison M. Okamura,et al.  Haptic Virtual Fixtures for Robot-Assisted Manipulation , 2005, ISRR.

[20]  D. Kortenkamp,et al.  Adjustable control autonomy for manned space flight , 2000, 2000 IEEE Aerospace Conference. Proceedings (Cat. No.00TH8484).

[21]  R. A. Leibler,et al.  On Information and Sufficiency , 1951 .

[22]  Brett Browning,et al.  A survey of robot learning from demonstration , 2009, Robotics Auton. Syst..

[23]  Albert N. Andry,et al.  A simple calculation for the critical damping matrix of a linear multidegree of freedom system , 1982 .

[24]  Bruno Siciliano,et al.  Six-DOF impedance control based on angle/axis representations , 1999, IEEE Trans. Robotics Autom..

[25]  Manuela M. Veloso,et al.  Interactive Policy Learning through Confidence-Based Autonomy , 2014, J. Artif. Intell. Res..

[26]  Mamoru Mitsuishi,et al.  Hybrid Rate—Admittance Control With Force Reflection for Safe Teleoperated Surgery , 2015, IEEE/ASME Transactions on Mechatronics.

[27]  Carme Torras,et al.  Learning Physical Collaborative Robot Behaviors From Human Demonstrations , 2016, IEEE Transactions on Robotics.

[28]  Antonio Franchi,et al.  Shared Control : Balancing Autonomy and Human Assistance with a Group of Quadrotor UAVs , 2012, IEEE Robotics & Automation Magazine.

[29]  Aude Billard,et al.  Incremental learning of gestures by imitation in a humanoid robot , 2007, 2007 2nd ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[30]  Pieter Abbeel,et al.  Autonomous Helicopter Aerobatics through Apprenticeship Learning , 2010, Int. J. Robotics Res..