Intelligent Human–Robot Interaction Systems Using Reinforcement Learning and Neural Networks

In this chapter, an intelligent human–robot system with adjustable robot autonomy is presented to assist the human operator to perform a given task with minimum workload demands and optimal performance. The proposed control methodology consists of two feedback loops: an inner loop that makes the robot with unknown dynamics behave like a prescribed impedance model as perceived by the operator, and an outer loop that finds the optimal parameters of this model to adjust the robot’s dynamics to the operator skills and minimize the tracking error . A nonlinear robust controller using neural networks is used in the inner loop to make the nonlinear unknown robot dynamics behave like a prescribed impedance model. The problem of finding the optimal parameters of the prescribed impedance model is formulated as an optimal control problem in the outer loop. The objective is to minimize the human effort and optimize the closed-loop behavior of the human–machine system for a given task. This design must take into account the unknown human dynamics as well as the desired overall performance of the human–robot system, which depends on the task. To obviate the requirement of the knowledge of the human model, reinforcement learning is used to learn the solution to the given optimal control problem online in real time.

[1]  Clément Gosselin,et al.  Safe, Stable and Intuitive Control for Physical Human-Robot Interaction , 2009, 2009 IEEE International Conference on Robotics and Automation.

[2]  Yoshiyuki Tanaka,et al.  Tracking control properties of human-robotic systems based on impedance control , 2005, IEEE Transactions on Systems, Man, and Cybernetics - Part A: Systems and Humans.

[3]  Shuzhi Sam Ge,et al.  Neural network based adaptive impedance control of constrained robots , 2002, Proceedings of the IEEE Internatinal Symposium on Intelligent Control.

[4]  Shuzhi Sam Ge,et al.  Human–Robot Collaboration Based on Motion Intention Estimation , 2014, IEEE/ASME Transactions on Mechatronics.

[5]  K. Furuta,et al.  Human adaptive mechatronics (HAM) for haptic system , 2004, 30th Annual Conference of IEEE Industrial Electronics Society, 2004. IECON 2004.

[6]  Toru Tsumugiwa,et al.  Variable impedance control based on estimation of human arm stiffness for human-robot cooperative calligraphic task , 2002, Proceedings 2002 IEEE International Conference on Robotics and Automation (Cat. No.02CH37292).

[7]  Frank L. Lewis,et al.  Optimized Assistive Human–Robot Interaction Using Reinforcement Learning , 2016, IEEE Transactions on Cybernetics.

[8]  Frank L. Lewis,et al.  Optimal Control: Lewis/Optimal Control 3e , 2012 .

[9]  Shahid Hussain,et al.  Adaptive Impedance Control of a Robotic Orthosis for Gait Rehabilitation , 2013, IEEE Transactions on Cybernetics.

[10]  Frank L. Lewis,et al.  Optimal Adaptive Control and Differential Games by Reinforcement Learning Principles , 2012 .

[11]  Keng Peng Tee,et al.  Continuous critic learning for robot control in physical human-robot interaction , 2013, 2013 13th International Conference on Control, Automation and Systems (ICCAS 2013).

[12]  Frank L. Lewis,et al.  Neural net robot controller with guaranteed tracking performance , 1995, IEEE Trans. Neural Networks.

[13]  A. Tustin,et al.  The nature of the operator's response in manual control, and its implications for controller design , 1947 .

[14]  Shuzhi Sam Ge,et al.  Adaptive Neural Network Control of Robotic Manipulators , 1999, World Scientific Series in Robotics and Intelligent Systems.

[15]  Shuzhi Sam Ge,et al.  Impedance control for multi-point human-robot interaction , 2011, 2011 8th Asian Control Conference (ASCC).

[16]  Sae Franklin,et al.  Visuomotor feedback gains upregulate during the learning of novel dynamics , 2012, Journal of neurophysiology.

[17]  Kiyoshi Ohishi,et al.  Variable mechanical stiffness control based on human stiffness estimation , 2011, 2011 IEEE International Conference on Mechatronics.

[18]  Frank L. Lewis,et al.  2009 Special Issue: Neural network approach to continuous-time direct adaptive optimal control for partially unknown nonlinear systems , 2009 .

[19]  Neville Hogan,et al.  Impedance control - An approach to manipulation. I - Theory. II - Implementation. III - Applications , 1985 .

[20]  Kyoungchul Kong,et al.  Frequency-Shaped Impedance Control for Safe Human–Robot Interaction in Reference Tracking Application , 2014, IEEE/ASME Transactions on Mechatronics.

[21]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[22]  Frank L. Lewis,et al.  Robot Manipulator Control: Theory and Practice , 2003 .

[23]  Katsuhisa Furuta,et al.  Adaptive impedance control to enhance human skill on a haptic interface system , 2012 .

[24]  Abderrahmane Kheddar,et al.  Motion learning and adaptive impedance for robot control during physical interaction with humans , 2011, 2011 IEEE International Conference on Robotics and Automation.

[25]  Neville Hogan,et al.  Impedance Control: An Approach to Manipulation: Part I—Theory , 1985 .

[26]  Shuzhi Sam Ge,et al.  Adaptive Neural Network Control For Smart Materials Robots Using Singular Perturbation Technique , 2008 .

[27]  Warren B. Powell,et al.  Approximate Dynamic Programming - Solving the Curses of Dimensionality , 2007 .

[28]  Stefan Schaal,et al.  Model-Free Reinforcement Learning of Impedance Control in Stochastic Environments , 2012, IEEE Transactions on Autonomous Mental Development.

[29]  Frank L. Lewis,et al.  Adaptive optimal control for continuous-time linear systems based on policy iteration , 2009, Autom..

[30]  Aiguo Song,et al.  Adaptive impedance control based on dynamic recurrent fuzzy neural network for upper-limb rehabilitation robot , 2009, 2009 IEEE International Conference on Control and Automation.

[31]  Kazuhiro Kosuge,et al.  Virtual internal model following control of robot arms , 1987, Proceedings. 1987 IEEE International Conference on Robotics and Automation.

[32]  Frank L. Lewis,et al.  Multilayer neural-net robot controller with guaranteed tracking performance , 1996, IEEE Trans. Neural Networks.

[33]  F. Lewis,et al.  Reinforcement Learning and Feedback Control: Using Natural Decision Methods to Design Optimal Adaptive Controllers , 2012, IEEE Control Systems.

[34]  Seul Jung,et al.  Neural network impedance force control of robot manipulator , 1998, IEEE Trans. Ind. Electron..

[35]  D. Kleinman,et al.  An optimal control model of human response part II: Prediction of human performance in a complex task , 1970 .

[36]  Ryojun Ikeura,et al.  Optimal variable impedance control for a robot and its application to lifting an object with a human , 2002, Proceedings. 11th IEEE International Workshop on Robot and Human Interactive Communication.

[37]  Warren B. Powell,et al.  Approximate Dynamic Programming: Solving the Curses of Dimensionality (Wiley Series in Probability and Statistics) , 2007 .

[38]  Katsuhisa Furuta,et al.  Assisting control in Human Adaptive Mechatronics — single ball juggling — , 2006, 2006 IEEE Conference on Computer Aided Control System Design, 2006 IEEE International Conference on Control Applications, 2006 IEEE International Symposium on Intelligent Control.