Deep reinforcement learning-based safe interaction for industrial human-robot collaboration using intrinsic reward function

[1]  Yuval Tassa,et al.  Continuous control with deep reinforcement learning , 2015, ICLR.

[2]  Andrew Y. Ng,et al.  Policy Invariance Under Reward Transformations: Theory and Application to Reward Shaping , 1999, ICML.

[3]  Kamel S. Saidi,et al.  Performance Evaluation of Human Detection Systems for Robot Safety , 2016, J. Intell. Robotic Syst..

[4]  Satinder Singh,et al.  On Learning Intrinsic Rewards for Policy Gradient Methods , 2018, NeurIPS.

[5]  Andrew Y. Ng,et al.  Pharmacokinetics of a novel formulation of ivermectin after administration to goats , 2000, ICML.

[6]  Gershon Elber,et al.  Continuous Collision Detection for Ellipsoids , 2009, IEEE Transactions on Visualization and Computer Graphics.

[7]  Andrea Maria Zanchettin,et al.  Trajectory generation algorithm for safe human-robot collaboration based on multiple depth sensor measurements , 2018, Mechatronics.

[8]  Rouhollah Rahmatizadeh,et al.  From Virtual Demonstration to Real-World Manipulation Using LSTM and MDN , 2016, AAAI.

[9]  Minna Lanz,et al.  Review of vision-based safety systems for human-robot collaboration , 2018 .

[10]  Matteo Parigi Polverini,et al.  Real-time collision avoidance in human-robot interaction based on kinetostatic safety field , 2014, 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[11]  Lihui Wang,et al.  Contact-less and Programming-less Human-Robot Collaboration , 2013 .

[12]  Duc Truong Pham,et al.  Dynamic risk assessment and active response strategy for industrial human-robot collaboration , 2020, Comput. Ind. Eng..

[13]  Jeremy A Marvel,et al.  Implementing Speed and Separation Monitoring in Collaborative Robot Workcells. , 2017, Robotics and computer-integrated manufacturing.

[14]  Duc Truong Pham,et al.  Human-robot collaboration in disassembly for sustainable manufacturing , 2019, Int. J. Prod. Res..

[15]  Alexei A. Efros,et al.  Curiosity-Driven Exploration by Self-Supervised Prediction , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).

[16]  Jeremy A. Marvel,et al.  Characterizing Task-Based Human–Robot Collaboration Safety in Manufacturing , 2015, IEEE Transactions on Systems, Man, and Cybernetics: Systems.

[17]  Oussama Khatib,et al.  A Depth Space Approach for Evaluating Distance to Objects , 2015, J. Intell. Robotic Syst..

[18]  Roland Siegwart,et al.  From perception to decision: A data-driven approach to end-to-end motion planning for autonomous ground robots , 2016, 2017 IEEE International Conference on Robotics and Automation (ICRA).

[19]  Filip De Turck,et al.  #Exploration: A Study of Count-Based Exploration for Deep Reinforcement Learning , 2016, NIPS.

[20]  Lihui Wang,et al.  Active collision avoidance for human–robot collaboration driven by vision sensors , 2017, Int. J. Comput. Integr. Manuf..

[21]  Herke van Hoof,et al.  Addressing Function Approximation Error in Actor-Critic Methods , 2018, ICML.

[22]  Richard L. Lewis,et al.  Where Do Rewards Come From , 2009 .

[23]  Andrea Maria Zanchettin,et al.  Safety in human-robot collaborative manufacturing environments: Metrics and control , 2016, IEEE Transactions on Automation Science and Engineering.

[24]  Francesco Leali,et al.  Survey on human–robot collaboration in industrial settings: Safety, intuitive interfaces and applications , 2018, Mechatronics.

[25]  Anca D. Dragan,et al.  Inverse Reward Design , 2017, NIPS.

[26]  Wolfram Burgard,et al.  Learning to Pour using Deep Deterministic Policy Gradients , 2018, 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[27]  Lovekesh Vig,et al.  Deterministic Policy Gradient Based Robotic Path Planning with Continuous Action Spaces , 2017, 2017 IEEE International Conference on Computer Vision Workshops (ICCVW).

[28]  Sergey Levine,et al.  Continuous Deep Q-Learning with Model-based Acceleration , 2016, ICML.

[29]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[30]  Han-Pang Huang,et al.  Virtual Impedance Control for Safe Human-Robot Interaction , 2016, J. Intell. Robotic Syst..

[31]  Richard L. Lewis,et al.  Intrinsically Motivated Reinforcement Learning: An Evolutionary Perspective , 2010, IEEE Transactions on Autonomous Mental Development.

[32]  Antonella Ferrara,et al.  Self-Configuring Robot Path Planning With Obstacle Avoidance via Deep Reinforcement Learning , 2021, IEEE Control Systems Letters.

[33]  Peter I. Corke,et al.  Towards Vision-Based Deep Reinforcement Learning for Robotic Motion Control , 2015, ICRA 2015.

[34]  Antonella Ferrara,et al.  Deep Reinforcement Learning for Collision Avoidance of Robotic Manipulators , 2018, 2018 European Control Conference (ECC).

[35]  Matteo Rubagotti,et al.  Semi-Autonomous Robot Teleoperation With Obstacle Avoidance via Model Predictive Control , 2019, IEEE Robotics and Automation Letters.

[36]  Keyu Wu,et al.  Depth-based Obstacle Avoidance through Deep Reinforcement Learning , 2019, ICMRE'19.

[37]  Guy Lever,et al.  Deterministic Policy Gradient Algorithms , 2014, ICML.

[38]  J. R. Llata,et al.  Working Together: A Review on Safe Human-Robot Collaboration in Industrial Environments , 2017, IEEE Access.

[39]  Alireza Akbarzadeh,et al.  Real-time velocity scaling and obstacle avoidance for industrial robots using fuzzy dynamic movement primitives and virtual impedances , 2018, Ind. Robot.

[40]  Richard L. Lewis,et al.  Reward Design via Online Gradient Ascent , 2010, NIPS.

[41]  M. G. Mohanan,et al.  A survey of robotic motion planning in dynamic environments , 2018, Robotics Auton. Syst..

[42]  Masayoshi Tomizuka,et al.  Human guidance programming on a 6-DoF robot with collision avoidance , 2016, 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[43]  Marc G. Bellemare,et al.  Count-Based Exploration with Neural Density Models , 2017, ICML.

[44]  Lorenzo Molinari Tosatti,et al.  Trajectory-dependent safe distances in human-robot interaction , 2014, Proceedings of the 2014 IEEE Emerging Technology and Factory Automation (ETFA).

[45]  Sergey Levine,et al.  Deep reinforcement learning for robotic manipulation with asynchronous off-policy updates , 2016, 2017 IEEE International Conference on Robotics and Automation (ICRA).