Self Learning in Flexible Manufacturing Units: A Reinforcement Learning Approach

This paper presents a novel approach for self-learning as well as plug-and-play control of highly flexible, modular manufacturing units. The approach is inspired by recent encouraging results of reinforcement learning (RL) in computer science. However, instead of learning the entire control behavior which results in long training times and requires huge data sets, we restrict the learning process to the supervisor level by defining appropriate parameter from the basic control level (BCL) to be learned by learning agents. To this end, we define a set of interface parameter to the BCL programmed by IEC61131 compatible code, which will be used for learning. Typical control parameters include switching thresholds, timing parameters and transition conditions. We apply the approach to a laboratory testbed consisting of different production modules which underlines the efficiency improvements for manufacturing units. In addition, plug-and-produce control is enabled by the approach as different configuration of production modules can efficiently be put in operation by re-learning the parameter sets.

[1]  Mahesan Niranjan,et al.  On-line Q-learning using connectionist systems , 1994 .

[2]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[3]  Ian H. Witten,et al.  An Adaptive Optimal Controller for Discrete-Time Markov Environments , 1977, Inf. Control..

[4]  Jürgen Beyerer,et al.  Plug & produce by modelling skills and service-oriented orchestration of reconfigurable manufacturing systems , 2015, Autom..

[5]  Yuan Zou,et al.  Reinforcement Learning–Based Energy Management Strategy for a Hybrid Electric Tracked Vehicle , 2015 .

[6]  Valeriy Vyatkin,et al.  Bridging Service-Oriented Architecture and IEC 61499 for Flexibility and Interoperability , 2015, IEEE Transactions on Industrial Informatics.

[7]  Dazhong Wu,et al.  Deep learning for smart manufacturing: Methods and applications , 2018, Journal of Manufacturing Systems.

[8]  Enrico Zio,et al.  Reinforcement learning for microgrid energy management , 2013 .

[9]  Tommaso Cucinotta,et al.  A Real-Time Service-Oriented Architecture for Industrial Automation , 2009, IEEE Transactions on Industrial Informatics.

[10]  Steven X. Ding,et al.  Self-optimization of energy consumption in complex bulk good processes using reinforcement learning , 2017, 2017 IEEE 15th International Conference on Industrial Informatics (INDIN).

[11]  L. Buşoniu,et al.  A comprehensive survey of multi-agent reinforcement learning , 2011 .

[12]  A. Zoitl,et al.  Is IEC 61499 in harmony with IEC 61131-3? , 2009, IEEE Industrial Electronics Magazine.

[13]  Shalabh Bhatnagar,et al.  Natural actorcritic algorithms. , 2009 .

[14]  Jürgen Jasperneite,et al.  Requirements and concept for Plug-and-Work , 2015, Autom..

[15]  Robert W. Brennan,et al.  Developments in dynamic and intelligent reconfiguration of industrial automation , 2008, Comput. Ind..

[16]  Paulo Leitão,et al.  Industrial automation based on cyber-physical systems technologies: Prototype implementations and challenges , 2016, Comput. Ind..

[17]  Yishay Mansour,et al.  Policy Gradient Methods for Reinforcement Learning with Function Approximation , 1999, NIPS.

[18]  Klaus-Dieter Thoben,et al.  Machine learning in manufacturing: advantages, challenges, and applications , 2016 .

[19]  Alex Graves,et al.  Asynchronous Methods for Deep Reinforcement Learning , 2016, ICML.

[20]  Bogdan Gabrys,et al.  Metalearning: a survey of trends and technologies , 2013, Artificial Intelligence Review.

[21]  Shane Legg,et al.  Human-level control through deep reinforcement learning , 2015, Nature.

[22]  Dorothea Schwung,et al.  System reconfiguration of modular production units using a SOA-based control structure , 2017, 2017 IEEE 15th International Conference on Industrial Informatics (INDIN).