Joint control of manufacturing and onsite microgrid system via novel neural-network integrated reinforcement learning algorithms

Microgrid is a promising technology of distributed energy supply system, which typically consists of storage devices, generation capacities including renewable sources, and controllable loads. It has been widely investigated and applied for residential & commercial end use customers as well as critical facilities. In this paper, we propose a joint dynamic control model of microgrids and manufacturing systems using Markov Decision Process (MDP) to identify an optimal control strategy for both microgrid components and manufacturing system so that the energy cost for production can be minimized without sacrificing production throughput. The proposed MDP model has a high dimensional state/action space and is complicated in that the state and action spaces have both discrete and continuous parts and are intertwined through constraints. To resolve these challenges, a novel reinforcement learning algorithm that leverages both on-policy temporal difference control (TDcontrol) and deterministic policy gradient (DPG) algorithms is proposed. In this algorithm, the values of discrete decision actions are learned through neural network integrated temporal difference iteration, while the parameterized values of continuous actions are learned from deterministic policy gradients. The constraints are then addressed via proximal projection operators at the policy gradient updates. Experiments for a manufacturing system with an onsite microgrid with renewable sources have been implemented to identify optimal control actions for both manufacturing system and microgrid components towards cost optimality. The experimental results show the effectiveness of combining TD control and policy gradient methodologies in addressing the “curse of dimensionality” in dynamic decisionmaking with high dimensional and complicated state and action spaces.

[1]  Sriram Raghavan,et al.  Integration of Sustainable Manufacturing Systems into Smart Grids with High Penetration of Renewable Energy Resources , 2016, 2016 IEEE Green Technologies Conference (GreenTech).

[2]  John N. Tsitsiklis,et al.  Neuro-Dynamic Programming , 1996, Encyclopedia of Machine Learning.

[3]  Haibo He,et al.  Dynamic Energy Management of a Microgrid Using Approximate Dynamic Programming and Deep Recurrent Neural Network Learning , 2019, IEEE Transactions on Smart Grid.

[4]  Zeyi Sun,et al.  Inventory control for peak electricity demand reduction of manufacturing systems considering the tradeoff between production loss and energy savings , 2014 .

[5]  Oriol Gomis-Bellmunt,et al.  Trends in Microgrid Control , 2014, IEEE Transactions on Smart Grid.

[6]  Alexandre Oudalov,et al.  Microgrids enter the mainstream , 2016 .

[7]  Yves Dallery,et al.  On modeling failure and repair times in stochastic models of manufacturing systems using generalized exponential distributions , 1994, Queueing Syst. Theory Appl..

[8]  Y. Wang,et al.  Time-of-use based electricity demand response for sustainable manufacturing systems , 2013 .

[9]  Xufeng Yao,et al.  Simulation-Based Investigation for the Application of Microgrid with Renewable Sources in Manufacturing Systems Towards Sustainability , 2016 .

[10]  Lin Li,et al.  “Just-for-Peak” buffer inventory for peak electricity demand reduction of manufacturing systems , 2013 .

[11]  Adam Hawkes,et al.  Cost-effective operating strategy for residential micro-combined heat and power , 2007 .

[12]  B. Lasseter,et al.  Microgrids [distributed power generation] , 2001, 2001 IEEE Power Engineering Society Winter Meeting. Conference Proceedings (Cat. No.01CH37194).

[13]  Zeyi Sun,et al.  Dynamic Energy Control for Energy Efficiency Improvement of Sustainable Manufacturing Systems Using Markov Decision Process , 2013, IEEE Transactions on Systems, Man, and Cybernetics: Systems.

[14]  Cassiano Rech,et al.  Design of a sustainable residential microgrid system including PHEV and energy storage device , 2011, Proceedings of the 2011 14th European Conference on Power Electronics and Applications.

[15]  Wenqing Hu,et al.  Joint Manufacturing and Onsite Microgrid System Control Using Markov Decision Process and Neural Network Integrated Reinforcement Learning , 2019, Procedia Manufacturing.

[16]  Guy Lever,et al.  Deterministic Policy Gradient Algorithms , 2014, ICML.

[17]  Nithiyananthan Kannan Microgrid , 2020, Research Trends and Challenges in Smart Grids.

[18]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[19]  Stephen P. Boyd,et al.  Proximal Algorithms , 2013, Found. Trends Optim..

[20]  Md. Monirul Islam,et al.  Onsite generation system sizing for manufacturing plant considering renewable sources towards sustainability , 2019, Sustainable Energy Technologies and Assessments.

[21]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[22]  Sami Kara,et al.  Towards Energy and Resource Efficient Manufacturing: A Processes and Systems Approach , 2012 .

[23]  Alex Graves,et al.  Playing Atari with Deep Reinforcement Learning , 2013, ArXiv.

[24]  Huaguang Zhang,et al.  Real-Time Energy Management of a Microgrid Using Deep Reinforcement Learning , 2019, Energies.

[25]  F. Katiraei,et al.  Planned islanding on rural feeders — utility perspective , 2008, 2008 IEEE Power and Energy Society General Meeting - Conversion and Delivery of Electrical Energy in the 21st Century.

[26]  Haoyi Xiong,et al.  Design the Capacity of Onsite Generation System with Renewable Sources for Manufacturing Plant , 2017 .

[27]  Yong Wang,et al.  A Novel Modeling Method for Both Steady-State and Transient Analyses of Serial Bernoulli Production Systems , 2015, IEEE Transactions on Systems, Man, and Cybernetics: Systems.

[28]  Enrico Zio,et al.  Reinforcement learning for microgrid energy management , 2013 .