NEW DEVELOPMENTS IN THE APPLICATION OF AUTOMATIC LEARNING TO POWER SYSTEM CONTROL

In this paper we present the basic principles of supervised learning and reinforcement learning as two complementary frameworks to design control laws or deci- sion policies within the context of power system control. We also review recent developments in the realm of automatic learning methods and discuss their applicability to power system decision and control problems. Simulation results il- lustrating the potentials of the recently introduced fitted Q iteration learning algorithm in controlling a TCSC device aimed to damp electro-mechanical oscillations in a synthetic 4-machine system, are included in the paper.

[1]  Nello Cristianini,et al.  An introduction to Support Vector Machines , 2000 .

[2]  Alan F. Blackwell,et al.  Programming , 1973, CSC '73.

[3]  Manfred Morari,et al.  Model predictive control: Theory and practice , 1988 .

[4]  Louis Wehenkel Whither dynamic congestion management , 2004 .

[5]  Damien Ernst,et al.  Near Optimal Closed-Loop Control. Application to Electric Power Systems , 2003 .

[6]  D. Ruppert The Elements of Statistical Learning: Data Mining, Inference, and Prediction , 2004 .

[7]  Pierre Geurts,et al.  Iteratively Extending Time Horizon Reinforcement Learning , 2003, ECML.

[8]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.

[9]  Graham Rogers,et al.  Power System Oscillations , 1999 .

[10]  Pierre Geurts,et al.  Contributions to decision tree induction: bias/variance tradeoff and time series classification , 2002 .

[11]  Eric Bauer,et al.  An Empirical Comparison of Voting Classification Algorithms: Bagging, Boosting, and Variants , 1999, Machine Learning.

[12]  D. Ernst,et al.  Power systems stability control: reinforcement learning framework , 2004, IEEE Transactions on Power Systems.

[13]  Liming Xiang,et al.  Kernel-Based Reinforcement Learning , 2006, ICIC.

[14]  Pierre Geurts,et al.  Tree-Based Batch Mode Reinforcement Learning , 2005, J. Mach. Learn. Res..

[15]  Chris Watkins,et al.  Learning from delayed rewards , 1989 .

[16]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[17]  Manfred Morari,et al.  Model predictive control: Theory and practice - A survey , 1989, Autom..

[18]  Michael P. Wellman,et al.  Nash Q-Learning for General-Sum Stochastic Games , 2003, J. Mach. Learn. Res..

[19]  R. Bellman Dynamic programming. , 1957, Science.

[20]  D. Ernst,et al.  Combining a stability and a performance-oriented control in power systems , 2005, IEEE Transactions on Power Systems.

[21]  Yishay Mansour,et al.  Learning Rates for Q-learning , 2004, J. Mach. Learn. Res..

[22]  D. Ernst,et al.  Approximate Value Iteration in the Reinforcement Learning Context. Application to Electrical Power System Control. , 2005 .

[23]  Andrew W. Moore,et al.  Reinforcement Learning: A Survey , 1996, J. Artif. Intell. Res..

[24]  Louis Wehenkel,et al.  Automatic Learning Techniques in Power Systems , 1997 .

[25]  Louis Wehenkel,et al.  FACTS devices controlled by means of reinforcement learning algorithms , 2002 .

[26]  Dimitri P. Bertsekas,et al.  Dynamic Programming and Optimal Control, Two Volume Set , 1995 .

[27]  Richard S. Sutton,et al.  Dimensions of Reinforcement Learning , 1998 .

[28]  Pierre Geurts,et al.  Extremely randomized trees , 2006, Machine Learning.

[29]  Shaopeng Wang,et al.  Power system damping controller design-using multiple input signals , 2000 .

[30]  P. Kundur,et al.  Power system stability and control , 1994 .