Reinforcement Learning for Engine Idle Speed Control

A control method of neural network controller with reinforcement learning is proposed to implement idle speed control of an automobile engine to reduce fluctuation of the idle speed. Firstly, the reinforcement-learning neural network is demonstrated in detail. Then, the control scheme of the reinforcement-learning controller is designed to experiment. Q learning algorithm, as one of methods of reinforcement learning, is used for learning of the neural network, which is based on evaluating the system performance and giving credit for successful actions. After the proposed controller is trained fully, the contrast experiments are implemented on an engine test bench between the proposed controller and the original controller. Experimental results show that the reinforcement learning controller has better performance in speed fluctuation and its frequency and fuel economy than that of the original controller. And, the transition of the transient idle speed controlled by the proposed controller is more smooth and stable. Meanwhile, exhaust emissions are tested during the conditions controlled by the two types of controllers respectively. And results demonstrate that the proposed controller has better fuel economy because of its lower exhaust emissions.

[1]  Sarah K. Spurgeon,et al.  Automotive engine speed control: A robust nonlinear control framework , 2001 .

[2]  M. Livshiz,et al.  Nonlinear engine model for idle speed control , 1994, Proceedings of 1994 33rd IEEE Conference on Decision and Control.

[3]  Jaehong Park,et al.  Application of adaptive control to the fluctuation of engine speed at idle , 2007, Inf. Sci..

[4]  Christopher Edwards,et al.  Sliding mode configurations for automotive engine control , 1999 .

[5]  Jianqiang Yi,et al.  Stabilization fuzzy control of parallel-type double inverted pendulum system , 2000, Ninth IEEE International Conference on Fuzzy Systems. FUZZ- IEEE 2000 (Cat. No.00CH37063).

[6]  Sridhar Mahadevan,et al.  Recent Advances in Hierarchical Reinforcement Learning , 2003, Discret. Event Dyn. Syst..

[7]  Ralph Neuneier,et al.  Risk-Sensitive Reinforcement Learning , 1998, Machine Learning.

[8]  Andrew James Smith,et al.  Applications of the self-organising map to reinforcement learning , 2002, Neural Networks.

[9]  Jean-Michel Jolion,et al.  Normalized radial basis function networks and bilinear discriminant analysis for face recognition , 2005, IEEE Conference on Advanced Video and Signal Based Surveillance, 2005..

[10]  Pu Sun,et al.  Optimal idle speed control of an automotive engine , 2000, Proceedings of the 2000 American Control Conference. ACC (IEEE Cat. No.00CH36334).

[11]  Kenji Doya,et al.  Meta-learning in Reinforcement Learning , 2003, Neural Networks.

[12]  Z. Ye,et al.  Modeling, Identification, Design, and Implementation of Nonlinear Automotive Idle Speed Control Systems—An Overview , 2007, IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews).

[13]  Martin A. Riedmiller Concepts and Facilities of a Neural Reinforcement Learning Control Architecture for Technical Process Control , 1999, Neural Computing & Applications.

[15]  Jaehong Park,et al.  Reducing automotive engine speed fluctuation at idle , 1996, IEEE Trans. Control. Syst. Technol..

[16]  Jianqiang Yi,et al.  A new fuzzy controller for stabilization of parallel-type double inverted pendulum system , 2002, Fuzzy Sets Syst..

[17]  Robert N. K. Loh,et al.  Modeling, design and implementation of discrete sliding mode control for an engine idle speed control system , 2002, Proceedings of the 2002 American Control Conference (IEEE Cat. No.CH37301).