Interpretable AI Agent Through Nonlinear Decision Trees for Lane Change Problem

The recent years have witnessed a surge in application of deep neural networks (DNNs) and reinforcement learning (RL) methods to various autonomous control systems and game playing problems. While they are capable of learning from real-world data and produce adequate actions to various state conditions, their internal complexity does not allow an easy way to provide an explanation for their actions. In this paper, we generate state-action pair data from a trained DNN/RL system and employ a previously proposed nonlinear decision tree (NLDT) framework to decipher hidden simplistic rule sets that interpret the working of DNN/RL systems. The complexity of the rule sets are controllable by the user. In essence, the inherent bilevel optimization procedure that finds the NLDTs is capable of reducing the complexities of the state-action logic to a minimalist and intrepretable level. Demonstrating the working principle of the NLDT method to a revised mountain car control problem, this paper applies the methodology to the lane changing problem involving six critical cars in front and rear in left, middle, and right lanes of a pilot car. NLDTs are derived to have simplistic relationships of 12 decision variables involving relative distances and velocities of the six critical cars. The derived analytical decision rules are then simplified further by using a symbolic analysis tool to provide English-like interpretation of the lane change problem. This study makes a scratch to the issue of interpretability of modern machine learning based tools and it now deserves further attention and applications to make the overall approach more integrated and effective.

[1]  Yashesh D. Dhebar,et al.  Toward Interpretable-AI Policies Using Evolutionary Nonlinear Decision Trees for Discrete-Action Systems. , 2020, IEEE transactions on cybernetics.

[2]  Kalyanmoy Deb,et al.  Interpretable Rule Discovery Through Bilevel Optimization of Split-Rules of Nonlinear Decision Trees for Classification Problems , 2020, IEEE Transactions on Cybernetics.

[3]  Dimitar Filev,et al.  Autonomous Highway Driving using Deep Reinforcement Learning , 2019, 2019 IEEE International Conference on Systems, Man and Cybernetics (SMC).

[4]  Tom Schaul,et al.  Rainbow: Combining Improvements in Deep Reinforcement Learning , 2017, AAAI.

[5]  David Silver,et al.  Deep Reinforcement Learning with Double Q-Learning , 2015, AAAI.

[6]  Alex Graves,et al.  Playing Atari with Deep Reinforcement Learning , 2013, ArXiv.

[7]  L. Rutkowski,et al.  Decision Trees for Mining Data Streams Based on the McDiarmid's Bound , 2013, IEEE Trans. Knowl. Data Eng..

[8]  Jan Peters,et al.  Relative Entropy Policy Search , 2010, AAAI.

[9]  Leslie Pack Kaelbling,et al.  Practical Reinforcement Learning in Continuous Spaces , 2000, ICML.

[10]  R. Sutton,et al.  Advances in Neural Information Processing Systems pp MIT Press Generalization in Reinforcement Learning Successful Examples Using Sparse Coarse Coding , 2010 .

[11]  Kalyanmoy Deb,et al.  Simulated Binary Crossover for Continuous Search Space , 1995, Complex Syst..

[12]  Mahesan Niranjan,et al.  On-line Q-learning using connectionist systems , 1994 .

[13]  Andrew W. Moore,et al.  Efficient memory-based learning for robot control , 1990 .