GP-RVM: Genetic Programing-Based Symbolic Regression Using Relevance Vector Machine

This paper proposes a hybrid basis function construction method (GP-RVM) for Symbolic Regression problem, which combines an extended version of Genetic Programming called Kaizen Programming and Relevance Vector Machine to evolve an optimal set of basis functions. Different from traditional evolutionary algorithms where a single individual is a complete solution, our method proposes a solution based on linear combination of basis functions built from individuals during the evolving process. RVM which is a sparse Bayesian kernel method selects suitable functions to constitute the basis. RVM determines the posterior weight of a function by evaluating its quality and sparsity. The solution produced by GP-RVM is a sparse Bayesian linear model of the coefficients of many non-linear functions. Our hybrid approach is focused on nonlinear white-box models selecting the right combination of functions to build robust predictions without prior knowledge about data. Experimental results show that GP-RVM outperforms conventional methods, which suggest that it is an efficient and accurate technique for solving SR. The computational complexity of GP-RVM scales in O(M3), where M is the number of functions in the basis set and is typically much smaller than the number N of training patterns.

[1]  Ayan Chatterjee,et al.  Efficiency Analysis of Genetic Algorithm and Genetic Programming in Data Mining and Image Processing , 2017 .

[2]  Vinicius Veloso de Melo,et al.  Predicting High-Performance Concrete Compressive Strength Using Features Constructed by Kaizen Programming , 2015, 2015 Brazilian Conference on Intelligent Systems (BRACIS).

[3]  Henri Luchian,et al.  Symbolic regression on noisy data with genetic and gene expression programming , 2005, Seventh International Symposium on Symbolic and Numeric Algorithms for Scientific Computing (SYNASC'05).

[4]  John R. Koza,et al.  Genetic programming as a means for programming computers by natural selection , 1994 .

[5]  George Eastman House,et al.  Sparse Bayesian Learning and the Relevance Vector Machine , 2001 .

[6]  Peter Nordin,et al.  Genetic programming - An Introduction: On the Automatic Evolution of Computer Programs and Its Applications , 1998 .

[7]  Özgür Kisi,et al.  Predicting river daily flow using wavelet-artificial neural networks based on regression analyses in comparison with artificial neural networks and support vector machine models , 2017, Neural computing & applications (Print).

[8]  Chang-an Yuan,et al.  An improved Gene Expression Programming approach for symbolic regression problems , 2014, Neurocomputing.

[9]  Josh C. Bongard,et al.  Improving genetic programming based symbolic regression using deterministic machine learning , 2013, 2013 IEEE Congress on Evolutionary Computation.

[10]  Michael D. Todd,et al.  Automated Feature Design for Numeric Sequence Classification by Genetic Programming , 2015, IEEE Transactions on Evolutionary Computation.

[11]  Michael E. Tipping,et al.  Fast Marginal Likelihood Maximisation for Sparse Bayesian Models , 2003 .

[12]  Vinicius Veloso de Melo,et al.  Studying bloat control and maintenance of effective code in linear genetic programming for symbolic regression , 2016, Neurocomputing.

[13]  John Fulcher,et al.  Computational Intelligence: An Introduction , 2008, Computational Intelligence: A Compendium.

[14]  Vinicius Veloso de Melo,et al.  Breast cancer detection with logistic regression improved by features constructed by Kaizen programming in a hybrid approach , 2016, 2016 IEEE Congress on Evolutionary Computation (CEC).

[15]  Krzysztof Krawiec,et al.  Geometric Semantic Genetic Programming , 2012, PPSN.

[16]  Victor Ciesielski,et al.  Linear genetic programming , 2008, Genetic Programming and Evolvable Machines.

[17]  R. Tibshirani,et al.  Least angle regression , 2004, math/0406456.

[18]  Vinicius Veloso de Melo,et al.  Improving the prediction of material properties of concrete using Kaizen Programming with Simulated Annealing , 2017, Neurocomputing.

[19]  Krzysztof Krawiec,et al.  Multiple regression genetic programming , 2014, GECCO.

[20]  Michael D. Schmidt,et al.  Symbolic Regression of Implicit Equations , 2010 .

[21]  Nasser M. Nasrabadi,et al.  Pattern Recognition and Machine Learning , 2006, Technometrics.

[22]  Leonardo Vanneschi,et al.  Genetic programming needs better benchmarks , 2012, GECCO '12.

[23]  Vinicius Veloso de Melo,et al.  Classification of Cardiac Arrhythmia by Random Forests with Features Constructed by Kaizen Programming with Linear Genetic Programming , 2016, GECCO.

[24]  Xu Tan,et al.  Discovery scientific laws by hybrid evolutionary model , 2015, Neurocomputing.

[25]  Arthur K. Kordon,et al.  Prime-Time: Symbolic Regression Takes Its Place in the Real World , 2016 .

[26]  W. Banzhaf,et al.  Improving Logistic Regression Classification of Credit Approval with Features Constructed by Kaizen Programming , 2016, GECCO.

[27]  Leonardo Vanneschi,et al.  Geometric Semantic Genetic Programming with Local Search , 2015, GECCO.

[28]  Hod Lipson,et al.  Distilling Free-Form Natural Laws from Experimental Data , 2009, Science.

[29]  Mohamed A. Saleh,et al.  Transfer learning with long term artificial neural network memory (LTANN-MEM) and neural symbolization algorithm (NSA) for solving high dimensional multi-objective symbolic regression problems , 2017, 2017 34th National Radio Science Conference (NRSC).

[30]  Radford M. Neal Pattern Recognition and Machine Learning , 2007, Technometrics.

[31]  Vinicius Veloso de Melo,et al.  Kaizen Programming for Feature Construction for Classification , 2016 .

[32]  Dervis Karaboga,et al.  Artificial bee colony programming for symbolic regression , 2012, Inf. Sci..

[33]  B. Schieber,et al.  Globally Optimal Symbolic Regression , 2017, 1710.10720.

[34]  Nello Cristianini,et al.  An Introduction to Support Vector Machines and Other Kernel-based Learning Methods , 2000 .