Multi-objective genetic programming for manifold learning: balancing quality and dimensionality

Manifold learning techniques have become increasingly valuable as data continues to grow in size. By discovering a lower-dimensional representation (embedding) of the structure of a dataset, manifold learning algorithms can substantially reduce the dimensionality of a dataset while preserving as much information as possible. However, state-of-the-art manifold learning algorithms are opaque in how they perform this transformation. Understanding the way in which the embedding relates to the original high-dimensional space is critical in exploratory data analysis. We previously proposed a Genetic Programming method that performed manifold learning by evolving mappings that are transparent and interpretable. This method required the dimensionality of the embedding to be known a priori, which makes it hard to use when little is known about a dataset. In this paper, we substantially extend our previous work, by introducing a multi-objective approach that automatically balances the competing objectives of manifold quality and dimensionality. Our proposed approach is competitive with a range of baseline and state-of-the-art manifold learning methods, while also providing a range (front) of solutions that give different trade-offs between quality and dimensionality. Furthermore, the learned models are shown to often be simple and efficient, utilising only a small number of features in an interpretable manner.

[1]  Lawrence Cayton,et al.  Algorithms for manifold learning , 2005 .

[2]  Eric Medvet,et al.  An analysis of dimensionality reduction techniques for visualizing evolution , 2019, GECCO.

[3]  Dario Floreano,et al.  Neuroevolution: from architectures to learning , 2008, Evol. Intell..

[4]  Zhang Yi,et al.  Evolving Unsupervised Deep Neural Networks for Learning Meaningful Representations , 2017, IEEE Transactions on Evolutionary Computation.

[5]  Leland McInnes,et al.  UMAP: Uniform Manifold Approximation and Projection for Dimension Reduction , 2018, ArXiv.

[6]  Hiroshi Motoda,et al.  Feature Extraction, Construction and Selection: A Data Mining Perspective , 1998 .

[7]  Janez Demsar,et al.  Statistical Comparisons of Classifiers over Multiple Data Sets , 2006, J. Mach. Learn. Res..

[8]  Hiroshi Motoda,et al.  Feature Selection for Knowledge Discovery and Data Mining , 1998, The Springer International Series in Engineering and Computer Science.

[9]  Michel Verleysen,et al.  Nonlinear Dimensionality Reduction , 2021, Computer Vision.

[10]  R. Storn,et al.  Differential Evolution , 2004 .

[11]  Yadolah Dodge The Concise Encyclopedia of Statistics , 2008 .

[12]  Xiangliang Zhang,et al.  An up-to-date comparison of state-of-the-art classification algorithms , 2017, Expert Syst. Appl..

[13]  Mengjie Zhang,et al.  A Filter Approach to Multiple Feature Construction for Symbolic Learning Classifiers Using Genetic Programming , 2012, IEEE Transactions on Evolutionary Computation.

[14]  Mengjie Zhang,et al.  Automatically evolving difficult benchmark feature selection datasets with genetic programming , 2018, GECCO.

[15]  Parikshit Sondhi,et al.  Feature Construction Methods : A Survey , 2009 .

[16]  Gaël Varoquaux,et al.  Scikit-learn: Machine Learning in Python , 2011, J. Mach. Learn. Res..

[17]  Mengjie Zhang,et al.  A survey on evolutionary machine learning , 2019, Journal of the Royal Society of New Zealand.

[18]  Chandan Singh,et al.  Definitions, methods, and applications in interpretable machine learning , 2019, Proceedings of the National Academy of Sciences.

[19]  Mengjie Zhang,et al.  Can Genetic Programming Do Manifold Learning Too? , 2019, EuroGP.

[20]  Kay Chen Tan,et al.  Visualizing the Evolution of Computer Programs for Genetic Programming [Research Frontier] , 2018, IEEE Computational Intelligence Magazine.

[21]  Ivan Zelinka,et al.  Handbook of Optimization - From Classical to Modern Approach , 2012, Handbook of Optimization.

[22]  Andrew M. Webb,et al.  Evolving controllably difficult datasets for clustering , 2019, GECCO.

[23]  Andrew Rosenberg,et al.  Multi-objective Genetic Programming for Visual Analytics , 2011, EuroGP.

[24]  André L. V. Coelho,et al.  Multi-objective design of hierarchical consensus functions for clustering ensembles via genetic programming , 2011, Decis. Support Syst..

[25]  Ashutosh Kumar Singh,et al.  The Elements of Statistical Learning: Data Mining, Inference, and Prediction , 2010 .

[26]  Leland McInnes,et al.  UMAP: Uniform Manifold Approximation and Projection , 2018, J. Open Source Softw..

[27]  Paul Vos,et al.  Inference and Prediction , 2018 .

[28]  Pablo A. Estévez,et al.  Genetic programming-based clustering using an information theoretic fitness measure , 2007, 2007 IEEE Congress on Evolutionary Computation.

[29]  Krzysztof Michalak Low-Dimensional Euclidean Embedding for Visualization of Search Spaces in Combinatorial Optimization , 2019, IEEE Transactions on Evolutionary Computation.

[30]  Huan Liu,et al.  Toward integrating feature selection algorithms for classification and clustering , 2005, IEEE Transactions on Knowledge and Data Engineering.

[31]  Pascal Vincent,et al.  Representation Learning: A Review and New Perspectives , 2012, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[32]  Cardona Alzate,et al.  Predicción y selección de variables con bosques aleatorios en presencia de variables correlacionadas , 2020 .

[33]  James McDermott,et al.  Why Is Auto-Encoding Difficult for Genetic Programming? , 2019, EuroGP.

[34]  Geoffrey E. Hinton,et al.  Visualizing Data using t-SNE , 2008 .

[35]  Carla E. Brodley,et al.  Feature Selection for Unsupervised Learning , 2004, J. Mach. Learn. Res..

[36]  Chandan Singh,et al.  Definitions, methods, and applications in interpretable machine learning , 2019, Proceedings of the National Academy of Sciences.

[37]  Mengjie Zhang,et al.  Genetic programming for feature construction and selection in classification on high-dimensional data , 2016, Memetic Comput..

[38]  H. Frigui,et al.  Clustering: Algorithms and Applications , 2008, 2008 First Workshops on Image Processing Theory, Tools and Applications.

[39]  Qingfu Zhang,et al.  MOEA/D: A Multiobjective Evolutionary Algorithm Based on Decomposition , 2007, IEEE Transactions on Evolutionary Computation.

[40]  M. Kramer Nonlinear principal component analysis using autoassociative neural networks , 1991 .

[41]  S T Roweis,et al.  Nonlinear dimensionality reduction by locally linear embedding. , 2000, Science.

[42]  VARUN CHANDOLA,et al.  Anomaly detection: A survey , 2009, CSUR.

[43]  Pierre Comon,et al.  Handbook of Blind Source Separation: Independent Component Analysis and Applications , 2010 .

[44]  Xin Yao,et al.  A Survey on Evolutionary Computation Approaches to Feature Selection , 2016, IEEE Transactions on Evolutionary Computation.

[45]  Huan Liu,et al.  Feature selection for classification: A review , 2014 .

[46]  Hiroshi Motoda,et al.  Feature Extraction, Construction and Selection , 1998 .

[47]  Heng Tao Shen,et al.  Principal Component Analysis , 2009, Encyclopedia of Biometrics.

[48]  J. Kruskal Multidimensional scaling by optimizing goodness of fit to a nonmetric hypothesis , 1964 .

[49]  Charu C. Aggarwal,et al.  Data Clustering , 2013 .

[50]  Hugo Jair Escalante,et al.  Structurally Layered Representation Learning: Towards Deep Learning Through Genetic Programming , 2018, EuroGP.

[51]  Ron Kohavi,et al.  Feature Selection for Knowledge Discovery and Data Mining , 1998 .

[52]  Yu Zhang,et al.  Unsupervised Learning of Disentangled and Interpretable Representations from Sequential Data , 2017, NIPS.

[53]  E. Alzate Modelos de mezclas Bernoulli con regresión logística: una aplicación en la valoración de carteras de crédito , 2020 .

[54]  Robert Tibshirani,et al.  The Elements of Statistical Learning: Data Mining, Inference, and Prediction, 2nd Edition , 2001, Springer Series in Statistics.

[55]  Tsuyoshi Murata,et al.  {m , 1934, ACML.